Qwen3-30B-A3B-Instruct-2507-AWQ is the AWQ int4 quantized version of Qwen3-30B-A3B-Instruct-2507, with a total of 30.5 billion parameters and 3.3 billion active parameters. This model has significant improvements in instruction following, logical reasoning, text understanding, mathematics, science, coding, and tool usage. It supports the ability to understand 256K long contexts and is compatible with Transformers and vLLM for efficient text generation.
Natural Language Processing
Transformers