Qwen3-Coder-30B-A3B-Instruct is an outstanding example of a high-quality large language model with advanced specialization in programming. This Mixture-of-Experts model has 30.5 billion total parameters, of which only 3.3 billion are activated per token, and out of 128 experts, only 8 are activated per token. The model comprises 48 hidden layers with grouped query attention (32 heads for Q and 4 for KV), delivering exceptional processing efficiency with minimal computational resource consumption. Native support for a 262,144-token context window—expandable up to 1 million tokens via Yarn—makes the model ideal for working with large code repositories within complex projects.
The key unique feature of Qwen3-Coder-30B-A3B-Instruct lies in its superior agent capabilities. The model does not merely generate code; it autonomously interacts with development tools, executes multi-step programming tasks, and is capable of solving complex problems without human intervention. On the LiveCodeBench v6 benchmark, the model achieves an impressive 66.0%, significantly outperforming the base version Qwen3-30B-A3B (57.4%). In AIME25 tasks (advanced mathematics for programming), it demonstrates 85.0% accuracy, surpassing Gemini-2.5-Flash-Thinking (72.0%) and confidently competing with much larger models. The model outperforms DeepSeek V3 on most coding tasks and delivers agent workflow performance comparable to Claude Sonnet 4, a remarkable achievement for an open-source solution.
Qwen3-Coder-30B-A3B-Instruct unlocks entirely new possibilities in software development. The model is integrated with popular agent-based programming platforms, including Qwen Code, CLINE, Roo Code, and Kilo Code, offering a unified function-calling format for seamless operation within CI/CD pipelines. Support for 358 programming languages makes it a universal reference tool for developers. The model particularly excels in repository-scale understanding scenarios, where it can analyze and modify massive codebases, automatically refactor legacy code, and create complex full-stack applications with minimal developer intervention.
Model Name | Context | Type | GPU | TPS | Status | Link |
---|
There are no public endpoints for this model yet.
Rent your own physically dedicated instance with hourly or long-term monthly billing.
We recommend deploying private instances in the following scenarios:
Name | vCPU | RAM, MB | Disk, GB | GPU | |||
---|---|---|---|---|---|---|---|
16 | 65536 | 160 | 2 | $0.93 | Launch | ||
16 | 65536 | 160 | 4 | $1.18 | Launch | ||
16 | 65536 | 160 | 4 | $1.48 | Launch | ||
16 | 65536 | 160 | 2 | $1.67 | Launch | ||
16 | 65536 | 160 | 2 | $2.19 | Launch | ||
16 | 65536 | 160 | 1 | $2.58 | Launch | ||
16 | 65536 | 160 | 2 | $2.93 | Launch | ||
16 | 65536 | 160 | 1 | $5.11 | Launch |
Name | vCPU | RAM, MB | Disk, GB | GPU | |||
---|---|---|---|---|---|---|---|
16 | 98304 | 160 | 3 | $1.34 | Launch | ||
16 | 65536 | 160 | 4 | $1.48 | Launch | ||
16 | 65535 | 240 | 2 | $2.22 | Launch | ||
16 | 98304 | 160 | 3 | $2.45 | Launch | ||
16 | 65536 | 160 | 1 | $2.58 | Launch | ||
16 | 65536 | 160 | 2 | $2.93 | Launch | ||
16 | 98304 | 160 | 3 | $3.23 | Launch | ||
16 | 65536 | 160 | 1 | $5.11 | Launch |
Name | vCPU | RAM, MB | Disk, GB | GPU | |||
---|---|---|---|---|---|---|---|
16 | 131072 | 160 | 4 | $1.75 | Launch | ||
16 | 131072 | 160 | 4 | $3.23 | Launch | ||
16 | 131072 | 160 | 4 | $4.26 | Launch | ||
16 | 98304 | 160 | 3 | $4.34 | Launch | ||
24 | 262144 | 160 | 2 | $5.35 | Launch | ||
24 | 262144 | 160 | 2 | $10.40 | Launch |
Contact our dedicated neural networks support team at nn@immers.cloud or send your request to the sales department at sale@immers.cloud.