Qwen3-Coder: A Powerful Coding Assistant with 480B Parameters
The article introduces Tongyi Qianwen's latest open-source Qwen3-Coder, especially its flagship version Qwen3-Coder-480B-A35B-Instruct. This model is an MoE (Mixture of Experts) model with 480B parameters and 35B effective parameters. It achieves open-source model SOTA in agentic programming, intelligent browser operation, and basic coding tasks. The article elaborates on the model's breakthroughs in data expansion during the pre-training phase (7.5T of high-quality code), context expansion (native 256K, up to 1M), and synthetic data optimization (using Qwen2.5-Coder to clean data). In the post-training phase, Scaling Code RL and Scaling Long-Horizon RL technologies are used to give the model 'agent thinking', enabling it to solve complex software engineering problems through multiple rounds of interaction and achieve high scores on SWE-Bench Verified. In addition, the article provides Qwen Code command-line tools and API call examples to facilitate developers' quick start. Overall, Qwen3-Coder demonstrates powerful code understanding and generation capabilities, aiming to advance intelligent programming.