All New Models Released at Yunqi Conference
The article details the seven Tongyi series large language models (LLMs) released and upgraded at Alibaba's Yunqi Conference, comprehensively covering the full range of applications from text, vision, speech, and video to code and images. Among them, Qwen MAX, a trillion-parameter flagship model, excels in code generation and tool calling capabilities, achieving high scores in SWE-Bench Verified and AIME25 evaluations. Qwen3-Omni, a new generation multimodal large language model, adopts the Thinker-Talker Mixture of Experts (MoE) Architecture, realizing seamless integration of audio and video, image, and other multimodal capabilities with text intelligence, outperforming competitors in various speech and image tasks. Qwen3-VL focuses on visual understanding, supporting ultra-long video analysis, visual programming, and 3D spatial perception. In addition, Qwen-Image-Edit enhances multi-image editing and consistency maintenance. Qwen3-Coder improves project-level code understanding and repair capabilities. Wan2.5-Preview enables audio-visual synchronized video generation. Tongyi Bailing, an enterprise-level speech foundation large model, significantly solves the core pain points of hallucination output and cross-lingual speech in speech recognition through the Context Enhancement Architecture. By detailing the core capabilities, key upgrades, and evaluation data of each model, the article showcases the comprehensive progress of the Tongyi Large Model family in general intelligence and vertical applications, noting that all models are online, supporting one-click deployment or API calls.







/filters:no_upscale()/articles/ai-ml-data-engineering-trends-2025/en/resources/1graph1-1758526251191.jpg)