PANews reported on September 12th that Alibaba's Tongyi Qianwen released its next-generation basic model architecture, Qwen3-Next, and open-sourced the Qwen3-Next-80B-A3B series of models based on this architecture. Compared to the Qwen3 MoE model architecture, this architecture features the following core improvements: a hybrid attention mechanism, a highly sparse MoE structure, a series of optimizations for stable and user-friendly training, and a multi-token prediction mechanism to improve inference efficiency. Based on the Qwen3-Next model architecture, Alibaba trained the Qwen3-Next-80B-A3B-Base model, which has 80 billion parameters but only activates 3 billion. This Base model achieves performance similar to or slightly better than the Qwen3-32B dense model, while its training cost (GPU hours) is less than one-tenth of that of the Qwen3-32B. Its inference throughput for contexts above 32k is over ten times that of the Qwen3-32B, achieving exceptional cost-effectiveness for both training and inference.PANews reported on September 12th that Alibaba's Tongyi Qianwen released its next-generation basic model architecture, Qwen3-Next, and open-sourced the Qwen3-Next-80B-A3B series of models based on this architecture. Compared to the Qwen3 MoE model architecture, this architecture features the following core improvements: a hybrid attention mechanism, a highly sparse MoE structure, a series of optimizations for stable and user-friendly training, and a multi-token prediction mechanism to improve inference efficiency. Based on the Qwen3-Next model architecture, Alibaba trained the Qwen3-Next-80B-A3B-Base model, which has 80 billion parameters but only activates 3 billion. This Base model achieves performance similar to or slightly better than the Qwen3-32B dense model, while its training cost (GPU hours) is less than one-tenth of that of the Qwen3-32B. Its inference throughput for contexts above 32k is over ten times that of the Qwen3-32B, achieving exceptional cost-effectiveness for both training and inference.

Alibaba launches more efficient Qwen3-Next artificial intelligence model

2025/09/12 07:27

PANews reported on September 12th that Alibaba's Tongyi Qianwen released its next-generation basic model architecture, Qwen3-Next, and open-sourced the Qwen3-Next-80B-A3B series of models based on this architecture. Compared to the Qwen3 MoE model architecture, this architecture features the following core improvements: a hybrid attention mechanism, a highly sparse MoE structure, a series of optimizations for stable and user-friendly training, and a multi-token prediction mechanism to improve inference efficiency. Based on the Qwen3-Next model architecture, Alibaba trained the Qwen3-Next-80B-A3B-Base model, which has 80 billion parameters but only activates 3 billion. This Base model achieves performance similar to or slightly better than the Qwen3-32B dense model, while its training cost (GPU hours) is less than one-tenth of that of the Qwen3-32B. Its inference throughput for contexts above 32k is over ten times that of the Qwen3-32B, achieving exceptional cost-effectiveness for both training and inference.

Market Opportunity
Moonveil Logo
Moonveil Price(MORE)
$0.002159
$0.002159$0.002159
-10.56%
USD
Moonveil (MORE) Live Price Chart
Disclaimer: The articles reposted on this site are sourced from public platforms and are provided for informational purposes only. They do not necessarily reflect the views of MEXC. All rights remain with the original authors. If you believe any content infringes on third-party rights, please contact service@support.mexc.com for removal. MEXC makes no guarantees regarding the accuracy, completeness, or timeliness of the content and is not responsible for any actions taken based on the information provided. The content does not constitute financial, legal, or other professional advice, nor should it be considered a recommendation or endorsement by MEXC.