How Alibaba builds its most efficient AI model to date

2025-09-14 01:00:20 英文原文

A technical innovation has allowed Alibaba Group Holding, one of the leading players in China’s artificial intelligence boom, to develop a new generation of foundation models that match the strong performance of larger predecessors while being significantly smaller and more cost efficient.

Alibaba Cloud, the AI and cloud computing division of Alibaba, unveiled on Friday a new generation of large language models that it said heralded “the future of efficient LLMs”. The new models are nearly 13 times smaller than the company’s largest AI model, released just a week earlier.

Despite its compact size, Qwen3-Next-80B-A3B is among Alibaba’s best models to date, according to developers. The key lies in its efficiency: the model is said to perform 10 times faster in some tasks than the preceding Qwen3-32B released in April, while achieving a 90 per cent reduction in training costs.

Emad Mostaque, co-founder of the UK-based start-up Stability AI, said on X that Alibaba’s new model outperformed “pretty much any model from last year” despite an estimated training cost of less than US$500,000.

For comparison, training Google’s Gemini Ultra, released in February 2024, cost an estimated US$191 million, according to Stanford University’s AI Index.

Alibaba says its new generation of AI foundation models heralds the “the future of efficient LLMs”. Photo: Handout

Alibaba says its new generation of AI foundation models heralds the “the future of efficient LLMs”. Photo: Handout

Artificial Analysis, a leading AI benchmarking firm, said Qwen3-Next-80B-A3B surpassed the latest versions of both DeepSeek R1 and Alibaba-backed start-up Moonshot AI’s Kimi-K2. Alibaba owns the South China Morning Post.

关于《How Alibaba builds its most efficient AI model to date》的评论


暂无评论

发表评论

摘要

Alibaba Group has developed a new generation of smaller, more efficient foundation models for artificial intelligence, matching the performance of larger predecessors at significantly reduced costs. Alibaba Cloud introduced Qwen3-Next-80B-A3B, which is nearly 13 times smaller than its largest AI model but performs tasks up to ten times faster and reduces training costs by 90%. The new model outperforms many previous models while costing less than $500,000 to train, according to Stability AI's Emad Mostaque. According to Artificial Analysis, Qwen3-Next-80B-A3B surpasses the latest versions of DeepSeek R1 and Moonshot AI’s Kimi-K2.