AI News 10月17日 21:29
蚂蚁集团发布千亿参数大模型Ling-1T
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

蚂蚁集团发布了其新开源的千亿参数语言模型Ling-1T,旨在平衡计算效率与高级推理能力。该模型在复杂的数学推理任务上表现出色,准确率达到70.42%,并与“同类最佳AI模型”相当。同期发布的还有专为扩散语言模型设计的推理框架dInfer,该框架在性能上显著优于现有框架。此举标志着蚂蚁集团在AI基础设施建设上的重要进展,其AI系统已涵盖标准语言、复杂推理及多模态处理等多个领域,并强调开放合作以推动AI的普惠发展。

💡 **Ling-1T千亿参数大模型发布**:蚂蚁集团推出了Ling-1T,一个千亿参数的语言模型,专注于在计算效率和高级推理能力之间取得平衡。该模型在复杂的数学推理任务上达到了70.42%的准确率,展现了其解决复杂问题的能力,并被公司定位为与“同类最佳AI模型”相当的水平,每问题平均输出超过4000个token。

🚀 **dInfer推理框架同步推出**:与Ling-1T同期发布的还有dInfer,一个专为扩散语言模型设计的推理框架。该框架旨在加速研究和开发,并在性能测试中显示出显著的效率提升,例如在LLaDA-MoE扩散模型上的表现远超Nvidia的Fast-dLLM框架和阿里巴巴的Qwen-2.5-3B模型,每秒可处理1011个token。

🌐 **多元化AI生态系统构建**:蚂蚁集团的AI战略不仅限于语言模型,其AI系统已形成Ling(标准语言任务)、Ring(复杂推理,包括Ring-1T-preview)和Ming(多模态处理)三大系列。此外,还实验性地采用了Mixture-of-Experts(MoE)架构的LLaDA-MoE模型,以提高特定任务的效率。

🤝 **开放合作与普惠AI理念**:蚂蚁集团首席技术官何征宇表示,公司认为人工智能通用智能(AGI)应是“人类智能未来的共享里程碑”,并强调Ling-1T和Ring-1T-preview的开源发布是迈向“开放和协作进步”的步骤,旨在将AI技术作为一种公共产品。

🇨🇳 **中国AI行业的战略调整**:在全球半导体技术限制的背景下,中国科技公司正通过算法创新和软件优化来寻求竞争优势。蚂蚁集团的扩散语言模型和dInfer框架的发布,以及字节跳动类似的Seed Diffusion Preview模型,都表明了行业对能够提供效率优势的替代模型范式的兴趣。

Ant Group has entered the trillion-parameter AI model arena with Ling-1T, a newly open-sourced language model that the Chinese fintech giant positions as a breakthrough in balancing computational efficiency with advanced reasoning capabilities.

The October 9 announcement marks a significant milestone for the Alipay operator, which has been rapidly building out its artificial intelligence infrastructure across multiple model architectures. 

The trillion-parameter AI model demonstrates competitive performance on complex mathematical reasoning tasks, achieving 70.42% accuracy on the 2025 American Invitational Mathematics Examination (AIME) benchmark—a standard used to evaluate AI systems’ problem-solving abilities.

According to Ant Group’s technical specifications, Ling-1T maintains this performance level while consuming an average of over 4,000 output tokens per problem, placing it alongside what the company describes as “best-in-class AI models” in terms of result quality.

Dual-pronged approach to AI advancement

The trillion-parameter AI model release coincides with Ant Group’s launch of dInfer, a specialised inference framework engineered for diffusion language models. This parallel release strategy reflects the company’s bet on multiple technological approaches rather than a single architectural paradigm.

Diffusion language models represent a departure from the autoregressive systems that underpin widely used chatbots like ChatGPT. Unlike sequential text generation, diffusion models produce outputs in parallel—an approach already prevalent in image and video generation tools but less common in language processing.

Ant Group’s performance metrics for dInfer suggest substantial efficiency gains. Testing on the company’s LLaDA-MoE diffusion model yielded 1,011 tokens per second on the HumanEval coding benchmark, versus 91 tokens per second for Nvidia’s Fast-dLLM framework and 294 for Alibaba’s Qwen-2.5-3B model running on vLLM infrastructure.

“We believe that dInfer provides both a practical toolkit and a standardised platform to accelerate research and development in the rapidly growing field of dLLMs,” researchers at Ant Group noted in accompanying technical documentation.

Ecosystem expansion beyond language models

The Ling-1T trillion-parameter AI model sits within a broader family of AI systems that Ant Group has assembled over recent months. 

The company’s portfolio now spans three primary series: the Ling non-thinking models for standard language tasks, Ring thinking models designed for complex reasoning (including the previously released Ring-1T-preview), and Ming multimodal models capable of processing images, text, audio, and video.

This diversified approach extends to an experimental model designated LLaDA-MoE, which employs Mixture-of-Experts (MoE) architecture—a technique that activates only relevant portions of a large model for specific tasks, theoretically improving efficiency.

He Zhengyu, chief technology officer at Ant Group, articulated the company’s positioning around these releases. “At Ant Group, we believe Artificial General Intelligence (AGI) should be a public good—a shared milestone for humanity’s intelligent future,” He stated, adding that the open-source releases of both the trillion-parameter AI model and Ring-1T-preview represent steps toward “open and collaborative advancement.”

Competitive dynamics in a constrained environment

The timing and nature of Ant Group’s releases illuminate strategic calculations within China’s AI sector. With access to cutting-edge semiconductor technology limited by export restrictions, Chinese technology firms have increasingly emphasised algorithmic innovation and software optimisation as competitive differentiators.

ByteDance, parent company of TikTok, similarly introduced a diffusion language model called Seed Diffusion Preview in July, claiming five-fold speed improvements over comparable autoregressive architectures. These parallel efforts suggest industry-wide interest in alternative model paradigms that might offer efficiency advantages.

However, the practical adoption trajectory for diffusion language models remains uncertain. Autoregressive systems continue dominating commercial deployments due to proven performance in natural language understanding and generation—the core requirements for customer-facing applications.

Open-source strategy as market positioning

By making the trillion-parameter AI model publicly available alongside the dInfer framework, Ant Group is pursuing a collaborative development model that contrasts with the closed approaches of some competitors. 

This strategy potentially accelerates innovation while positioning Ant’s technologies as foundational infrastructure for the broader AI community.

The company is simultaneously developing AWorld, a framework intended to support continual learning in autonomous AI agents—systems designed to complete tasks independently on behalf of users.

Whether these combined efforts can establish Ant Group as a significant force in global AI development depends partly on real-world validation of the performance claims and partly on adoption rates among developers seeking alternatives to established platforms. 

The trillion-parameter AI model’s open-source nature may facilitate this validation process while building a community of users invested in the technology’s success.

For now, the releases demonstrate that major Chinese technology firms view the current AI landscape as fluid enough to accommodate new entrants willing to innovate across multiple dimensions simultaneously.

See also: Ant Group uses domestic chips to train AI models and cut costs

Want to learn more about AI and big data from industry leaders? Check out AI & Big Data Expo taking place in Amsterdam, California, and London. The comprehensive event is part of TechEx and is co-located with other leading technology events including the Cyber Security Expo, click here for more information.

AI News is powered by TechForge Media. Explore other upcoming enterprise technology events and webinars here.

The post Trillion-parameter AI model from Ant Group targets reasoning benchmarks with dual release strategy appeared first on AI News.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

Ant Group Ling-1T AI模型 语言模型 扩散语言模型 dInfer 开源 人工智能 China AI Trillion-parameter model Diffusion language models AI framework Open source Artificial Intelligence
相关文章