Nvidia Developer 09月03日
NVIDIA Nemotron Super 49B v1.5:AI智能体的新标杆
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

NVIDIA推出了Nemotron Super 49B v1.5,一款在推理和智能体任务方面表现卓越的AI模型。该模型基于强大的开源模型,通过NVIDIA的合成数据集和先进技术进行增强,实现了更高的准确性、效率和透明度。Llama Nemotron Super v1.5在数学、科学、编码、指令遵循和聊天等核心能力上均有显著提升,并荣登Artificial Analysis Intelligence Index排行榜榜首。该模型优化了吞吐量,能够运行在单块NVIDIA H100 GPU上,降低了推理成本,为开发者构建高效的AI智能体系统提供了有力支持。此外,NVIDIA还发布了训练数据集,以提高模型训练的透明度。

🚀 **卓越的推理与智能体能力**:Llama Nemotron Super v1.5在数学、科学、编码、函数调用、指令遵循和聊天等一系列核心推理及智能体任务上取得了显著进步,其性能在Artificial Analysis Intelligence Index排行榜上名列前茅,展现了其在处理复杂任务时的强大实力。

💡 **高效的架构与优化**:该模型通过利用神经架构搜索(Neural Architecture Search)等后训练技术,显著提升了吞吐量表现,降低了推理成本。其高度优化的设计使其能够部署在单块NVIDIA H100 Tensor Core GPU上,为构建更经济高效的AI智能体系统提供了可能。

📁 **透明的训练数据与方法**:NVIDIA公开了用于训练Nemotron模型的合成数据集,该数据集包含超过2600万行高质量的函数调用、指令遵循、推理、聊天、数学和代码数据。这种透明度使用户能够对模型的训练过程有清晰的了解,并为开发者提供了构建自身高性能模型的基础。

🛠️ **先进的后训练技术**:Llama Nemotron Super v1.5采用了包括监督微调(SFT)、奖励感知偏好优化(RPO)、直接偏好优化(DPO)和可验证奖励强化学习(RLVR)在内的多种先进后训练流水线。这些技术确保了模型在各项能力上的针对性优化,进一步提升了推理准确性。

AI agents now solve multi-step problems, write production-level code, and act as general assistants across multiple domains. But to reach their full potential, the systems need advanced reasoning models without being prohibitively expensive. 

​​The NVIDIA Nemotron family builds on the strongest open models in the ecosystem by enhancing them with greater accuracy, efficiency, and transparency using NVIDIA open synthetic datasets, advanced techniques, and tools. This enables the creation of practical, right-sized, and high-performing AI agents. 

Llama Nemotron Super 49B v1.5, the latest version released Friday, brings significant improvements across core reasoning and agentic tasks like math, science, coding, function calling, instruction following, and chat, while maintaining strong throughput and compute efficiency.

It has now topped the Artificial Analysis Intelligence Index leaderboard.

In this blog post, we’ll cover the accuracy and inference performance of the latest NVIDIA Nemotron model, training methodology, data transparency, architectural optimizations, and deployment options. 

Llama Nemotron Super v1.5 tops Artificial Analysis leaderboard

The new model was built with the same methodology as the original Llama Nemotron Ultra v1, but has undergone further refinement and post-training by using additional high quality reasoning data. 


This model achieves best-in-class performance across a number of reasoning and agentic tasks, topping the Artificial Analysis Intelligence Index leaderboard, which measures accuracies across MMLU-Pro, GPQA Diamond, Humanity’s Last Exam, LiveCodeBench, SciCode, AIME, and MATH-500.

Figure 1. Artificial Analysis Intelligence Index Leaderboard

Evaluated by a third party on a suite of industry-standard benchmarks for reasoning, and instruction and function-calling tasks, Llama Nemotron Super v1.5 outpaces open models across advanced math, coding, reasoning, and chat metrics—firmly placing it as the top model in the 70-billion parameter range.

Beyond just being best-in-class in reasoning and agentic capabilities, the model also achieves significantly higher throughput by leveraging post-training methods to improve throughput performance (Neural Architecture Search).

Figure 3. Llama Nemotron Super v1.5 provides the highest accuracy and throughput for agentic tasks, lowering the cost of inference

The result is a highly performant model that fits on a single NVIDIA H100 Tensor Core GPU, letting developers build more effective and more efficient agentic systems.

Built for reasoning and agentic workloads

Building Llama Nemotron Super v1.5 required the combination of several key NVIDIA technologies:

Llama Nemotron post-training open dataset

This dataset was created entirely through synthetic data generation using advanced reasoning models like Qwen3 235B and DeepSeek R1 671B 0528. It allowed our team to create over 26 million rows of high-quality function calling, instruction following, reasoning, chat, math, and code data. 

Releasing the data allows us to be transparent about exactly what went into training our models, which can help developers and enterprises be confident in their selection of the Llama Nemotron Super v1.5 as the engine for their agentic systems. 

Beyond transparency, releasing the dataset allows developers to build their own models without expending the effort and time required to produce a high-quality dataset—thereby lowering the barrier to entry to produce highly-capable new models. 

This dataset is now available on Hugging Face, and the dataset card provides a more detailed breakdown.

Post-training process

As shared in this previous blog post, the post-training pipeline looks as follows:

Figure 4. Llama Nemotron Super v1.5 post-training pipeline

The team leveraged reinforcement learning to push the model to the limits and achieve the reasoning capabilities outlined above. The model underwent a number of post-training pipelines, all tailored for the desired capability enhancements. Aside from Supervised Fine-Tuning (SFT), the model also underwent: 

    RPO (Reward-aware Preference Optimization) – leveraging best-in-class NVIDIA Reward Models for chat capabilities DPO (Direct Preference Optimization) – for tool-calling capabilitiesRLVR (Reinforcement Learning with Verifiable Rewards) – for instruction-following, math, science, and more

The comprehensive post-training pipeline ensured that the model was ideally trained for each capability, further pushing the boundaries of reasoning accuracy. 

The teams also used NeMo Skills to evaluate and validate the model checkpoints, allowing for tight iteration and research cycles, as well as reproducibility. 

Llama Nemotron Super v1.5 available as a NIM

Llama Nemotron Super v1.5 will soon be available as an NVIDIA NIM microservice for rapid, reliable deployment on your preferred NVIDIA accelerated infrastructure. You can deploy it with a few simple commands and immediately integrate private, OpenAI API-compatible endpoints to level-up AI agents and reasoning apps. Plus, the high-performance Llama Nemotron Super inference autoscales on demand.

Get started with Llama Nemotron Super v1.5

The Llama Nemotron Super v1.5 model delivers powerful reasoning capabilities while remaining compute-efficient. It’s ready to power agentic applications from individual developers, all the way to huge enterprises. 

You can get started by trying out the model on build.nvidia.com. Once you’ve had some time to test the model, you can download the checkpoint from Hugging Face or follow the model card and run the model through the instructions provided there.

This post originally ran July 25. Updated July 29 with leaderboard information.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

NVIDIA Nemotron AI Agents Reasoning Models Large Language Models AI Machine Learning Llama Nemotron Super v1.5 Artificial Analysis Intelligence Index Synthetic Data AI Development
相关文章