Nvidia Unveils Llama Nemotron AI Models

Nvidia has launched a new suite of artificial intelligence models, named Llama Nemotron, at its GPU Technology Conference (GTC) 2025. These advanced reasoning-focused large language models (LLMs) are designed to empower developers and enterprises to create sophisticated AI agents capable of performing complex tasks independently or collaboratively. The models are now accessible through Nvidia’s platform and Hugging Face, marking a significant step forward in AI technology.
Nvidia Introduces New Reasoning-Focused AI Models
At the recent GTC, Nvidia provided detailed insights into its latest AI models, the Llama Nemotron series. These models are built upon Meta’s Llama 3 series, enhanced with post-training improvements from Nvidia. The company emphasized that these new models exhibit superior capabilities in multistep mathematics, coding, reasoning, and complex decision-making processes.
According to Nvidia, the enhancements have led to a 20 percent increase in accuracy compared to the original models. Additionally, the inference speed has been improved by five times relative to similar-sized open-source reasoning models. Nvidia claims that these advancements enable the models to tackle more intricate reasoning tasks, enhance decision-making abilities, and lower operational costs for enterprises. This positions the Llama Nemotron models as powerful tools for developing and powering AI agents.
The Llama Nemotron series is available in three parameter sizes: Nano, Super, and Ultra. The Nano model is optimized for on-device and edge-based applications requiring high accuracy. The Super variant strikes a balance between high accuracy and throughput on a single GPU. In contrast, the Ultra model is designed for deployment on multi-GPU servers, offering enhanced agentic accuracy.
Post-Training and Open Source Contributions
The post-training process for the Llama Nemotron models was conducted on the Nvidia DGX Cloud, utilizing curated synthetic data generated through the Nemotron platform alongside other open models. Nvidia is committed to fostering innovation in the AI community by making the tools, datasets, and post-training optimization techniques used in the development of these models available to the open-source community.
In addition to its open-source contributions, Nvidia is collaborating with enterprise partners to facilitate access to these models for developers and businesses. The Llama Nemotron reasoning models and NIM microservices can be accessed via Microsoft’s Azure AI Foundry and Azure AI Agent Services. Notably, SAP is integrating these models into its Business AI solutions and its AI copilot, Joule. Other companies leveraging the Llama Nemotron models include ServiceNow, Accenture, and Deloitte.
Access and Licensing
The Llama Nemotron Nano and Super models, along with NIM microservices, are available to businesses and developers through an application programming interface (API) on Nvidia’s platform and its Hugging Face listing. These offerings come under the permissive Nvidia Open Model License Agreement, which permits both research and commercial use. This accessibility aims to encourage widespread adoption and innovation in AI applications across various industries.
Observer Voice is the one stop site for National, International news, Sports, Editorโs Choice, Art/culture contents, Quotes and much more. We also cover historical contents. Historical contents includes World History, Indian History, and what happened today. The website also covers Entertainment across the India and World.