nvidia llama nemotron 1742453796860.jpg

Nvidia releases llama nemotron ai reasoning models for agentic workflows

Nvidia released a new family of artificial intelligence (ai) models on tuesday at its gpu technology conference (GTC) 2025. Dubbed llama nemotron, these are the company Reasoning-Focused Large Language Models (llms) that are designed to offer a foundation for agentic ai workflows. The Santa Clara-Based Tech Giant Said these models were aimed at developers and enterprises to enable them to make advanced ai agents that can eather work independent oor as connected teams to perform Tasks. The llama nemotron models are currently available via nvidia’s platform and hugging face.

Nvidia introduces new reasoning-focused ai models

In a newsroom postThe tech giant detailed the new ai models. The llama nemotron reasoning models are based on meta’s llama 3 series models, with post-training enhancements added by nvidia. The company highlighted that family of ai models display improved capabilitys in multistep mathematics, coding, reasons, and complex decision-making.

The company highlighted that the process improved the according to the models by up to 20 percent compared to the based models. The infection speed is also said to have been improved by five times compared to similar-sized open-source reasoning models. Nvidia claimed that “The models can handle more complex reasoning tasks, enhance decision-making capabilities, and Reduce Operational Costs for Enterprises.” With these advancements, the llm can be used to build and power ai agents.

LLAma Nemotron Reasoning Models are available in three parameter sizes – Nano, Super, and Ultra. The nano model is best suited for on-device and edge-based tasks that require high accuracy. The super variant is placed in the middle to offer high accuracy and throughput on a single gpu. Finally, the ultra model is meant to be run on Multi-GPU Servers and offers agentic accuracy.

The post-training of the reasoning models was done on the nvidia dgx cloud using synthetic data generated using the nemotron platform as well as other models. The tech giant is also making the tools, datasets, and post-training optimization techniques used to develop to develop the llama nemotron models available to the open-source communication.

Nvidia is also working with enterprise partners to brings the models to developers and businesses. These reasons models and the nim microservices can be accessed via microsoft’s azure ai foundry as well as an option via the azure ai agent services. Sap is also using the models for its business ai solutions and the AI ​​Copilot Dubbed Joule, the company said. Other Enterprises Using Lalama Nemotron Models Include Servicenow, Accenture, and Deloitte.

The llama nemotron nano and super models and nim microservices are available for businesses and developers as an application programming interface (api) via nvidia’s’ platform as well as its hugging face ListingIt is available with the permissive nvidia open model license agrement which allows both research and commercial usage.

Similar Posts