Nvidia’s AI agent play is here with new models, orchestration blueprints

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More


The industry’s push into agentic AI continues, with Nvidia announcing several new services and models to facilitate the creation and deployment of AI agents. 

Today, Nvidia launched Nemotron, a family of models based on Meta’s Llama and trained on the company’s techniques and datasets. The company also announced new AI orchestration blueprints to guide AI agents. These latest releases bring Nvidia, a company more known for the hardware that powers the generative AI revolution, to the forefront of agentic AI development.

Nemotron comes in three sizes: Nano, Super and Ultra. It also comes in two flavors: the Llama Nemotron for language tasks and the Cosmos Nemotron vision model for physical AI projects. The Llama Nemotron Nano has 4B parameters, the Super 49B parameters and the Ultra 253B parameters. 

All three work best for agentic tasks including “instruction following, chat, function calling, coding and math,” according to the company.

Rev Lebaredian, VP of Omniverse and simulation technology at Nvidia, said in a briefing with reporters that the three sizes are optimized for different Nvidia computing resources. Nano is for cost-efficient low latency applications on PC and edge devices, Super is for high accuracy and throughput on a single GPU and Ultra is for highest accuracy at data center scale. 

“AI agents are the digital workforce that will work for us and work with us, and so the Nemotron model family is for agentic AI,” said Lebaredian. 

The Nemotron models are available as hosted APIs on Hugging Face and Nvidia’s website. Nvidia said enterprises can access the models through its AI Enterprise software platform. 

Nvidia is no stranger to foundation models. Last year, it quietly released a version of Nemotron, Llama-3.1-Nemotron-70B-Instruct, that outperformed similar models from OpenAI and Anthropic. It also unveiled NVLM 1.0, a family of multimodal language models. 

More support for agents

AI agents became a big trend in 2024 as enterprises began exploring how to deploy agentic systems in their workflow. Many believe that momentum will continue this year. 

Companies like Salesforce, ServiceNow, AWS and Microsoft have all called agents the next wave of gen AI in enterprises. AWS has added multi-agent orchestration to Bedrock, while Salesforce released its Agentforce 2.0, bringing more agents to its customers. 

However, agentic workflows still need other infrastructure to work efficiently. One such infrastructure revolves around orchestration, or managing multiple agents crossing different systems. 

Orchestration blueprints 

Nvidia has also entered the emerging field of AI orchestration with its blueprints that guide agents through specific tasks. 

The company has partnered with several orchestration companies, including LangChain, LlamaIndex, CrewAI, Daily and Weights and Biases, to build blueprints on Nvidia AI Enterprise. Each orchestration framework has developed its own blueprint with Nvidia. For example, CrewAI created a blueprint for code documentation to ensure code repositories are easy to navigate. LangChain added Nvidia NIM microservices to its structured report generation blueprint to help agents return internet searches in different formats. 

“Making multiple agents work together smoothly or orchestration is key to deploying agentic AI,” said Lebaredian. “These leading AI orchestration companies are integrating every Nvidia agentic building block, NIM, Nemo and Blueprints with their open-source agentic orchestration platforms.”

Nvidia’s new PDF-to-podcast blueprint aims to compete with Google’s NotebookLM by converting information from PDFs to audio. Another new blueprint will help build agents to search for and summarize videos. 

Lebaredian said Blueprints aims to help developers quickly deploy AI agents. To that end, Nvidia unveiled Nvidia Launchables, a platform that lets developers test, prototype and run blueprints in one click. 

Orchestration could be one of the bigger stories of 2025 as enterprises grapple with multi-agent production. 

Related Posts

TikTok is no longer available in the US

The switch has flipped on the TikTok ban. TikTok’s app stoped working and was removed from the App Store and Google Play on Saturday night, just hours before the January…

Read more

TikTok is down in the US

/ TikTok and CapCut’s apps have started to shut down in the US after telling users the ban would ‘force us to make our services temporarily unavailable.’ p>span:first-child]:text-gray-13 [&_.duet–article-byline-and]:text-gray-13″> By…

Read more

3 Ways To Password Protect Your USB Drive (And Why You Should)

monte_a/Shutterstock USB drives offer unmatched convenience and portability, making them a go-to solution for transferring files between computers, storing backups, or carrying sensitive information on the go. However, this portability…

Read more

TikTok ban: all the news on the app’s shutdown in the US

President Joe Biden signed a bill last year that would ban TikTok, the popular video app owned by Chinese company ByteDance if the company didn’t sell the platform within a…

Read more

TikTok says it’s going offline in the US

/ The app failed to avoid a US federal ban that is now going into effect. p>span:first-child]:text-gray-13 [&_.duet–article-byline-and]:text-gray-13″> By Alex Heath, a deputy editor and author of the Command Line…

Read more

What Does ‘Roger’ Mean In Military Aviation And Why Do Pilots Say It?

Jeremy Poland/Getty If you’ve ever seen shows with military pilots, you’ve likely heard them saying “Roger” a few times when talking to air traffic control, fellow pilots, and other people…

Read more

Leave a Reply