Ad
Skip to content

Nvidia's new 'Blueprints' promise plug-and-play AI agents

Image description
Nvidia

Nvidia has unveiled a new approach to building AI agents called "blueprints" - essentially pre-built templates that give developers everything they need to create autonomous AI systems without starting from scratch.

These blueprints come packaged as "launchables" - pre-configured development environments that teams can spin up instantly. The goal is simple: help developers build AI systems that can perform complex tasks on their own, without reinventing the wheel each time.

Five AI companies have already jumped on board with their own specialized blueprints. CrewAI created one for automated code documentation, Daily focused on language AI agents, LangChain developed templates for structured reports, LlamaIndex designed research tools for blog posts, and Weights & Biases contributed blueprints for virtual AI assistants.

Nvidia has added two blueprints of its own: one that turns PDFs into podcasts and another for video analysis. According to the company, the video blueprint can analyze footage 30 times faster than real-time playback.

Ad
DEC_D_Incontent-1

There's a catch, though: developers need Nvidia AI Enterprise Software to use these blueprints. The software works with data centers from Dell, HPE, Lenovo, and Supermicro, plus cloud services from AWS, Google, Microsoft, and Oracle.

New Nemotron models based on Llama and optimized for "agentic" AI

Under the hood, these blueprints run on Nvidia's new Nemotron models, built on Meta's Llama technology. The family includes two main members: Llama Nemotron for text processing and Cosmos Nemotron for image analysis.

Nvidia says they've specifically optimized these Llama-based models for AI agents, enhancing their ability to follow instructions, engage in conversation, make function calls, and handle programming and math tasks.

Infographic: Comparison of NVIDIA Llama and Cosmos Nemotron models in three sizes (Nano, Super, Ultra) with respective parameter numbers.
The Llama-based Nemoton models are available in three sizes to meet different needs: Nano for real-time PC applications, Super for high-accuracy tasks on single GPUs, and Ultra for data center use. | Image: Nvidia

For developers who want to try it out, Nvidia offers two paths: members of the Nvidia Developer Program will soon be able to access the models through hosted APIs, or they can download them directly from build.nvidia.com and Hugging Face for "testing and research."

Ad
DEC_D_Incontent-2

Business users can get the complete package - including NeMo, NeMo Retriever, and both Nemotron models as NIM microservices - through Nvidia's AI Enterprise platform mentioned earlier.

AI News Without the Hype – Curated by Humans

As a THE DECODER subscriber, you get ad-free reading, our weekly AI newsletter, the exclusive "AI Radar" Frontier Report 6× per year, access to comments, and our complete archive.

AI news without the hype
Curated by humans.

  • Over 20 percent launch discount.
  • Read without distractions – no Google ads.
  • Access to comments and community discussions.
  • Weekly AI newsletter.
  • 6 times a year: “AI Radar” – deep dives on key AI topics.
  • Up to 25 % off on KI Pro online events.
  • Access to our full ten-year archive.
  • Get the latest AI news from The Decoder.
Subscribe to The Decoder