Natural language processing

Nvidia’s new ‘Blueprints’ promise plug-and-play AI agents

Nvidia’s new ‘Blueprints’ promise plug-and-play AI agents



summary
Summary

Nvidia has unveiled a new approach to building AI agents called “blueprints” – essentially pre-built templates that give developers everything they need to create autonomous AI systems without starting from scratch.

These blueprints come packaged as “launchables” – pre-configured development environments that teams can spin up instantly. The goal is simple: help developers build AI systems that can perform complex tasks on their own, without reinventing the wheel each time.

Five AI companies have already jumped on board with their own specialized blueprints. CrewAI created one for automated code documentation, Daily focused on language AI agents, LangChain developed templates for structured reports, LlamaIndex designed research tools for blog posts, and Weights & Biases contributed blueprints for virtual AI assistants.

Nvidia has added two blueprints of its own: one that turns PDFs into podcasts and another for video analysis. According to the company, the video blueprint can analyze footage 30 times faster than real-time playback.

Ad

There’s a catch, though: developers need Nvidia AI Enterprise Software to use these blueprints. The software works with data centers from Dell, HPE, Lenovo, and Supermicro, plus cloud services from AWS, Google, Microsoft, and Oracle.

New Nemotron models based on Llama and optimized for “agentic” AI

Under the hood, these blueprints run on Nvidia’s new Nemotron models, built on Meta’s Llama technology. The family includes two main members: Llama Nemotron for text processing and Cosmos Nemotron for image analysis.

Nvidia says they’ve specifically optimized these Llama-based models for AI agents, enhancing their ability to follow instructions, engage in conversation, make function calls, and handle programming and math tasks.

Infographic: Comparison of NVIDIA Llama and Cosmos Nemotron models in three sizes (Nano, Super, Ultra) with respective parameter numbers.
The Llama-based Nemoton models are available in three sizes to meet different needs: Nano for real-time PC applications, Super for high-accuracy tasks on single GPUs, and Ultra for data center use. | Image: Nvidia

For developers who want to try it out, Nvidia offers two paths: members of the Nvidia Developer Program will soon be able to access the models through hosted APIs, or they can download them directly from build.nvidia.com and Hugging Face for “testing and research.”

Business users can get the complete package – including NeMo, NeMo Retriever, and both Nemotron models as NIM microservices – through Nvidia’s AI Enterprise platform mentioned earlier.

Recommendation

Nvidia's new 'Blueprints' promise plug-and-play AI agents

Source link