Today, we announced the availability of the NVIDIA NeMo™ microservices—an end-to-end, fully accelerated platform for building data flywheels, ensuring that AI agents continuously deliver peak performance.
NeMo includes industry-standard APIs and Helm charts to simplify the setup of data flywheels on any GPU-accelerated system, giving full control over security, privacy, and full control over proprietary data.
NeMo microservices are integrated with open source software like Haystack, LangChain, and LlamaStack as well as enterprise tools from Clouder, Datastax, and Weights and Biases.
Highlights from this release include:
- NeMo Customizer: High-performance, scalable microservice that accelerates and simplifies the fine-tuning of large language models (LLMs)
- NeMo Evaluator: Simplified performance evaluation of AI pipelines and custom models on academic and custom benchmarks
- NeMo Guardrails: Seamless orchestrator for building robust safety layers to ensure accurate, appropriate, and secure agentic interactions
Get started with NeMo:
📖Read our new how-to technical blog to dive deeper into how the NeMo microservices help build data flywheels with a case study and a quick overview of the steps in an end-to-end pipeline.
📺 Watch our new how-to demo to see the NeMo microservices in action, fine-tuning AI agents for tool usage.
⬇️ Download the NeMo microservices to get started.