×
Nvidia launches AI safety microservices for autonomous agents
Written by
Published on
Join our daily newsletter for breaking news, product launches and deals, research breakdowns, and other industry-leading AI coverage
Join Now

Nvidia has introduced three new microservices designed to enhance safety and control for AI agents, addressing key concerns around enterprise adoption of autonomous AI systems.

The core announcement: Nvidia has expanded its NeMo Guardrails software toolkit with new inference microservices (NIMs) that leverage small language models to improve AI agent security and compliance.

  • The three new microservices focus on content safety, topic control, and jailbreak detection
  • These tools are designed to help organizations maintain control over AI agents while ensuring fast, responsive performance
  • According to Nvidia, 10% of organizations currently use AI agents, with 80% planning adoption within three years

Technical specifications: The microservices utilize small language models (SLMs), which offer lower latency compared to larger language models and can operate effectively in environments with limited resources.

  • The content safety NIM uses the Aegis Content Safety Data Set, containing 35,000 human-annotated samples, to prevent harmful or biased AI outputs
  • The topic control NIM keeps AI agents focused on approved subjects and prevents unwanted content discussion
  • The jailbreak detection NIM, built on Nvidia’s Garak toolkit, uses 17,000 known jailbreak examples to protect against security circumvention attempts

Practical applications: These guardrails enable organizations to implement AI agents while maintaining strict control over their behavior and outputs.

  • Automotive manufacturers can create AI agents for vehicle operations while preventing discussion of competitor brands
  • Healthcare, manufacturing, and other regulated industries can deploy AI agents while ensuring compliance with industry-specific requirements
  • Organizations can customize guardrails based on their unique needs, policies, and geographic regulations

Implementation framework: The NeMo platform provides a comprehensive system for managing AI agent policies and behavior.

  • The platform allows for both default configurations and extensive customization options
  • Multiple guardrails can be implemented simultaneously to address various security and compliance requirements
  • IT departments will take on new responsibilities as “HR for agents,” managing AI behavior and compliance

Looking ahead: While these guardrails address current enterprise concerns about AI agent deployment, their effectiveness will ultimately depend on how well organizations can customize and implement them to match their specific use cases and regulatory requirements. The growing adoption of AI agents suggests that tools like these will become increasingly crucial for maintaining control and safety in autonomous AI systems.

Nvidia intros new guardrail microservices for agentic AI

Recent News

Apple’s cheapest iPad is bad for AI

Apple's budget tablet lacks sufficient RAM to run upcoming AI features, widening the gap with pricier models in the lineup.

Mira Murati’s AI venture recruits ex-OpenAI leader among first hires

Former OpenAI exec's new AI startup lures top talent and seeks $100 million in early funding.

Microsoft is cracking down on malicious actors who bypass Copilot’s safeguards

Tech giant targets cybercriminals who created and sold tools to bypass AI security measures and generate harmful content.