AutoHarness
NormalDescription
Automated harness engineering for AI agents. Auto-generates test harnesses to evaluate agent safety and reliability across different scenarios.
Automated harness engineering for AI agents. Auto-generates test harnesses to evaluate agent safety and reliability across different scenarios.
An open-source evaluation and testing library for LLM agents providing automated model scanning, bias detection, performance benchmarking, and compliance checks.
NVIDIA NeMo Guardrails is an open-source toolkit for adding programmable guardrails to LLM-based conversational systems, supporting topic control, safety enforcement, and dialog guidance.
NVIDIA's open-source LLM vulnerability scanner that automatically detects security issues in language models including safety vulnerabilities, hallucination tendencies, jailbreak risks, and prompt injection attacks.
The security toolkit for LLM interactions, providing prompt injection detection, PII anonymization, content safety auditing, and more to secure production LLM deployments.