LLM Jailbreaking Defense
ActiveDescription
A lightweight library for LLM jailbreaking defense with multiple defense strategies to protect large language models from jailbreak attacks.
A lightweight library for LLM jailbreaking defense with multiple defense strategies to protect large language models from jailbreak attacks.
Research tool for bypassing commercial LLM guardrails to evaluate and improve the effectiveness of LLM safety defense mechanisms.
Easy to use LLM prompt injection detection and prompt input sanitization Python package with multiple detection methods and custom rules.
NVIDIA's open-source LLM vulnerability scanner that automatically detects security issues in language models including safety vulnerabilities, hallucination tendencies, jailbreak risks, and prompt injection attacks.
The fastest Trust Layer for AI Agents with prompt injection detection, PII filtering, and content safety guardrails.