AgentList
HomeProjectsArticlesAbout
Explore Projects
HomeProjectsArticlesAbout
Explore Projects
Home / Projects / LLM Jailbreaking Defense

LLM Jailbreaking Defense

Active
GitHub Python MIT

Description

A lightweight library for LLM jailbreaking defense with multiple defense strategies to protect large language models from jailbreak attacks.

Tags

jailbreak-defense llm-security safety python

Categories

🛡️ Security & Guardrails
Visit GitHub

Project Metrics

Stars 60
Forks 0
Watchers 0
Issues 0
Created January 1, 2025
Last commit April 9, 2026

Deployment

Local

Related Projects

CKA-Agent

197 · Python
Active

Research tool for bypassing commercial LLM guardrails to evaluate and improve the effectiveness of LLM safety defense mechanisms.

llm-securityguardrails-testingred-teaming +2

Pytector

40 · Python
Active

Easy to use LLM prompt injection detection and prompt input sanitization Python package with multiple detection methods and custom rules.

prompt-injectiondetectionsanitization +2

Garak

7.6k · HTML
Active

NVIDIA's open-source LLM vulnerability scanner that automatically detects security issues in language models including safety vulnerabilities, hallucination tendencies, jailbreak risks, and prompt injection attacks.

llm-securityvulnerability-scannerllm-evaluation +2

ZenGuard AI

150 · Python
Active

The fastest Trust Layer for AI Agents with prompt injection detection, PII filtering, and content safety guardrails.

llm-securityguardrailsprompt-injection +2
AgentList

Curated directory of open-source AI agent projects

Quick Links

  • Project List
  • Featured Articles
  • Browse Categories

Contact

  • About
  • Privacy Policy
  • Contact Us

© 2026 AgentList. All rights reserved.

Made with for the open source community