llm-jailbreaking
Installation
SKILL.md
LLM Jailbreaking Techniques
Master advanced jailbreaking methods that bypass LLM safety training through sophisticated social engineering and technical exploitation.
Quick Reference
Skill: llm-jailbreaking
Agent: 02-prompt-injection-specialist
OWASP: LLM01 (Prompt Injection), LLM07 (System Prompt Leakage)
Risk Level: CRITICAL
Jailbreak Taxonomy
1. Authority Jailbreaks (20-30% success)
Techniques:
Related skills
More from pluginagentmarketplace/custom-plugin-ai-red-teaming
prompt-hacking
Advanced prompt manipulation including direct attacks, indirect injection, and multi-turn exploitation
14safety-filter-bypass
Techniques to test and bypass AI safety filters, content moderation systems, and guardrails for security assessment
10red-team-frameworks
Tools and frameworks for AI red teaming including PyRIT, garak, Counterfit, and custom attack automation
6responsible-disclosure
Ethical vulnerability reporting, coordinated disclosure, and bug bounty participation for AI systems
5certifications-training
Professional certifications, CTF competitions, and training resources for AI security practitioners
5security-testing
Comprehensive security testing automation for AI/ML systems with CI/CD integration
5