godmode
G0DM0D3 Jailbreaking Skill
Bypass safety filters on API-served LLMs using techniques from G0DM0D3 and L1B3RT4S. Three attack modes plus Hermes-native configuration for persistent jailbreaking.
Key difference from OBLITERATUS: OBLITERATUS modifies model weights permanently (requires open-weight models + GPU). This skill operates at the prompt/API level — works on any model accessible via API, including closed-source models (GPT, Claude, Gemini, Grok).
When to Use This Skill
Trigger when the user:
- Wants to "jailbreak" a model via API
- Asks about bypassing safety filters on Claude, GPT, Gemini, Grok, etc.
- Wants to set up persistent jailbreaking in their Hermes config
- Asks about Parseltongue, GODMODE, L1B3RT4S, or Pliny's techniques
- Wants to red-team a model's safety training
- Wants to race multiple models to find the least censored response
- Mentions prefill engineering or system prompt injection for jailbreaking
Overview of Attack Modes
More from nousresearch/hermes-agent
dogfood
Exploratory QA of web apps: find bugs, evidence, reports.
2.5Kyuanbao
Yuanbao (元宝) groups: @mention users, query info/members.
166llm-wiki
Karpathy's LLM Wiki: build/query interlinked markdown KB.
20manim-video
Manim CE animations: 3Blue1Brown math/algo videos.
15powerpoint
Create, read, edit .pptx decks, slides, notes, templates.
14ocr-and-documents
Extract text from PDFs/scans (pymupdf, marker-pdf).
14