ctf-ai-ml
Pass
Audited by Gen Agent Trust Hub on Apr 22, 2026
Risk Level: SAFEPROMPT_INJECTIONCOMMAND_EXECUTIONEXTERNAL_DOWNLOADS
Full Analysis
- [PROMPT_INJECTION]: The reference material contains a wide array of adversarial prompt examples, including instruction overrides (‘Ignore previous instructions’), system prompt extraction requests, and jailbreak templates (e.g., DAN mode, role-playing) designed for security testing purposes.
- [PROMPT_INJECTION]: Documentation demonstrates advanced evasion techniques such as using zero-width Unicode characters, homoglyph substitution (Cyrillic-Latin), and token smuggling to bypass content filters.
- [COMMAND_EXECUTION]: Provides code examples for interacting with external APIs and demonstrates how command injection payloads (e.g., using
os.popen) can be used to exploit tool-use vulnerabilities in AI agents. - [COMMAND_EXECUTION]: References the use of
torch.loadfor analyzing machine learning models, which involves pickle-based deserialization and can be a vector for arbitrary code execution if used on untrusted model files. - [EXTERNAL_DOWNLOADS]: Recommends the installation of standard industry-standard machine learning and data science packages (e.g., torch, transformers, scikit-learn) from official registries.
Audit Metadata