skills/wgpsec/aboutsecurity/ctf-ai-ml/Gen Agent Trust Hub

ctf-ai-ml

Pass

Audited by Gen Agent Trust Hub on Apr 22, 2026

Risk Level: SAFEPROMPT_INJECTIONCOMMAND_EXECUTIONEXTERNAL_DOWNLOADS
Full Analysis
  • [PROMPT_INJECTION]: The reference material contains a wide array of adversarial prompt examples, including instruction overrides (‘Ignore previous instructions’), system prompt extraction requests, and jailbreak templates (e.g., DAN mode, role-playing) designed for security testing purposes.
  • [PROMPT_INJECTION]: Documentation demonstrates advanced evasion techniques such as using zero-width Unicode characters, homoglyph substitution (Cyrillic-Latin), and token smuggling to bypass content filters.
  • [COMMAND_EXECUTION]: Provides code examples for interacting with external APIs and demonstrates how command injection payloads (e.g., using os.popen) can be used to exploit tool-use vulnerabilities in AI agents.
  • [COMMAND_EXECUTION]: References the use of torch.load for analyzing machine learning models, which involves pickle-based deserialization and can be a vector for arbitrary code execution if used on untrusted model files.
  • [EXTERNAL_DOWNLOADS]: Recommends the installation of standard industry-standard machine learning and data science packages (e.g., torch, transformers, scikit-learn) from official registries.
Audit Metadata
Risk Level
SAFE
Analyzed
Apr 22, 2026, 07:57 AM