prompt-test

Installation
SKILL.md

/dm:prompt-test

Purpose

A/B test content output variations by comparing quality scores across different prompt approaches, headline styles, CTA phrasing, or complete content strategy variations. Create named tests, log variants with their evaluation scores, and determine which approach produces the best quality results.

This command brings experimental rigor to content creation. Instead of guessing which headline style, subject line approach, or content structure works best, you run a structured test: define the experiment, log each variant with its quality scores, and get a statistically grounded recommendation on which approach to adopt. Useful for testing subject line styles (curiosity vs. benefit-driven), headline approaches (question vs. statement vs. how-to), CTA phrasing (urgency vs. value vs. social proof), tone variations (formal vs. conversational), or complete content strategy A/B comparisons.

Input Required

The user must provide (or will be prompted for):

  • Action: What to do — create (set up a new test), log (add a variant to an existing test), results (get comparison and winner), or list (show all tests)
  • Test name: A descriptive name for the experiment (e.g., "Q1 email subject line style", "homepage headline approach") — required for create, log, and results
  • Variant label: Identifier for this variant (e.g., "A", "B", "C", "control", "curiosity-driven", "benefit-led") — required for log
  • Content for the variant: The actual content to evaluate — text inline, file path, or pasted content block — required for log
  • Variant description: Brief explanation of the approach or strategy this variant represents (e.g., "Uses curiosity gap with no product mention", "Leads with quantified benefit") — required for log
  • Content type: The type of content being tested (email subject line, headline, ad copy, CTA, full article, etc.) — optional, applied during evaluation for dimension weighting
  • Evidence file: Supporting data or research that informs the test hypothesis — optional, passed to evaluation for context
Related skills
Installs
31
GitHub Stars
100
First Seen
Feb 27, 2026