apastra-eval
Apastra Eval
Run prompt evaluations locally. Your IDE agent is the harness — no external tools, APIs, or CI needed.
When to Use
Use this skill when you want to:
- Evaluate a prompt against test cases
- Run a quick eval file (single-file prompt + cases + assertions)
- Compare results against a baseline to detect regressions
- Get a scorecard with metrics for a prompt change
How Evaluation Works
More from bintzgavin/apastra
apastra
PromptOps skills for versioning, evaluating, and shipping AI prompts as disciplined software assets. Agent-as-harness — your IDE agent runs evals, compares baselines, and gates quality.
10apastra-validate
Validate all promptops files against JSON schemas. Catch formatting errors before running evaluations.
5apastra-scaffold
Generate new prompt specs, datasets, evaluators, and suites from templates. Creates correctly-formatted files that pass schema validation.
5apastra-baseline
Establish and manage evaluation baselines for regression detection. A baseline is a known-good scorecard that future runs are compared against.
5apastra-getting-started
Quick setup guide for apastra PromptOps. Create your first prompt spec, dataset, evaluator, and suite in 5 minutes.
5apastra-setup-ci
Upgrade from local-first evaluation to automated GitHub Actions CI. Installs workflows for PR gating, release promotion, and auto-merge.
4