addon-llm-judge-evals
Add-on: LLM Judge Evals
Use this skill when you need qualitative evaluation (clarity, domain fit, UX coherence, docs quality) in addition to deterministic checks.
Compatibility
- Works with all stacks.
- Best paired with
addon-deterministic-eval-suite.
Inputs
Collect:
JUDGE_BACKEND:auto|langchain|google-adk(defaultauto).JUDGE_MODEL: model id to run scoring.JUDGE_TIMEOUT_SECONDS: default60.JUDGE_MAX_RETRIES: default2.JUDGE_TEMPERATURE: default0.JUDGE_FAIL_ON_BACKEND_MISMATCH:yes|no(defaultyes).JUDGE_RUBRIC_MODE:product|security|developer-experience|custom.
More from ajrlewis/ai-skills
architect-python-uv-fastapi-sqlalchemy
Use when scaffolding production-ready FastAPI services with uv, SQLAlchemy, Alembic, Postgres, Docker, and CI gates.
11addon-rag-ingestion-pipeline
Use when adding multi-format RAG ingest, chunk, embed, and retrieval pipelines; pair with architect-python-uv-batch or architect-python-uv-fastapi-sqlalchemy.
11addon-docling-legal-chunk-embed
Use when you need legal PDF to markdown extraction plus clause chunking and embedding prep; pair with addon-rag-ingestion-pipeline and architect-python-uv-batch.
10addon-llm-ancient-greek-translation
Use when adding Koine or Attic Greek translation to Next.js content flows; pair with ui-editorial-writing-surface and addon-nostr-nip23-longform.
10architect-python-uv-batch
Use when scaffolding production-ready Python uv batch or worker projects with Docker required by default.
10addon-human-pr-review-gate
Use when agent-generated code must pass a human PR review gate with trusted checks and merge blocks; pair with addon-decision-justification-ledger and architect-stack-selector.
9