together-evaluations

Installation
SKILL.md

Together AI Evaluations

Overview

Use Together AI evaluations when the user wants a managed LLM-as-a-judge workflow rather than an ad hoc prompt loop.

Core evaluation types:

  • Classify: assign outputs to labels
  • Score: grade outputs on a numeric scale
  • Compare: compare two candidate outputs with bias controls

This skill also covers external providers used as judges or targets when the workflow still runs through Together AI's evaluation system.

When This Skill Wins

  • Benchmark prompt variants, models, or product responses
Related skills

More from zainhas/skills

Installs
1
Repository
zainhas/skills
First Seen
Mar 30, 2026