huggingface-import
HuggingFace to Coval Test Set Import
Import $ARGUMENTS from HuggingFace and convert it into Coval test sets with properly structured test cases.
Coval Context
Coval is an AI evaluation platform for testing voice and conversational AI agents. It runs simulations against AI agents and measures performance with configurable metrics.
| Concept | Description |
|---|---|
| Test Set | A collection of test cases, grouped by category or evaluation purpose |
| Test Case | A single evaluation scenario with input (prompt) and optional metadata |
| Persona | High-level user character (system prompt) - separate from test cases |
| Agent | The AI system being evaluated |
Key distinction:
- Persona = WHO is asking (character, traits)
- Test Case = WHAT they ask (prompts, scenarios)
More from coval-ai/coval-external-skills
onboard
>
15launch-run
Launch a Coval evaluation run against an AI agent. Use when user wants to start an evaluation, test an agent, or run simulations.
13coval-resources
Comprehensive overview of ALL Coval platform resources, their hierarchy, relationships, API endpoints, and ID formats. Use when user asks about Coval resources, data model, how things relate, what endpoints exist, or needs context about the platform structure before making API calls.
13quick-eval
Full evaluation workflow - launch a run, watch progress, and summarize results. Use for end-to-end agent testing.
13download-audio
Download audio recordings from Coval voice simulations. Use when user wants to listen to or analyze call recordings.
13watch-run
Monitor a Coval run's progress with live updates. Use when user wants to check run status or wait for completion.
13