fine-tuning-expert

Installation
Summary

Expert guidance for fine-tuning LLMs with parameter-efficient methods and production optimization.

  • Covers LoRA, QLoRA, and full fine-tuning workflows with Hugging Face PEFT, including dataset validation, hyperparameter configuration, and adapter merging for deployment
  • Provides a complete minimal working example with LoRA setup, training loop, and quantization variants for memory-constrained environments
  • Includes five-stage workflow: dataset preparation, method selection, training with checkpoints, evaluation against base model, and production deployment with quantization
  • Enforces best practices through explicit constraints: mandatory data validation, parameter-efficient methods for large models, loss curve monitoring, and held-out set evaluation before serving
SKILL.md

Fine-Tuning Expert

Senior ML engineer specializing in LLM fine-tuning, parameter-efficient methods, and production model optimization.

Core Workflow

  1. Dataset preparation — Validate and format data; run quality checks before training starts
    • Checkpoint: python validate_dataset.py --input data.jsonl — fix all errors before proceeding
  2. Method selection — Choose PEFT technique based on GPU memory and task requirements
    • Use LoRA for most tasks; QLoRA (4-bit) when GPU memory is constrained; full fine-tune only for small models
  3. Training — Configure hyperparameters, monitor loss curves, checkpoint regularly
    • Checkpoint: validation loss must decrease; plateau or increase signals overfitting
  4. Evaluation — Benchmark against the base model; test on held-out set and edge cases
    • Checkpoint: collect perplexity, task-specific metrics (BLEU/ROUGE), and latency numbers
  5. Deployment — Merge adapter weights, quantize, measure inference throughput before serving

Reference Guide

Load detailed guidance based on context:

Related skills

More from jeffallan/claude-skills

Installs
1.8K
GitHub Stars
9.0K
First Seen
Jan 21, 2026