runpod-deployment

Installation
SKILL.md
  1. Serverless Workers - Scale-to-zero handlers with pay-per-second billing
  2. vLLM Endpoints - OpenAI-compatible LLM serving with 2-3x throughput
  3. Pod Management - Dedicated GPU instances for development/training
  4. Cost Optimization - GPU selection, spot instances, budget controls

Key deliverables:

  • Production-ready serverless handlers with streaming
  • vLLM deployment with OpenAI API compatibility
  • Cost-optimized GPU selection for any model size
  • Health monitoring and auto-scaling configuration

<quick_start> Minimal Serverless Handler (v1.8.1):

import runpod
Related skills

More from scientiacapital/skills

Installs
87
GitHub Stars
14
First Seen
Jan 22, 2026