ml-inference-optimization
ML inference latency optimization, model compression, distillation, caching strategies, and edge deployment patterns. Use when optimizing inference performance, reducing model size, or deploying ML at
Also installable via skills CLI
npx skills add melodic-software/claude-code-plugins/devops/ml-inference-optimization