ml-inference-optimization

ML inference latency optimization, model compression, distillation, caching strategies, and edge deployment patterns. Use when optimizing inference performance, reducing model size, or deploying ML at

by melodic-software· Repository·devops
Also installable via skills CLI
npx skills add melodic-software/claude-code-plugins/devops/ml-inference-optimization

Source

Path:devops/ml-inference-optimization(main)

Related in devops