vllm-deployment
3Deploy vLLM for high-performance LLM inference. Covers Docker CPU/GPU deployments and cloud VM provisioning with OpenAI-compatible API endpoints.
vllm
deployment
Also installable via skills CLI
npx skills add stakpak/community-paks/data/vllm-deployment