Post
223
Want to deploy open models using vLLM as the inference engine?
We just released a step-by-step guide on how to do it with @huggingface Inference Endpoints, now available in the vLLM docs.
let the gpus go brrr
https://docs.vllm.ai/en/latest/deployment/frameworks/hf_inference_endpoints.html
We just released a step-by-step guide on how to do it with @huggingface Inference Endpoints, now available in the vLLM docs.
let the gpus go brrr
https://docs.vllm.ai/en/latest/deployment/frameworks/hf_inference_endpoints.html