Can't deploy Qwen/Qwen2.5-14B-Instruct-1M on serverless
Steps to reproduce:
1. Use Serverless vLLM quick deploy for Qwen/Qwen2.5-14B-Instruct-1M (image attached)
2. Proceed with default config.
3. Try and send a request.
Error:
How do I fix this?
I've been trying to troubleshoot this all morning. All help appreciated 🙏

4 Replies
Unknown User•5mo ago
Message Not Public
Sign In & Join Server To View
Yes - doesn't appear to be env variables on worker vLLM – https://github.com/runpod-workers/worker-vllm
GitHub
GitHub - runpod-workers/worker-vllm: The RunPod worker template for...
The RunPod worker template for serving our large language model endpoints. Powered by vLLM. - runpod-workers/worker-vllm
Unknown User•5mo ago
Message Not Public
Sign In & Join Server To View
For my use case a env variable was missing to but I did a PR and they merged it in like 2h.
Try that.