how can I upgrade vllm version in serverless worker image?

https://docs.runpod.io/serverless/endpoints/model-caching Hi, I was trying the new Qwen3-next-80B-a3b-instruct in the example, and I selected vllm worker image to run it, but it seems that the vllm version in the image doesn't support the model yet. so I added the pip upgrade transformers and vllm in the docker start command, and the image still tries to load the model in the old version.
Runpod Documentation
Cached models - Runpod Documentation
Accelerate cold starts and reduce costs by using cached models.
3 Replies
Vazyrac_Carlcox
Vazyrac_CarlcoxOP2d ago
Plus, I have a network storage with fine tuned model of my company. Can I change the model path in the serverless image to the /workspace in the network volume so that it can skip the model pulling processes? Cheers.
Dj
Dj19h ago
Inside serverless it will always be mounted to /runpod-volume, sorry :(
Vazyrac_Carlcox
Vazyrac_CarlcoxOP7h ago
I understand. Then what's the point of enabling network in the 'Advanced' in the serverless config? I wonder if i can use it for some other workarounds.

Did you find this page helpful?