Did Runpod break vLLM bnb 4-bit?

I used Runpod a year ago and was able load a Llama3-8B finetune into vLLM and quantized using BNB on the fly to 4bit. I've been trying that with a Qwen3-14B finetune recently and I can't seem to get it to work. I also merged my finetune to 4-bit bnb safetensors and it also refuses to load. Is there some new configuration I need to use to get this to work now?
0 Replies
No replies yetBe the first to reply to this messageJoin

Did you find this page helpful?