RunpodR
Runpodβ€’14mo ago
Arad

Deploying bitsandbytes-quantized Models on RunPod Serverless using Custom Docker Image

Hey everyone πŸ‘‹
Looking for tips from anyone who's worked with bitsandbytes-quantized models on RunPod's serverless setup. It's not available out of the box with vLLM, and I was wondering if anyone's got it working? Saw a post in the serverless forum about maybe using a custom Docker image for this.

For context: I've fine-tuned LLaMA-3.1 70B-instruct using the unsloth library (which utilizes bitsandbytes for quantization) and am looking to deploy it.
Any insights would be greatly appreciated! πŸ™
Was this page helpful?