© 2026 Hedgehog Software, LLC
Twitter
GitHub
Discord
System
Light
Dark
More
Communities
Docs
About
Terms
Privacy
Search
Star
Feedback
Setup for Free
GGUF in serverless vLLM - Runpod
R
Runpod
•
2y ago
•
57 replies
Armyk
GGUF in serverless vLLM
How do I run a GGUF quantized model
?
I need to run this LLM
:
https://huggingface.co/mradermacher/OpenBioLLM-Llama3-70B-GGUF
What parameters should I specify
?
Thank you
mradermacher/OpenBioLLM-Llama3-70B-GGUF · Hugging Face
Recent Announcements
Next page
Similar Threads
GGUF vllm
R
Runpod / ⚡|serverless
2y ago
Serverless VLLM batching
R
Runpod / ⚡|serverless
10mo ago
Serverless vllm - lora
R
Runpod / ⚡|serverless
2y ago
vLLM Serverless error
R
Runpod / ⚡|serverless
2y ago