Search
Setup for Free
R
Runpod
•
10mo ago
Yobin
Can you now run gemma 3 in the vllm container?
In the serverless
, its seems im getting an error
, any help on this
Runpod
Join
We're a community of enthusiasts, engineers, and enterprises, all sharing insights on AI, Machine Learning and GPUs!
20,859
Members
View on Discord
Similar Threads
Was this page helpful?
Yes
No
© 2026 Hedgehog Software, LLC
Twitter
GitHub
Discord
System
Light
Dark
More
Communities
Docs
About
Terms
Privacy
J
Jason
•
3/17/25, 8:45 AM
can you send the error
Y
Yobin
OP
•
3/17/25, 9:23 AM
I deleted it but it seems bc gemma3 is a new model so transformer is relatively outdated afaik
?
J
Jason
•
3/17/25, 9:32 AM
hmm did you use vllm
? or only transformer
?
J
Jason
•
3/17/25, 9:32 AM
yeah maybe its a good thing to check first for compatibility
J
Jason
hmm did you use vllm? or only transformer?
Y
Yobin
OP
•
3/17/25, 10:22 AM
I used the preset vllm
, llama 3
.2b worked but the new gemma 3 didnt
D
Dj
•
3/17/25, 4:52 PM
vLLM needs to publish an update first unfortunately
D
Dj
•
3/17/25, 4:53 PM
You can use vLLM directly from the main branch
, but that
's not super easy if you
're using our vLLM template iirc
J
Jason
•
3/18/25, 8:01 AM
I think we can update and build our own vllm template from that vllm
-worker repo in github easily
J
Jason
•
3/18/25, 8:01 AM
Just update the requirement
.txt or wherever installs the vllm
B
Bj9000
•
3/24/25, 4:59 PM
Looks like vllm v0
.8
.0 added gemma3 support will the serverless vllm be updated soon
?
J
Jason
•
3/24/25, 5:05 PM
usually its delayed
, so probably a few days
/ weeks late
Y
Yobin
In the serverless, its seems im getting an error, any help on this
A
Aizen
•
3/30/25, 10:51 AM
Hi
, i have the same issue
, have you resolved it
?
, if then please help me out with it too
A
Aizen
Hi, i have the same issue , have you resolved it? , if then please help me out w...
Y
Yobin
OP
•
3/30/25, 11:19 AM
I used ollana
Y
Yobin
OP
•
3/30/25, 11:19 AM
Ollama
A
Aizen
•
3/30/25, 12:07 PM
Okay
J
Jason
•
3/30/25, 1:17 PM
yes i think vllm is updated
J
Jason
•
3/30/25, 1:17 PM
already
J
Javier
•
3/31/25, 4:56 PM
I deployed an end point to try to call
g
e
m
m
a
3
:4b but nothing is happening when I call it
, anybody managed
?
J
Jason
•
4/1/25, 1:28 AM
Yes it works
J
Jason
•
4/1/25, 1:28 AM
just now i
've tried for you
!
Output-from-gemma3.txt
5.42KB
J
Jason
•
4/1/25, 1:29 AM
you need access from hf
+ hf token to access it
J
Jason
•
4/1/25, 1:31 AM
use vllm to configure
+ check the allow remote code options in the config
( in runpod menu when configuring vllm
)
J
Jason
•
4/1/25, 1:32 AM
second image is the next page
Similar Threads
vLLM jobs not processing: "deferring container creation"
R
Runpod / ⚡|serverless
3mo ago
I can't run the pod with container start command
R
Runpod / ⛅|pods
17mo ago
Run container only once
R
Runpod / ⛅|pods
2y ago
Run Mixtral 8x22B Instruct on vLLM worker
R
Runpod / ⚡|serverless
2y ago