Search
Setup for Free
R
Runpod
•
10mo ago
sahir
queue delay times
Hi
, I
'm seeing really long delay times
. even though there
's nothing in the queue
, and this is a really small CPU serverless endpoint
. Any idea what causes this
?
Runpod
Join
We're a community of enthusiasts, engineers, and enterprises, all sharing insights on AI, Machine Learning and GPUs!
20,859
Members
View on Discord
Similar Threads
Was this page helpful?
Yes
No
© 2026 Hedgehog Software, LLC
Twitter
GitHub
Discord
System
Light
Dark
More
Communities
Docs
About
Terms
Privacy
Y
youssef
•
2/25/25, 9:34 PM
I
'm having same issue
, on 16gb
/24gb
, my request stays a shit ton of time in the queue
, these are only two items
Y
youssef
•
2/25/25, 9:34 PM
cc
@PRB
- any issue on the statuspage
?
P
PRB
•
2/26/25, 4:34 PM
@sahir which datacenter are you running this in
?
P
PRB
•
2/26/25, 4:35 PM
@youssef can you open a support ticket
? we are already looking at this but will be nice to keep track and get back to you
P
PRB
@sahir which datacenter are you running this in?
S
sahir
OP
•
2/27/25, 3:52 AM
all locations were selected
, so its made workers here
S
sahir
OP
•
2/27/25, 6:19 AM
This is happening to my other endpoints too now
K
Kays
•
2/27/25, 3:40 PM
Same here
, almost 2 minutes cold start every time
K
Kays
•
2/27/25, 3:41 PM
but once every few requests it goes
<5 seconds again
P
PRB
•
2/27/25, 5:37 PM
are you guys on cpu endpoints or GPU endpoints
?
P
PRB
•
2/27/25, 5:38 PM
@Kays please reply so i can resolve faster
K
Kays
•
2/27/25, 5:39 PM
GPU endpoints
K
Kays
•
2/27/25, 5:39 PM
Its mostly A100 for me
K
Kays
•
2/27/25, 5:39 PM
I mean h100
*
P
PRB
•
2/27/25, 5:43 PM
endpoint id will help
K
Kays
•
2/27/25, 5:52 PM
pury32p7r6r4wf
K
Kays
•
2/27/25, 5:53 PM
I can give you an example test request if you like
K
Kays
pury32p7r6r4wf
F
flash-singh
•
2/27/25, 6:40 PM
your cold starts are high
, are you loading model from network volume or is the model just too big
?
F
flash-singh
your cold starts are high, are you loading model from network volume or is the m...
K
Kays
•
2/27/25, 6:42 PM
I
'm not using network volumes
, the model is flux
-dev
(24gb
)
K
Kays
•
2/27/25, 6:42 PM
But what
's weird is that cold start sometimes is extremely quick
, like under 5 seconds
K
Kays
•
2/28/25, 2:02 PM
Hey there
, any updates on this
? Is it just the model being too big
?
@PRB
@flash
-singh thanks
!
K
Kays
•
2/28/25, 5:33 PM
seems to be fixed now somehow
K
Kays
But what's weird is that cold start sometimes is extremely quick, like under 5 s...
F
flash-singh
•
2/28/25, 5:54 PM
thats just flashboot
, anything over 10s should be your ideal cold start
, is your model baked into the container image
?
K
Kays
•
2/28/25, 7:22 PM
yes it is on the container
K
Kays
•
2/28/25, 7:22 PM
right now I
'm getting around 50
/50 flashboots
F
flash-singh
•
2/28/25, 11:50 PM
thats about right
, depends on workload and capacity
, for h100s thats really good if your p50 is hitting flashboot
K
Kays
•
3/1/25, 12:18 AM
Cool
, yes I
’m happy with that rate
, yesterday was more like 90
-10 that
’s why I mentioned it
Similar Threads
Queue Delay Time
R
Runpod / ⚡|serverless
7mo ago
Mixed Delay Times
R
Runpod / ⚡|serverless
2y ago
Delay times on requests
R
Runpod / ⚡|serverless
14mo ago
Insane delay times as of late
R
Runpod / ⚡|serverless
3mo ago