Eren
RRunPod
•Created by Jackie on 4/21/2025 in #⚡|serverless
Runpod down?
Updated Apr 21 at 11:04am PDT
The team has identified the issue and is deploying a fix at this time.
20 replies
RRunPod
•Created by jvm-cb on 6/27/2024 in #⚡|serverless
Maximum queue size
>Is 50 just a safety measure in case WorkersMax is set to 0?
Correct
62 replies
RRunPod
•Created by michin on 4/10/2025 in #⚡|serverless
Serverless endpoints dissapeared
its back
35 replies
RRunPod
•Created by michin on 4/10/2025 in #⚡|serverless
Serverless endpoints dissapeared
They are checking
35 replies
RRunPod
•Created by dbtr on 4/6/2025 in #⚡|serverless
Serverless endpoint fails with Out Of Memory despite no changes
I don't have a1111 ready deployment but i assume it should return that above "error" key in the request fail response, that might be read and then kill worker
26 replies
RRunPod
•Created by dbtr on 4/6/2025 in #⚡|serverless
Serverless endpoint fails with Out Of Memory despite no changes
Yeah this applies using your own pipeline mostly wrapping inside of this logic
26 replies
RRunPod
•Created by hammer on 4/8/2025 in #⚡|serverless
It seems that serverless does not have an option to customize the CPU?
the hardware specs are varying between data-centers on GPU deployments
Check your workers tab and see which region it runs, remove that insufficient specs region from Edit>Advanced tab and see what happens. it will assign another worker from another region check it's specs then
9 replies
RRunPod
•Created by dbtr on 4/6/2025 in #⚡|serverless
Serverless endpoint fails with Out Of Memory despite no changes
like this:
26 replies
RRunPod
•Created by dbtr on 4/6/2025 in #⚡|serverless
Serverless endpoint fails with Out Of Memory despite no changes
if it polls/checks the request execution result and if it returns FAILED, the error key has the value, it also has this raised "e" variable as string so another way might be that
26 replies
RRunPod
•Created by dbtr on 4/6/2025 in #⚡|serverless
Serverless endpoint fails with Out Of Memory despite no changes
many ways but simple as this
26 replies
RRunPod
•Created by dbtr on 4/6/2025 in #⚡|serverless
Serverless endpoint fails with Out Of Memory despite no changes
I also strongly recommend implementing torch cache clear support and periodically clearing the garbage collector
26 replies
RRunPod
•Created by dbtr on 4/6/2025 in #⚡|serverless
Serverless endpoint fails with Out Of Memory despite no changes
Yes you can catch the OOM exception and use GraphQL API to kill the worker. OOM can occur due to several reasons and that doesn't mean that worker is a poor worker but yeah you can do that programmatically
26 replies
RRunPod
•Created by dbtr on 4/6/2025 in #⚡|serverless
Serverless endpoint fails with Out Of Memory despite no changes
I can approve that, 1 out of 100 worker having problem to run the code ending up with OOM, just kill the worker and move on
26 replies
RRunPod
•Created by dalton57918 on 3/31/2025 in #⚡|serverless
Buil docker with environment variables
Your only option is building the docker image, pushing to dockerhub, and pulling from there
12 replies
RRunPod
•Created by dalton57918 on 3/31/2025 in #⚡|serverless
Buil docker with environment variables
You can write them inside your Dockerfile for a quick fix, I remember Runpod had this on their roadmap like couple months ago
12 replies
RRunPod
•Created by Eren on 3/27/2025 in #⚡|serverless
Meaning of -u1 -u2 at the end of request id?
yeah I just wanted to get more insight about what u1 and u2
5 replies