Runpod

R

Runpod

We're a community of enthusiasts, engineers, and enterprises, all sharing insights on AI, Machine Learning and GPUs!

Join

⚡|serverless

⛅|pods

🔧|api-opensource

📡|instant-clusters

🗂|hub

RunPod Serverless Endpoint Issue - Jobs Complete But No Output Returned

Problem: My serverless endpoint jobs are completing successfully but returning empty results. Endpoint ID: su6ufhaephnw03 (Stable Diffusion XL) Symptoms:...

Serverless Endpoint with vLLM (Qwen2.5-VL-3B-Instruct)

I’m trying to set up a Serverless Endpoint on RunPod with vLLM (with Qwen2.5-VL-3B-Instruct). My goal is to get a lot of images descriptions. Here is how i set it up: Docker Image:...

Build in pending for hours

Has anyone else experienced insanely long pending time before the build starts? This just happened to me a few days ago as is still happening. Every build is in pending state for hours or starts at very unpredictable times....

5090 disappeared in Serverless

Hey 🖐 Is it just me, or has the 5090 disappeared when creating an endpoint? Or is it a temporary limitation due to lack of available resources?...
No description

Setting up a serverless endpoint for a custom model

Hi, complete beginner here. I've been trying to set up a custom model endpoint, but keep getting issues with having my requests answered (using the web application). I've gotten delay times of 40+ minutes so clearly something is wrong but I am having a hard time troubleshooting. Here are some details: - trained a model (base model deepseek-ai/DeepSeek-R1-Distill-Llama-8B) and saved to a huggingface repo - trying to publish endpoint via GitHub integration (already have folder containing Dockerfile, handler.py, requirements.txt)...

Regions with better/guaranteed bandwidth

Hi, I was interested in running serverless tasks but I need to upload the results of them reasonably fast (ideally in the 10 Gbps per gpu range). Are there regions that have faster or guaranteed bandwidth in the secure cloud, or is it possible to filter by that? Thanks!

Updating CMD override after creating endpoint

I noticed that it's not possible to edit the CMD override after creating a serverless endpoint. I have to create an entirely new endpoint if I misconfigure / need to update this. Why?

Configuring endpoints via API if publishing via Github integration

It seems that there's no way to create / update endpoints via the API, unless you host the images on a docker registry. Are there plans to extend API support to endpoints that are published from a github branch?...

s3 path for serverless image gen uploads (comfyui)

is it possible to set it via request body? if yes, how? we are using flux-schnel workflow

Most Available GPUs

Hi, I am using serverless with Network Volume attached to it, I need to know which region has the most available GPUs in the day, No matter what type of GPUs, The problem is my volume at US-CA-2 but most of the GPUs are not available and it usually initializing then throttled and most of the time unavailable, I know it's generic question and GPUs can't be trackable as they are shared but if anyone tried this before can help. Thank you....

Serverless worker won't even start but counts as running

Hi so lately I've been dealing with various issues regarding serverless workers (i am using a custom docker file). At first i used the official base image with cuda 11.8 (and an older version of pytorch) and it worked fine with the 3090 but not with the 5090 (i have two serverless endpoints, one with "lower end gpus" and the other one with "higher end gpus"). So i used the base image with the latest version of everything (pytorch 2.7.1, cuda 12.9, ubuntu 24.04) but for some reason now the 3090 pod didn't work and the 5090 pod gave error CUDA_ERROR_NO_DEVICE: no CUDA-capable device is detected. I tried to tweak the docker image a bit but no success. Then i made the docker image install the package nvidia-cuda-toolkit and now, on the 5090 pod i get error CUDA error: no kernel image is available for execution on the device....

Hi All,

Has anyone faced a problem reading some models like Pulid from Network Volume attached to the serverless? I am getting this error PulidFluxModelLoader 45: - Value not in list: pulid_file: 'pulid_flux_v0.9.1.safetensors' not in [] Any help please?...
No description

For 24G Pro Machines: Memory Allocation Must Not Be Less Than 60GB

I use the GPU of the 24 Pro, and the system allocates memory randomly. When the allocated memory is 46GB, I encounter an out-of-memory error. Could the allocated memory be at least 60GB or more? Please do not assign me a machine with 46GB of memory.

Requests vs Jobs

On serverless, what's the relationship between a "request" and a "job"? I cancelled all of a worker's requests over 10 minutes ago and yet the logs show it still going (attached).
No description

No cuda driver

RuntimeError: CUDA driver initialization failed, you might not have a CUDA gpu.

How to create serverless template

Hey guys. Where is the option to pick the type to be "serverless"??? I created a template and it is still not seen when i want to create a serverless endpoint

Automation of Image Updates for a Serverless Endpoint

Hello everyone, I would like to automate the image update for a serverless endpoint. After reading the API and runpodctl documentation, I couldn’t find such an option. Is it possible to create new releases using the API or runpodctl? If there are other possible solutions, I would greatly appreciate your suggestions....

Billing history for endpoints not working

Solution:
its serverless
Message Not Public
Sign In & Join Server To View
No description

Some tasks are consistently in the IN_PROGRESS state

I'm trying out runpod and I've simply created a whisper task, but some tasks are consistently in the IN_PROGRESS state, and I don't know how to fix it.