dstack apply: Resource name should match regex
volumeEncrypted Broken in API
RTX 5090 Pod Availability
Team access to authorized ssh keys
comfyui
Pod overwrites my project code in "Volume Mount Path"
UserWarning: CUDA initialization: CUDA driver initialization failed, you might not have a CUDA gpu.
Network volumes temporarily disabled for CPU Pods - timeline query
Pod memory limits
cat /proc/meminfoAnd then runpod relies on
cgroups to enforce resource limits.
So far-so-good....AI Toolkit Lora Training torch.OutOfMemoryError
Where is My Network Volume Mounted?
Lost Workspace running (official) Stable Diffusion Template
omp.h unable to access all processors.
How to queue requests to vLLM pods?
Container Image: vllm/vllm-openai:latest) because serverless was getting very expensive.
Currently I have three pods spun up and a Next.js API which uses the Vercel ai SDK to call one of the three pods (I just choose one of the three randomly). This works okay as a fake load balancer, but sometimes the pods are all busy and I fail with:...Pod stuck trying to start custom docker image
sleep infinity.
My pod is taking forever to download the image
Pods stuck on “Waiting for logs”
I am having issues with running jupyter lab on my pod, it was running before but just got disconnect
Container Registry Auth not working for private docker images
Model Maximum Context Length Error
Container Image: vllm/vllm-openai:latest. Here is my configuration:
--host 0.0.0.0 --port 8000 --model LoneStriker/Fimbulvetr-11B-v2-AWQ --enforce-eager --gpu-memory-utilization 0.95 --api-key foo --max-model-len 4096 --max-seq-len-to-capture 4096 --trust-remote-code --chat-template "{{ (messages|selectattr('role', 'equalto', 'system')|list|last).content|trim if (messages|selectattr('role', 'equalto', 'system')|list) else '' }} {% for message in messages %} {% if message['role'] == 'user' %} ### Instruction: {{ message['content']|trim -}} {% if not loop.last %} {% endif %} {% elif message['role'] == 'assistant' %} ### Response: {{ message['content']|trim -}} {% if not loop.last %} {% endif %} {% elif message['role'] == 'user_context' %} ### Input: {{ message['content']|trim -}} {% if not loop.last %} {% endif %} {% endif %} {% endfor %} {% if add_generation_prompt and messages[-1]['role'] != 'assistant' %} ### Response: {% endif %}"
--host 0.0.0.0 --port 8000 --model LoneStriker/Fimbulvetr-11B-v2-AWQ --enforce-eager --gpu-memory-utilization 0.95 --api-key foo --max-model-len 4096 --max-seq-len-to-capture 4096 --trust-remote-code --chat-template "{{ (messages|selectattr('role', 'equalto', 'system')|list|last).content|trim if (messages|selectattr('role', 'equalto', 'system')|list) else '' }} {% for message in messages %} {% if message['role'] == 'user' %} ### Instruction: {{ message['content']|trim -}} {% if not loop.last %} {% endif %} {% elif message['role'] == 'assistant' %} ### Response: {{ message['content']|trim -}} {% if not loop.last %} {% endif %} {% elif message['role'] == 'user_context' %} ### Input: {{ message['content']|trim -}} {% if not loop.last %} {% endif %} {% endif %} {% endfor %} {% if add_generation_prompt and messages[-1]['role'] != 'assistant' %} ### Response: {% endif %}"