© 2026 Hedgehog Software, LLC

TwitterGitHubDiscord
More
CommunitiesDocsAboutTermsPrivacy
Search
Star
Setup for Free
RunpodR
Runpod•2y ago•
81 replies
Thibaud

vllm seems not use GPU

i'm using vllm
and on the graph, when i launch some request, only cpu usage increase.
if i open a terminal and launch nvidia-smi, i didn't see any process too.

settings line
--model NousResearch/Meta-Llama-3-8B-Instruct --max-model-len 8192 --port 8000 --dtype half --enable-chunked-prefill true --max-num-batched-tokens 6144 --gpu-memory-utilization 0.97
image.png
Runpod banner
RunpodJoin
We're a community of enthusiasts, engineers, and enterprises, all sharing insights on AI, Machine Learning and GPUs!
21,202Members
Resources
Recent Announcements

Similar Threads

Was this page helpful?
Next page

Similar Threads

vLLM and multiple GPUs
RunpodRRunpod / ⛅|pods
11mo ago
Runpod VLLM - How to use GGUF with VLLM
RunpodRRunpod / ⛅|pods
17mo ago
Issue with vLLM on L40s GPU on RunPod
RunpodRRunpod / ⛅|pods
7mo ago
GPU don't use
RunpodRRunpod / ⛅|pods
2y ago