Search
Star
Feedback
Setup for Free
© 2026 Hedgehog Software, LLC
Twitter
GitHub
Discord
System
Light
Dark
More
Communities
Docs
About
Terms
Privacy
Running llama 3.3 70b using vLLM and 160gb network volume - Runpod
R
Runpod
•
15mo ago
•
48 replies
Nickbkl
Running llama 3.3 70b using vLLM and 160gb network volume
Hi
, I want to check if 160 gb is enough for llama 70b and whether I can use use a smaller network volume
Runpod
Join
We're a community of enthusiasts, engineers, and enterprises, all sharing insights on AI, Machine Learning and GPUs!
21,202
Members
View on Discord
Resources
ModelContextProtocol
ModelContextProtocol
MCP Server
Similar Threads
Was this page helpful?
Yes
No
Similar Threads
Llama-3.1-Nemotron-70B-Instruct in Serverless
R
Runpod / ⚡|serverless
16mo ago
Incredibly long startup time when running 70b models via vllm
R
Runpod / ⚡|serverless
16mo ago
Not getting 100s of req/sec serving for Llama 3 70B models with default vLLM serverless template
R
Runpod / ⚡|serverless
15mo ago
Llama-70B 3.1 execution and queue delay time much larger than 3.0. Why?
R
Runpod / ⚡|serverless
2y ago