Search
Star
Feedback
Setup for Free
© 2026 Hedgehog Software, LLC
Twitter
GitHub
Discord
System
Light
Dark
More
Communities
Docs
About
Terms
Privacy
The total token limit at 131 - Runpod
R
Runpod
•
9mo ago
•
20 replies
tanawatl
The total token limit at 131
I use vLLM and set max model length to 8000 a2048 but out is just 131
(total out
+ in
)
, although i have set max tokens to 2048
. I try with 2 models and result is the same
.
Runpod
Join
We're a community of enthusiasts, engineers, and enterprises, all sharing insights on AI, Machine Learning and GPUs!
21,202
Members
View on Discord
Resources
ModelContextProtocol
ModelContextProtocol
MCP Server
Recent Announcements
Similar Threads
Was this page helpful?
Yes
No
Similar Threads
Serverless SGLang - 128 max token limit problem.
R
Runpod / ⚡|serverless
14mo ago
Total crash | HELP!!!!
R
Runpod / ⚡|serverless
4mo ago
Automate the generation of the ECR token in Serverless endpoint?
R
Runpod / ⚡|serverless
3y ago
How can I limit the queue "in progress"?
R
Runpod / ⚡|serverless
2y ago