Reading the serverless documentation, under the autoscale section, it says "Dynamically scale workers from 0 to 100 on the Secure Cloud platform, which is highly available and distributed globally. This provides users with the computational resources exactly when needed." Not sure if 0 to 100 is meant literally or figuratively-
Our current provider has around 50 H100s available so this is an active point of investigation for us.
TLDR: Can we scale past 100 GPUs on enterprise plans? Is there an enterprise POC I can reach out to?
Recent Announcements
Continue the conversation
Join the Discord to ask follow-up questions and connect with the community
R
Runpod
We're a community of enthusiasts, engineers, and enterprises, all sharing insights on AI, Machine Learning and GPUs!