Hello All, We have multiple serverless endpoints that downloads the model and generate the inference. Is there is a way to mount a common volume to all the serverless endpoint system. We don't want to down the model every time endpoint boots up. It would be nice if you can please share a concrete example
Continue the conversation
Join the Discord to ask follow-up questions and connect with the community
R
Runpod
We're a community of enthusiasts, engineers, and enterprises, all sharing insights on AI, Machine Learning and GPUs!