R
Runpod2y ago
RK

how to host 20gb models + fastapi code on serverless

I have 20gb model files and a fastapi pipeline code to perform preprocessing and inference+ training. How can I use runpods serverless?
14 Replies
Unknown User
Unknown User2y ago
Message Not Public
Sign In & Join Server To View
RK
RKOP2y ago
It's Pytorch + tensorflow
Unknown User
Unknown User2y ago
Message Not Public
Sign In & Join Server To View
RK
RKOP2y ago
Do I have to dockerize the models with code?.. The docker image is around 50GB
Unknown User
Unknown User2y ago
Message Not Public
Sign In & Join Server To View
RK
RKOP2y ago
Yeah.. that's why thought to keep model out of docker
Unknown User
Unknown User2y ago
Message Not Public
Sign In & Join Server To View
RK
RKOP2y ago
Perfect that sounds a plan Thanks Can u share links for the network storage access and deploy too plz
Unknown User
Unknown User2y ago
Message Not Public
Sign In & Join Server To View
digigoblin
digigoblin2y ago
Don't use fastAPI on serverless, its already an API layer.
Neuraldivergent
and don't put your models in your image most likely, def use network drives
Unknown User
Unknown User2y ago
Message Not Public
Sign In & Join Server To View
Neuraldivergent
just several reasons to prefer smaller images, it will work but with a lot of overhead
digigoblin
digigoblin2y ago
Network drives are about 1 million percent slower than baking things into the image, so I don't know why you are saying its better, because you are wrong. Its always better to bake a model into the image wherever possible every single time, you should only use the garbage network storage if you absoultely have to.

Did you find this page helpful?