R
RunPod14mo ago
Adam?

is there any method to deploy bert architecture models serverlessly?

is there any method to deploy bert architecture models serverlessly?
Solution:
@Adam? https://www.runpod.io/console/explore then select this...
No description
Jump to solution
9 Replies
Jason
Jason13mo ago
Hi There is using huggingface's transformers you can cache the model first in some path, then load it using relative / absolute path. put the model in the image / use a network volume
Adam?
Adam?OP13mo ago
Hi, thanks for help Is there any template for this? Or I should write the handler by myself?
Jason
Jason13mo ago
yes there is use VLLM template
Solution
Jason
Jason13mo ago
Jason
Jason13mo ago
If you need help with configuring that template theres a setup menu after you click on it
Adam?
Adam?OP13mo ago
Yeah I have try this
Jason
Jason13mo ago
oh wait bert isnt compatible with that isnt it?
Adam?
Adam?OP13mo ago
But it doesn't supports bart models Exactly
Jason
Jason13mo ago
Hmm then you'll have to write an custom handler first for beert using transformers pipeline works too

Did you find this page helpful?