And Quick Deploy (under https://www.runpod.io/console/serverless) shows multiple option, what should I choose? Should it be "Serverless vLLM" or something else?
NOTE: this model is NOT a LLM. It is based on the RoBERTa architecture. And RoBERTa is not directly listed as a supported model architecture in vLLM