LLM inference on serverless solution - Runpod