Can I use LoRA in vLLM serverless with OpenAI API?
I need both LoRA and Structured Outputs, but it seems like LoRA is only supported by Runpod API and Structured Outputs are only (poorly) supported by OpenAI API?
lora should be able to be load in the configuration via environment variable if i remember correctly, and whats the problem with structured output in openai api?