llama.cpp is indeed more performant which is why it’s better suited for edge and serverless devices
llama.cpp is indeed more performant which is why it’s better suited for edge and serverless devices

(error) Error processing message: AiError: 3010: Invalid or incomplete input for the model: model returned: Failed to decode image: cannot identify image file <_io.BytesIO object at 0x7ee1b5d57a10>