error: '4002: could not route request to AI model', with model: '@cf/baai/bge-m3',Error: 3030: 1 validation error for VllmBatchRequest does anyone know what doesn't mean and how to resolve it?InferenceUpstreamError: error code: 1031 and now I'm getting InferenceUpstreamError: <!DOCTYPE html>... where it is returning CF html error page, with 500 error. I've tried using my production worker, and then in dev with --remote and dev locally. In all cases it returns some variation of the above but all fail. The code worked fine yesterday and hasn't changed. It fails at the call "@cf/qwen/qwen2.5-coder-32b-instruct". All other requests appear to work.remote = true included in my wrangler.toml I had it set in both queues and r2_buckets, but oddly not in ai. Removing them fixed the problem.remote = true solved the problem in dev but not in production. I redeployed worker, but the request still hangs. Nothing in logs...Workers AI: 9003: unknown internal error for exampleWorkers AI: WORKERS AI: Operation timed out after 40000 msNotes on input format
Smart Turn takes 16kHz PCM audio as input. Up to 8 seconds of audio is supported, and we recommend providing the full audio of the user's current turn.
The model is designed to be used in conjunction with a lightweight VAD model such as Silero. Once the VAD model detects silence, run Smart Turn on the entire recording of the user's turn, truncating from the beginning to shorten the audio to around 8 seconds if necessary.
If additional speech is detected from the user before Smart Turn has finished executing, re-run Smart Turn on the entire turn recording, including the new audio, rather than just the new segment. Smart Turn works best when given sufficient context, and is not designed to run on very short audio segments.
Note that audio from previous turns does not need to be included.
.venv/lib/python3.10/site-packages/llama_index/embeddings/cloudflare_workersai/base.py", line 118, in _aget_text_embeddings
return resp["result"]["data"]
KeyError: 'data'