if it's a text generation model you can lower the `max_tokens`
if it's a text generation model you can lower the
max_tokensmax_tokens@cf/qwen/qwen1.5-7b-chat-awq has a context length way above 8k. Managed to chop up and squeeze in ~20k tokens, ~100k characters, and ask a question about a detail buried in the beginning of the text.@cf/llava-hf/llava-1.5-7b-hf and @cf/unum/uform-gen2-qwen-500m accept an image as input.bge-large-en-v1.5 appears on the model list, but not on the dashboard.

@cf/openchat/openchat-3.5-0106, @cf/qwen/qwen1.5-14b-chat-awq, @hf/google/gemma-7b-it are also gone. Seems like the dashboard has a limit of 50 models.eval or new Function cannot be used in Workers for security reasons. Based on the error message I assume it's trying to do some sort of schema validation which often uses new Function.
p is just random dataPOST http://workers-binding.ai/run?version=3 - Canceled @ 24/05/2024, 06:28:16[ERROR] Error compiling schema, function code: const schema2 = scope.schema[2];const schema1 = scope.schema[1]......
X [ERROR] Error in fetch handler: EvalError: Code generation from strings disallowed for this context