@cf/mistral/mistral-7b-instruct-v0.2-lora and worker-aiInferenceUpstreamError: ERROR 3028: Unknown internal error
export interface Env {
AI: any;
}
export default {
async fetch(request, env): Promise<Response> {
const response = await env.AI.run(
"@cf/mistral/mistral-7b-instruct-v0.2-lora", //the model supporting LoRAs
{
messages: [{"role": "user", "content": "Hello world"}],
raw: true,
lora: "4e900000-0000-0000-0000-000000000",
}
);
return new Response(JSON.stringify(response));
},
} satisfies ExportedHandler<Env>;
Mistral-7B-Instruct-v0.2 as base Model (https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2)} after Hello world"mistral and not mistralai
raw: true but then providing the messages field which I don't believe is compatible?raw: true then you need to format it to ChatML yourself and provide it as an input string if I remember correctlyraw property also tried true/false using the correct message/prompt formatInferenceUpstreamError: ERROR 3028: Unknown internal error{
"adapter_file": null,
"adapter_path": "adapters",
"batch_size": 5,
"config": "lora.yml",
"data": "./data/",
"grad_checkpoint": false,
"iters": 1800,
"learning_rate": 1e-05,
"lora_layers": 19,
"lora_parameters": {
"keys": [
"self_attn.q_proj",
"self_attn.v_proj"
],
"rank": 8,
"alpha": 16.0,
"scale": 10.0,
"dropout": 0.05
},
"lr_schedule": {
"name": "cosine_decay",
"warmup": 100,
"warmup_init": 1e-07,
"arguments": [
1e-05,
1000,
1e-07
]
},
"max_seq_length": 32768,
"model": "mistralai/Mistral-7B-Instruct-v0.2",
"model_type": "mistral",
"resume_adapter_file": null,
"save_every": 100,
"seed": 0,
"steps_per_eval": 20,
"steps_per_report": 10,
"test": false,
"test_batches": 100,
"train": true,
"val_batches": -1
}adapter_config.json (click to show the full content)adapter_config.json is now: see message above workers-ai adapter_model.safetensors I recevie a new error from the wrangler✘ [ERROR] 🚨 Couldn't upload file: A request to the Cloudflare API (/accounts/1111122223334444/ai/finetunes/6a4a4a4a4a4a4a4a4-a5aa5a5a-aaaaaa/finetune-assets) failed. FILE_PARSE_ERROR: 'file' should be of valid safetensors type [code: 1000], quiting...InferenceUpstreamError: ERROR 3028: Unknown internal error when I try to run an inference.rank r <=8 or quantization = None) that has not been specified into the documentation ?
--target_modules q_proj,v_proj as target modules with autotrain
@cf/meta/llama-3-8b-instruct to be broken/unavailable? I can't get a response, but @cf/meta/llama-2-7b-chat-fp16 works fineActive Models filterFILE_PARSE_ERROR with my autotrained finetuneError: no available backend found. ERR: [wasm] RuntimeError: Aborted(both async and sync fetching of the wasm failed) . But the other files (config, onnx) are loading up fine. @Xenova I hope it's still okay to tag you! I'm so excited to get your library up and running!✘ [ERROR] 🚨 Couldn't upload file: A request to the Cloudflare API (/accounts/1111122223334444/ai/finetunes/6a4a4a4a4a4a4a4a4-a5aa5a5a-aaaaaa/finetune-assets) failed. FILE_PARSE_ERROR: 'file' should be of valid safetensors type [code: 1000], quiting...[wrangler:err] InferenceUpstreamError: AiError: undefined: ERROR 3001: Unknown internal error
at Ai.run (cloudflare-internal:ai-api:66:23)
at async Object.fetch (file:///C:/Users/Fathan/PetProject/cloudflare-demo/src/index.ts:15:21)
at async jsonError (file:///C:/Users/Fathan/PetProject/cloudflare-demo/node_modules/wrangler/templates/middleware/middleware-miniflare3-json-error.ts:22:10)
at async drainBody (file:///C:/Users/Fathan/PetProject/cloudflare-demo/node_modules/wrangler/templates/middleware/middleware-ensure-req-body-drained.ts:5:10)
[wrangler:inf] POST / 500 Internal Server Error (2467ms)@cf/mistral/mistral-7b-instruct-v0.2-lora
export interface Env {
AI: any;
}
export default {
async fetch(request, env): Promise<Response> {
const response = await env.AI.run(
"@cf/mistral/mistral-7b-instruct-v0.2-lora", //the model supporting LoRAs
{
messages: [{"role": "user", "content": "Hello world"}],
raw: true,
lora: "4e900000-0000-0000-0000-000000000",
}
);
return new Response(JSON.stringify(response));
},
} satisfies ExportedHandler<Env>;
Mistral-7B-Instruct-v0.2}Hello world"mistralmistralairaw: trueraw: trueraw{
"adapter_file": null,
"adapter_path": "adapters",
"batch_size": 5,
"config": "lora.yml",
"data": "./data/",
"grad_checkpoint": false,
"iters": 1800,
"learning_rate": 1e-05,
"lora_layers": 19,
"lora_parameters": {
"keys": [
"self_attn.q_proj",
"self_attn.v_proj"
],
"rank": 8,
"alpha": 16.0,
"scale": 10.0,
"dropout": 0.05
},
"lr_schedule": {
"name": "cosine_decay",
"warmup": 100,
"warmup_init": 1e-07,
"arguments": [
1e-05,
1000,
1e-07
]
},
"max_seq_length": 32768,
"model": "mistralai/Mistral-7B-Instruct-v0.2",
"model_type": "mistral",
"resume_adapter_file": null,
"save_every": 100,
"seed": 0,
"steps_per_eval": 20,
"steps_per_report": 10,
"test": false,
"test_batches": 100,
"train": true,
"val_batches": -1
}