Update Transformers Library
Hi, I am trying to run Qwen/Qwen3-Embedding-8B via serverless endpoints.
1. I select quick deploy, Infinity Vector Embeddings.
2. Set Qwen/Qwen3-Embedding-8B as the model.
3. Batch size 32, data type auto.
4. Torch backend
5. Endpoint type: queue
6. 24 Gb GPU selected
Inspecting worker logs:
This is a known issue as per HF: https://huggingface.co/Qwen/Qwen3-Embedding-8B#usage
I need to update the Transformers library. How can I do that? Any guidance is much appreciated.
2 Replies