I am getting the following error when building the new worker-vllm image with my model.
=> ERROR [vllm-base 6/7] RUN --mount=type=secret,id=HF_TOKEN,required=false if [ -f /run/secrets/HF_TOKEN ]; then export HF_TOKEN=$(cat /run/secrets/HF_TOKEN); fi && if [ -n "Pate 10.5s------ > [vllm-base 6/7] RUN --mount=type=secret,id=HF_TOKEN,required=false if [ -f /run/secrets/HF_TOKEN ]; then export HF_TOKEN=$(cat /run/secrets/HF_TOKEN); fi && if [ -n "PatentPilotAI/mistral-7b-patent-instruct-v2" ]; then python3 /download_model.py; fi:#10 9.713 Traceback (most recent call last):#10 9.713 File "/download_model.py", line 4, in <module>#10 9.715 from vllm.model_executor.weight_utils import prepare_hf_model_weights, Disabledtqdm#10 9.715 File "/vllm-installation/vllm/model_executor/__init__.py", line 2, in <module>#10 9.715 from vllm.model_executor.model_loader import get_model#10 9.715 File "/vllm-installation/vllm/model_executor/model_loader.py", line 10, in <module>#10 9.715 from vllm.model_executor.weight_utils import (get_quant_config,#10 9.715 File "/vllm-installation/vllm/model_executor/weight_utils.py", line 18, in <module>#10 9.715 from vllm.model_executor.layers.quantization import (get_quantization_config,#10 9.715 File "/vllm-installation/vllm/model_executor/layers/quantization/__init__.py", line 4, in <module>#10 9.716 from vllm.model_executor.layers.quantization.awq import AWQConfig#10 9.716 File "/vllm-installation/vllm/model_executor/layers/quantization/awq.py", line 6, in <module>#10 9.716 from vllm._C import ops#10 9.716 ImportError: libcuda.so.1: cannot open shared object file: No such file or directory
=> ERROR [vllm-base 6/7] RUN --mount=type=secret,id=HF_TOKEN,required=false if [ -f /run/secrets/HF_TOKEN ]; then export HF_TOKEN=$(cat /run/secrets/HF_TOKEN); fi && if [ -n "Pate 10.5s------ > [vllm-base 6/7] RUN --mount=type=secret,id=HF_TOKEN,required=false if [ -f /run/secrets/HF_TOKEN ]; then export HF_TOKEN=$(cat /run/secrets/HF_TOKEN); fi && if [ -n "PatentPilotAI/mistral-7b-patent-instruct-v2" ]; then python3 /download_model.py; fi:#10 9.713 Traceback (most recent call last):#10 9.713 File "/download_model.py", line 4, in <module>#10 9.715 from vllm.model_executor.weight_utils import prepare_hf_model_weights, Disabledtqdm#10 9.715 File "/vllm-installation/vllm/model_executor/__init__.py", line 2, in <module>#10 9.715 from vllm.model_executor.model_loader import get_model#10 9.715 File "/vllm-installation/vllm/model_executor/model_loader.py", line 10, in <module>#10 9.715 from vllm.model_executor.weight_utils import (get_quant_config,#10 9.715 File "/vllm-installation/vllm/model_executor/weight_utils.py", line 18, in <module>#10 9.715 from vllm.model_executor.layers.quantization import (get_quantization_config,#10 9.715 File "/vllm-installation/vllm/model_executor/layers/quantization/__init__.py", line 4, in <module>#10 9.716 from vllm.model_executor.layers.quantization.awq import AWQConfig#10 9.716 File "/vllm-installation/vllm/model_executor/layers/quantization/awq.py", line 6, in <module>#10 9.716 from vllm._C import ops#10 9.716 ImportError: libcuda.so.1: cannot open shared object file: No such file or directory
Recent Announcements
Continue the conversation
Join the Discord to ask follow-up questions and connect with the community
R
Runpod
We're a community of enthusiasts, engineers, and enterprises, all sharing insights on AI, Machine Learning and GPUs!