© 2026 Hedgehog Software, LLC
Twitter
GitHub
Discord
System
Light
Dark
More
Communities
Docs
About
Terms
Privacy
Search
Star
Feedback
Setup for Free
Error building worker-vllm docker image for mixtral 8x7b - Runpod
R
Runpod
•
3y ago
•
68 replies
wizardjoe
Error building worker-vllm docker image for mixtral 8x7b
I
'm running the following command to build and tag a docker worker image based off of worker
-vllm
:
docker build
-t lesterhnh
/mixtral
-8x7b
-instruct
-v0
.1
-runpod
-
s
e
r
v
e
r
l
e
s
s
:1
.0
-
-build
-arg MODEL
_NAME
=
"mistralai
/Mixtral
-8x7B
-Instruct
-v0
.1
"
-
-build
-arg MODEL
_BASE
_PATH
=
"
/models
"
.
I
'm getting the following error
:
-
-
-
-
-
-
D
o
c
k
e
r
f
i
l
e
:23
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
22
|
# Install torch and vllm based on CUDA version
23
|
>
>
> RUN if
[
[
"
$
{WORKER
_CUDA
_VERSION
}
"
=
= 11
.8
*
]
]
; then
\
24
|
>
>
> python3
.11
-m pip install
-U
-
-force
-reinstall torch
=
=2
.1
.2 xformers
=
=0
.0
.23
.post1
-
-index
-url
https://download.pytorch.org/whl/cu118
;
\
25
|
>
>
> python3
.11
-m pip install
-e git
+
https://github.com/runpod/
[email protected]
#egg=vllm
;
\
26
|
>
>
> else
\
27
|
>
>
> python3
.11
-m pip install
-e git
+
https://github.com/runpod/vllm-fork-for-sls-worker.git#egg=vllm
;
\
28
|
>
>
> fi
&
&
\
29
|
>
>
> rm
-rf
/root
/
.cache
/pip
30
|
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
ERROR
: failed to solve
: process
"
/bin
/bash
-o pipefail
-c if
[
[
"
$
{WORKER
_CUDA
_VERSION
}
"
=
= 11
.8
*
]
]
; then python3
.11
-m pip install
-U
-
-force
-reinstall torch
=
=2
.1
.2 xformers
=
=0
.0
.23
.post1
-
-index
-url
https://download.pytorch.org/whl/cu118
; python3
.11
-m pip install
-e git
+
https://github.com/runpod/
[email protected]
#egg=vllm
; else python3
.11
-m pip install
-e git
+
https://github.com/runpod/vllm-fork-for-sls-worker.git#egg=vllm
; fi
&
& rm
-rf
/root
/
.cache
/pip
" did not complete successfully
: exit code
: 1
GitHub
GitHub - runpod/vllm-fork-for-sls-worker: A high-throughput and mem...
A high
-throughput and memory
-efficient inference and serving engine for LLMs
- GitHub
- runpod
/vllm
-fork
-for
-sls
-worker
: A high
-throughput and memory
-efficient inference and serving engine for LLMs
Similar Threads
Next page
Similar Threads
Issues with building the new `worker-vllm` Docker Image
R
Runpod / ⚡|serverless
3y ago
Run Mixtral 8x22B Instruct on vLLM worker
R
Runpod / ⚡|serverless
2y ago
worker vllm 'build docker image with model inside' fails
R
Runpod / ⚡|serverless
3y ago
error downloading model? TheBloke/Mixtral-8x7B-MoE-RP-Story-AWQ
R
Runpod / ⚡|serverless
3y ago