ComfyUI Worker: FLUX.1 dev & Network Volume Setup Questions

Few questions: https://github.com/runpod-workers/worker-comfyui runpod/worker-comfyui:<version>-flux1-dev: Includes checkpoint, text encoders, and VAE for FLUX.1 dev <---- This model is using the fp8 version right now and not the full version, is that right? Thats why I want to now use this because I want to use the full size flux version: https://github.com/runpod-workers/worker-comfyui/blob/main/docs/customization.md I want to use network storage but I am not sure how i need to do it. What i wanted to do was create pod temporarily where the network storage is attached to and then just upload the full flux model etc. into the /workspace directory with this strucutre that is described in the customization.md Example structure inside the Network Volume: /models/checkpoints/your_model.safetensors /models/loras/your_lora.pt /models/vae/your_vae.safetensors But I think I need to use a symlink so it will work on server less later as well? Is that correct? Since here it mentions /runpod-volume <--- https://docs.runpod.io/serverless/storage/network-volumes So i am confused on what I need to do now so would something like this be correct then?
# Start from the base image
FROM runpod/worker-comfyui:5.1.0-base

# Install required packages and custom nodes
RUN pip install huggingface-hub
RUN comfy-node-install https://github.com/olduvai-jp/ComfyUI-HfLoader

RUN rm -rf /workspace && \
ln -s /runpod-volume /workspace
# Start from the base image
FROM runpod/worker-comfyui:5.1.0-base

# Install required packages and custom nodes
RUN pip install huggingface-hub
RUN comfy-node-install https://github.com/olduvai-jp/ComfyUI-HfLoader

RUN rm -rf /workspace && \
ln -s /runpod-volume /workspace
If not what would be the correct way to do it?
39 Replies
ScaledEcom
ScaledEcomOP4mo ago
uploaded my stuff to network drive manually with temporay pod and now using: this is my dockerfile:
FROM runpod/worker-comfyui:5.2.0-base

# Install your custom nodes
RUN pip install huggingface-hub
RUN comfy-node-install https://github.com/olduvai-jp/ComfyUI-HfLoader
FROM runpod/worker-comfyui:5.2.0-base

# Install your custom nodes
RUN pip install huggingface-hub
RUN comfy-node-install https://github.com/olduvai-jp/ComfyUI-HfLoader
another thing I am wondering about: see attached file I have my networkdrive conncted where i have models etc already but it looks like stuff is installed to /comfyui/models/ anyways even though stuff is already on the networkdrive? (isnt that unesscacry? and is causing start ups to take much longer? i have the models on like flux etc on the networkdrive already
ScaledEcom
ScaledEcomOP4mo ago
shortned log file, see latest message for compelte log file
ScaledEcom
ScaledEcomOP4mo ago
which pytorch version is used by this? since it seems it can not be used with rtx 5090 NVIDIA GeForce RTX 5090 with CUDA capability sm_120 is not compatible with the current PyTorch installation. The current PyTorch install supports CUDA capabilities sm_50 sm_60 sm_70 sm_75 sm_80 sm_86 sm_90. how to fix this? here is also compelte logs file
Unknown User
Unknown User4mo ago
Message Not Public
Sign In & Join Server To View
ScaledEcom
ScaledEcomOP4mo ago
@Elder Papa Madiator @Madiator2011 (Work) Lets continue the communication here so we don't spam your Better Comfy Ui Slim Thread.
ScaledEcom
ScaledEcomOP4mo ago
No description
ScaledEcom
ScaledEcomOP4mo ago
No description
ScaledEcom
ScaledEcomOP4mo ago
Still getting this error in the logs
ScaledEcom
ScaledEcomOP4mo ago
I see that the CUDA Version is 12.8.1 now but also seeing this in the logs
2025-07-17T15:31:10.699315136Z /opt/venv/lib/python3.12/site-packages/torch/cuda/__init__.py:287: UserWarning:
2025-07-17T15:31:10.699358195Z NVIDIA GeForce RTX 5090 with CUDA capability sm_120 is not compatible with the current PyTorch installation.
2025-07-17T15:31:10.699389844Z The current PyTorch install supports CUDA capabilities sm_50 sm_60 sm_70 sm_75 sm_80 sm_86 sm_90.
2025-07-17T15:31:10.699394994Z If you want to use the NVIDIA GeForce RTX 5090 GPU with PyTorch, please check the instructions at https://pytorch.org/get-started/locally/
2025-07-17T15:31:10.699404694Z warnings.warn(
2025-07-17T15:31:10.826379471Z Total VRAM 32120 MB, total RAM 1160740 MB
2025-07-17T15:31:10.826405240Z pytorch version: 2.7.1+cu126
2025-07-17T15:31:10.699315136Z /opt/venv/lib/python3.12/site-packages/torch/cuda/__init__.py:287: UserWarning:
2025-07-17T15:31:10.699358195Z NVIDIA GeForce RTX 5090 with CUDA capability sm_120 is not compatible with the current PyTorch installation.
2025-07-17T15:31:10.699389844Z The current PyTorch install supports CUDA capabilities sm_50 sm_60 sm_70 sm_75 sm_80 sm_86 sm_90.
2025-07-17T15:31:10.699394994Z If you want to use the NVIDIA GeForce RTX 5090 GPU with PyTorch, please check the instructions at https://pytorch.org/get-started/locally/
2025-07-17T15:31:10.699404694Z warnings.warn(
2025-07-17T15:31:10.826379471Z Total VRAM 32120 MB, total RAM 1160740 MB
2025-07-17T15:31:10.826405240Z pytorch version: 2.7.1+cu126
Madiator2011
Madiator20114mo ago
ye it's not yet working looks like someone forces outdated pytorch @Salexes so I think I fixed bug pushing update and will let you know
(venv) root@6ee5828b0118:/# python
Python 3.12.3 (main, Jun 18 2025, 17:59:45) [GCC 13.3.0] on linux
Type "help", "copyright", "credits" or "license" for more information.
>>> import torch
>>> print(f'PyTorch version: {torch.__version__}')
PyTorch version: 2.7.1+cu128
(venv) root@6ee5828b0118:/# python
Python 3.12.3 (main, Jun 18 2025, 17:59:45) [GCC 13.3.0] on linux
Type "help", "copyright", "credits" or "license" for more information.
>>> import torch
>>> print(f'PyTorch version: {torch.__version__}')
PyTorch version: 2.7.1+cu128
it was installing CUDA 12.6 PyTORCH building new image now
ScaledEcom
ScaledEcomOP4mo ago
Ah got it that's why it did not work. Will wait for the new image and test once available
Madiator2011
Madiator20114mo ago
@Salexes
No description
ScaledEcom
ScaledEcomOP4mo ago
Nice! It works! How can I use your image now and use one custom node that I need? RUN comfy-node-install https://github.com/olduvai-jp/ComfyUI-HfLoader
GitHub
GitHub - olduvai-jp/ComfyUI-HfLoader
Contribute to olduvai-jp/ComfyUI-HfLoader development by creating an account on GitHub.
Madiator2011
Madiator20114mo ago
it's same as orginal repo
Madiator2011
Madiator20114mo ago
GitHub
worker-comfyui/docs/customization.md at main · runpod-workers/work...
ComfyUI as a serverless API on RunPod. Contribute to runpod-workers/worker-comfyui development by creating an account on GitHub.
Madiator2011
Madiator20114mo ago
just need to push updated base image to dockerhub
ScaledEcom
ScaledEcomOP4mo ago
So I can just do this, and replace with correct image ofcourse, correct?
FROM runpod/worker-comfyui:5.2.0-base

# Install your custom nodes
RUN pip install huggingface-hub
RUN comfy-node-install https://github.com/olduvai-jp/ComfyUI-HfLoader
FROM runpod/worker-comfyui:5.2.0-base

# Install your custom nodes
RUN pip install huggingface-hub
RUN comfy-node-install https://github.com/olduvai-jp/ComfyUI-HfLoader
Madiator2011
Madiator20114mo ago
also pushed flux dev version FROM runpod/worker-comfyui:5.2.0-base replace to: FROM madiator2011/worker-comfyui:5090-base though wait first till I push updated image @Salexes done make sure to pull latest base first
ScaledEcom
ScaledEcomOP4mo ago
regarding that, does the flux dev version use the full 23 GB flux version or is it the fp8 version (that is i think 12gb) ? Nice, will do! Thank you for the help with this, couldn't have done it without you!
Madiator2011
Madiator20114mo ago
np let me know if that works now for you
ScaledEcom
ScaledEcomOP4mo ago
Just tested it it works now!
Madiator2011
Madiator20114mo ago
btw pushed some images with models
ScaledEcom
ScaledEcomOP4mo ago
thats awesome thank you very much! So happy right now! now I can deep dive into testing further
Madiator2011
Madiator20114mo ago
@Salexes btw pushing right now the flux dev version https://hub.docker.com/r/madiator2011/worker-comfyui/tags
ScaledEcom
ScaledEcomOP4mo ago
Thats perfect, thank you
Madiator2011
Madiator20114mo ago
will take a little as it's chunky model
ScaledEcom
ScaledEcomOP4mo ago
@Elder Papa Madiator Thank you for all the help again. I am actively testing a lot of stuff right now. Based on your experience, what is better to do use lightweight base docker image without flux and have the flux model on network drive OR use the flux dev docker image that you made and use no network drive ? and how does flashboot come into play here? Which of these solutions would be more efficient on serverless? Would one be loading faster/be ready faster than the other? I am trying to optimize the total time it needs to run. _ I hope you don't mind me asking these questions, just trying to get a deeper understanding of what the best approach would be so I can apply stuff that way in the future as well
Unknown User
Unknown User4mo ago
Message Not Public
Sign In & Join Server To View
Madiator2011
Madiator20114mo ago
I pushed docker image with model in
ScaledEcom
ScaledEcomOP3mo ago
Hi @Madiator2011 I do have a question is it normal that with 20 workers available it does not make use of all of them automatically?
No description
ScaledEcom
ScaledEcomOP3mo ago
I noticed that happening sometimes now, that there is 20 workers available and there is like 24 jobs
Madiator2011
Madiator20113mo ago
yes it depends on your settings
ScaledEcom
ScaledEcomOP3mo ago
No description
ScaledEcom
ScaledEcomOP3mo ago
Is there a setting somewhere else than here where I can set amount of workers? Since I would have expected it to use all 20 workers when there is 24 requests coming in Oh now suddenly after 3 minutes it started to use at least 13 workers at once. Still wondering why it does not utiize all 20 immedeatly, could that be a setting on my end as well? I have set the delay to 2 seconds "Queue Delay Queue Delay scaling strategy adjusts worker numbers based on request wait times. With zero workers initially, the first request adds one worker. Subsequent requests add workers only after waiting in the queue for 2 seconds." Sorry for the message spam, just tried to give you the full picture
Unknown User
Unknown User3mo ago
Message Not Public
Sign In & Join Server To View
ScaledEcom
ScaledEcomOP3mo ago
Where can I do that? Same question for this
Unknown User
Unknown User3mo ago
Message Not Public
Sign In & Join Server To View
yhlong00000
yhlong000003mo ago
1. If you plan to use 20, I’d suggest set max worker to be 30~. 2. If you want worker to scale as quickly as possible, use request count, and set scaler to be 1.

Did you find this page helpful?