Hello everyone. I am Dr. Furkan Gözükara. PhD Computer Engineer. SECourses is a dedicated YouTube channel for the following topics : Tech, AI, News, Science, Robotics, Singularity, ComfyUI, SwarmUI, ML, Artificial Intelligence, Humanoid Robots, Wan 2.2, FLUX, Krea, Qwen Image, VLMs, Stable Diffusion
Just a quick FYI for anyone interested, here's how long it took me to do a 20img/150 epoch training session on a 4090 this morning. Used 'Rank 4 best config better colours' script
While you guys are training Flux, I'm experiencing the interesting features of PonyRealism 2.1 under Prodigy. (I see that the Flux training is still being tweaked at Kohya, so I'll wait.) It turns out that the SDXL rule of only training Unet doesn't apply here, because then the finished Lora will only look a little like the person. However, if I train normally (i.e. without the --network_train_unet_only switch), I'm already on target. By using the 'fp8 base training' switch I can weaken the strength of the model a bit, i.e. sometimes it can come in handy in achieving a more accurate Lora. Without fp8, the Lora is stronger, so the d_coef switch needs to be reduced.
seems to be a big training effort, not just remerge: "Hello everyone! An early version of RealVisXL V5.0 (560k steps) is now available on Hugging Face. The model is still in training, and the final version will be released on Hugging Face and CivitAI at the end of August. You can leave your feedback on this version in the comments below."
@Dr. Furkan Gözükara Bro, I am going to sleep now its 6am haha, training loras all night, I fixed all my machines now. Anyway, I want to Train 1000 images (a style of art) I want to train that but doing it on 1 A6000 it would take like a week right? What system you recommend and how much would it cost me on Mass compute? 4-8 GPU A6000? If its $3-5 an HR I am down to pay it. Please do recommend the correct VM to use on there and do you have a video setting one up as I only use local, never rented hardware before. Let me know thanks and gn brother! Totally worth the Patreon fee!
@Dr. Furkan Gözükara another question, if I want to use custom nodes in swarm UI, I did not see the folder, do I have to create one? Autocompletions Users-log.ldb clip_vision tensorrt Embeddings Users.ldb comfy-auto-model.yaml unet Lora VAE controlnet upscale_models Stable-Diffusion clip diffusion_models, or where are the custom node folder?
@Dr. Furkan Gözükara I've test on windows on my PC with A4000 (8go) and it worked, however as I suspected it use shared memory during the step "move text encoder to gpu" as there is not enough memory on the GPU, that's why it OOM on linux, because it can't use shared memory, then it's passed back to the cpu freeing memory and the rest is loaded that require less than 8go so during training it's stay under 8go. Is there a way to not move the text encoder to gpu, like having it calculated on cpu directly, having it use shared memory or making it use less VRAM, so it will work also on linux ?
Replace SD3Tokenizer with the original CLIP-L/G/T5 tokenizers. Extend the max token length to 256 for T5XXL. Refactor caching for latents. Refactor caching for Text Encoder outputs Extract arch...