Hello everyone. I am Dr. Furkan Gözükara. PhD Computer Engineer. SECourses is a dedicated YouTube channel for the following topics : Tech, AI, News, Science, Robotics, Singularity, ComfyUI, SwarmUI, ML, Artificial Intelligence, Humanoid Robots, Wan 2.2, FLUX, Krea, Qwen Image, VLMs, Stable Diffusion
I’m planning to build a SaaS product based on some ComfyUI workflows I’ve created, using Flux as the main image generation model. I’ve spoken with the team at Fal.ai, and they confirmed I can use it commercially through their platform, as long as it runs through their API due to their arrangement with Blackforestlabs. They also mentioned they can host ComfyUI and that it’s possible to use ComfyUI as a backend for a SaaS product, as long as all models used in the workflow (like upscalers, facial detection, etc.) are licensed for commercial use. I’m still deciding whether to stick with ComfyUI or build a custom pipeline based on my workflows.
Hey Dr. Furka, my friend is training a flux character lora (not realistic) with default settings of ai-toolkit and I try training with kohya and your config rank 3, and he seems to get really good results, but I'm having a problem getting the proper body shape and some details properly. It seems like no matter what parameters I try, it feels undertrained. tried 1 repeat with 200 epochs (20 images) lr0.00005 tried 20 repeat with 10 epochs (20 images) lr0.0001 tried 150 repeat with 1 epoch, (save every 1000 steps) tried the output of 2000, 3000 ,4000, and unless I make the prompts very long and detailed, I'm getting something that resembles the character and clothes, but the body shape and face mostly gets random every seed and it won't really capture the character. Even if I put a lot of details in the prompt, it kinda works, but not really.
And yet, my friend's first attempt with default settings on ai-toolkit, which has much less parameters to change, gets awesome results and accuracy, and it's still very flexible while keeping the essence of the character pretty well.
forgot to mention, we trained the same character his captions are mostly automated with roughly manual fixes my captions were automated at first, then highly reconstructed manually in a consistent captioning way
Right now I'm trying to replicate his decent+ result of ai-toolkit, but with kohya_ss, just to have a solid starting point, but I can't manage to achieve that.
and even though the character is not accurate enough, I do find that my consistent captioning structure for the angle the character is standing and face expressions that are not the neutral ones works pretty well when I prompt for it so should I caption those?