Hello everyone. I am Dr. Furkan Gözükara. PhD Computer Engineer. SECourses is a dedicated YouTube channel for the following topics : Tech, AI, News, Science, Robotics, Singularity, ComfyUI, SwarmUI, ML, Artificial Intelligence, Humanoid Robots, Wan 2.2, FLUX, Krea, Qwen Image, VLMs, Stable Diffusion
Might be the prompt as well? I did the mistake adding man into my prompt when i wanted to replicate myself.. so instead of "ohwx man" just use your token.. in my case "ohwx" ..
@Dr. Furkan Gözükara My first test was a total success, I trained many people at the same time without bleeding between each other, It works perfect and the lora can be used on regular flux dev that is much faster on inference, I get a little bleeding with two subjects with a similar name "Diego man" "Dani man" is minimal so it can be fixed changing the name to "Daniel man" I saw a little class bleeding but I think it can be fixed using regularization images. the model behaves very similar to regular SDXL, Training many subjects at the same time was imposible with regular flux-dev. The model is still a little undertrained so I will continue and try with regularization of class people because my dataset contains people from different genders and ages. my captions are very simple, "name class" This model is awesome and very promising so far.
note: for this first test I used 512x512 resolution to speed up the training and I used a higher lr for the same reason 0.0001 for unet and TE, may be with lower lr and of couse at 1024x1024 will get better results
yes different images, but sdxl has the same problem to generate two different people in the same image, the good thing is that now I can train multiple subjects in one lora, I didn't try to generate two people in the same frame, I will try tomorrow
Are you doing a fine-tune then extracting a LORA? or is this a straight LORA training? I've been training finetunes/loras on family members, so having a bunch of subjects together with lower bleeding is a great use case for me. Even if there is still some bleeding, then I can segment or inpaint the faces if it can get the broad features correct.
I bought 2x used 3090 on the cheap. But I keep getting blue screens If I want to use both of them optimally. Please don't tell me that I have to upgrade my PSU as well
Note sure if this applies but if youre on massedcompute you need to copy model/clip/vae files into relevant swarm folders. Your checkpoint will go in diffussion_models, or the lora folder if that is what you trained in which case flux1_dev needs to be in diffusion_models
My GOD getting Krita AI Diffusion and SwarmUI to use the same comfyui installation as backend, then have a shared model folder for comfy AND swarmui ANd a1111 webui has been the most frustrating two day troubleshooting nightmare I've had in years
this a straight LORA training, this is my first test with the model flux-dev-de-distill, works much better than regular flux-dev for training, the only problem is the samples during training, don work correctly with this model because it uses distilled cfg and no regular cfg, but it doesn't interfere with the trainig.
oh, sorry I misunderstood. I see what you mean, when it generates samples from the lora for each x epoch, those aren't working correctly, so you don't see the progress?
yes, it generates images but look deformed because it uses real cfg 1, to display correctly you need the cfg at 3.5, the Distilled CFG Scale is disabled on this model so this parameter does nothing