Hello everyone. I am Dr. Furkan Gözükara. PhD Computer Engineer. SECourses is a dedicated YouTube channel for the following topics : Tech, AI, News, Science, Robotics, Singularity, ComfyUI, SwarmUI, ML, Artificial Intelligence, Humanoid Robots, Wan 2.2, FLUX, Krea, Qwen Image, VLMs, Stable Diffusion
I did my first test, very promising results, a little undertrained at the moment, at the end I resumed the training with regularization images still no bleeding, now I going to start a new training from scratch 11 people at the same time, family and friends, I going to use regularization images from the class person because I'm training people form different ages and genders the regularization images are captioned I downloaded the dataset including captions a time ago it contains very diverse people from different races and ages from baby to elderly people, male and female. the captions of my subjects are simple "name class" lets see how it behaves
I try to train myself but before that I need more info. Normally when I train a lora of one person if the background have some people behind, somehow the image of the person I train bleed to the background into the people. So in your theory, by training with open dev flux checkpoint we can eliminate this problem right ?
The most important thing and the main difference with this model is that I can train multiple people of the same class without bleeding, with regular flux dev I I trained 3 man for example and the model mixed all the faces together, and on inference prompting for one person returns a mix of all, now the problem is gone with flux-dev-de-distill, it behaves like sdxl but with much better quality. I still experimenting but is very promising so far.
One at a time works perfect, two or more people on the same frame loose likeness, can be fixed with inpainting, man and woman works much better, the same problem that sdxl have, but remember my lora is still undertrained, lets see with my new lora... I will start training later, it will take 2 days or more to finish
From the comparison posts on reddit, q8 seems to keep a similar consistency to fp16. So if we could use it for training that would be super. But that doesn't seem possible atm.
I'm looking to do a finetune with a fp8 but the Dr. setting for it uses shared memory, and is slow at around 10s/it. But speaking to koyha, it seems its an intended behavior to use shared memory.
when training with my local pc, I need 11 hours of 3090 run full throttle, my poor pc fan roar the hell of of my room and the temperature turn up a lot
I'm training this model now, Works much better for training, you can train multiple people on one lora without concept mixing, it behaves very much like sdxl, much more flexible
Is it possible to create a finetune, then extract the LORA? I've had good luck with that for single subjects, and sometimes it's also nice to have a checkpoint as well.