Hello everyone. I am Dr. Furkan Gözükara. PhD Computer Engineer. SECourses is a dedicated YouTube channel for the following topics : Tech, AI, News, Science, Robotics, Singularity, ComfyUI, SwarmUI, ML, Artificial Intelligence, Humanoid Robots, Wan 2.2, FLUX, Krea, Qwen Image, VLMs, Stable Diffusion
Your results are very similar to mine. I wonder why these results are coming out and I want to find out how to solve them. Does anyone else have similar errors like this?
Get more from SECourses: Tutorials, Guides, Resources, Training, MidJourney, Voice Clone, TTS, ChatGPT, GPT, LLM, Scripts by Furkan Gözükara on Patreon
I just solved a problem I had. When CompyUI connected vae to vae decoder through the vae loader node, the image was created normally. I was using the config file you shared with Patreon.
I trained the model using your config file, but when I didn't connect the VAE loader, the images came out the same as before. Is it possible that the model didn't train properly? I used the config file you shared on Patreon and trained with normalized images.
Postive - cinematic photo Portrait of ohwx man as super Mario, red hat, fantasy, highly detailed, digital painting, artstation, concept art, sharp focus, illustration, art by Tony Sart and artgerm and randy vargas. 35mm photograph, film, bokeh, professional, 4k, highly detailed
After 30 training and 5 dataset retake to have the best one, i finally achieve the - p e r f e c t - result thanks to @Furkan Gözükara SECourses however, mid-distance or distance shot are Totally destroyed, my dataset is perfect, used cinema grade camera, used cropper and resizer, tried mask training, adetailer, changing the number of step only for face, changing the denoising strengh, i can only make p r e f e c t close-up, selfie or whatever, but when the subject is in midrange like your picture in civtai, or fullbody, it's totally crap, even the face is not really "human" more like destroyed, any idea ? thanks for everything (also tried with and without reg)
i have to admit i spent the last weeks understanding every little things possible and even myself cannot recognize a (fake) picture from myself, ,even my family, but im limited to close-up
i tried different dataset style, always perfect quality with different clothing and background
i found out that it is not the best idea to add different facial expression since when using a prompt like "surprised"
if the model use an already surprised expression from the dataset it would be SUPER exaggerated and very weird, even using moderate prompt like "slightly"
I suppose that it would not be an issue if every picture from dataset were captioned, but i havnt found a good tutorial about dataset/captioning and method to use
Also i found that not be able to modify face is a pretty sad, adding a cyborg eye, or facepaint, makup, blood, dirt whatever, tried a lot of prompt even in the adetailer face parameter
I also found out that if your dataset is always with a sharp focused talent and blurry backround, whatever prompt you use, it will always be the same in Stablediffusion, always a sharp focused person, but blury background (i may doing something wrong), i feel like the "from single text file" method is pretty good for something quick, but if you want to achieve God level of training, you need to do the Super anoying long version of data training method