Hello everyone. I am Dr. Furkan Gözükara. PhD Computer Engineer. SECourses is a dedicated YouTube channel for the following topics : Tech, AI, News, Science, Robotics, Singularity, ComfyUI, SwarmUI, ML, Artificial Intelligence, Humanoid Robots, Wan 2.2, FLUX, Krea, Qwen Image, VLMs, Stable Diffusion
Get more from SECourses: Tutorials, Guides, Resources, Training, FLUX, MidJourney, Voice Clone, TTS, ChatGPT, GPT, LLM, Scripts by Furkan Gözü on Patreon
Everything like "anxious"? Like, if you put in a picture of yourself looking worried about something (dunno if you're that good of an actor ), do you think the model would pick it up and be able to reproduce it when the prompt asks for it?
repeats are a kohya thing, 1 epoch in the general context of ML usually means going once through your dataset. And look here, 2,5 minutes. Even if you used multiple GPUs, it cannot mean much more than going through your dataset once with the s/it I have seen so far on flux
I have question about training and captions please : How many images you think, for a Pose lora including 2 or multiple characters should be in the folder, and repeats ? And also my captionning, can I send one here with the related image ? It is Sfw but may contain violence, soo i prefer to ask ^^
sooo...I've tried some lora trainings recently. What can I say. Sometimes results are pretty good on 70-80 epochs (once on 50 epochs) and sometimes 200 epochs are not good enough. I guess it all depends on dataset.
and training on realistic dataset will not give you any cartoonish (comics) style output. Because it's too overtrained on realistic images. I did sample check every 40 epochs (5 checks during training) and after 80th epoch cartoon style becomes realistic
Hi @Furkan Gözükara SECourses can you create another train preset using 24gb Vram to train fp16 Lora. I trained using Rank_3_18950MB. The lora is fp8 compare with fp16 training in your Rank 1 or Rank 2 presets, the prompt understanding of the fp8 lora is reduce and the result of fp8 is so far not satisfy my need
I'm just saying that even during inference a low-precision base model doesn't make a difference, so how would it during training. The Lora weights are still bf16+