Hello everyone. I am Dr. Furkan Gözükara. PhD Computer Engineer. SECourses is a dedicated YouTube channel for the following topics : Tech, AI, News, Science, Robotics, Singularity, ComfyUI, SwarmUI, ML, Artificial Intelligence, Humanoid Robots, Wan 2.2, FLUX, Krea, Qwen Image, VLMs, Stable Diffusion
I get kind of close with Kling Lip Sync but it's not perfect and it's very tedious doing 5s videos and matching them together + Kling doesn't have infinite generation
It sort of works with Kling by using a few different 5s clips but it's kind of choppy, lips aren't perfect, have to stitch them together etc. Is there a better way?
Technique consisting in a new synthetically trained AI model [FLUX.D LORA], and some ComfyUI wizardry, with the objective of accurately reproducing a 'found footage/liminal' aesthetic.
Both music and visuals by myself.
You can access these [new FLUX.dev LORA] + ComfyUI workflow + 2350 images and prompts in metadata + 26 img-to-vid e...
Guys, what are your Wan2.1 img-to-vid 480 14B generation times like after the latest updates? On my 4060 Ti 16GB, I'm getting 22/50 [26:20<33:55, 72.71s/it]. Before, each generation used to take around 1 hour and 30 minutes.
Flux gym and kohya allow training from fine tuned models, ai-toolkit and maybe onetrainer require raw dev from huggingface...i believe. I haven't looked in awhile. I'm still using ai toolkit on cloud, but I also run q8 so don't really use a finetune.