Hello everyone. I am Dr. Furkan Gözükara. PhD Computer Engineer. SECourses is a dedicated YouTube channel for the following topics : Tech, AI, News, Science, Robotics, Singularity, ComfyUI, SwarmUI, ML, Artificial Intelligence, Humanoid Robots, Wan 2.2, FLUX, Krea, Qwen Image, VLMs, Stable Diffusion
Question. I want to make talking avatars, but D-ID and Pipio are too expensive to make the investment worth it. 15 minutes worth of vid for several dollars? Is there a better or at least cheaper alternative?
Anyone notice in generating images that nomatter how you replace your promts or even delete all text and replace new compostion, after web ui has been reset and still the composition and style always generates the same? something like your stuck on the same seed. How to totally get away with that?
I trained my model in Textual Inversion on SD 2.1. Results were not that bad, but I decided to check SD 1.5 because I was able to do hypernetwork on sd 1.5 and not on 2.1. I used same images and filewords in 1.5 as I used in 2.1. My embeddings have comparable loss value every 500 steps. But it is learning bad. Samples are much worse than in 2.1. Any idea what options or anything that I can change to improve my results? I used 22 images maybe I need to use more?
Hello, it's me again. I trained the model based on a cup, and it gave me pretty good sample images, but when prompted after training it doesn't give me the result i wanted. I used ohwx mug as a instance prompt and photo of ohwx mug as a sample image prompt
What have i done wrong ? Any ideas how to get the result from the training ?