Hello everyone. I am Dr. Furkan Gözükara. PhD Computer Engineer. SECourses is a dedicated YouTube channel for the following topics : Tech, AI, News, Science, Robotics, Singularity, ComfyUI, SwarmUI, ML, Artificial Intelligence, Humanoid Robots, Wan 2.2, FLUX, Krea, Qwen Image, VLMs, Stable Diffusion
I used a Learning rate of .001, thew results were much better! .001 might be high but because I'm doing parody I don't really need exact I just need to capture their likeness and make them something else
I need to know, the 16gb config, can i get it from you by any chance?? ..i would buy the Patreon way long ago, but for my country i could pay, because of the payment method restrictions.
I loaded the styles.csv into Automatic and have gotten some very interesting results using the updated Juggernaut 7 model. The advantage to this model is the refiner is built into the model, so you don’t need a separate refiner. The disadvantage to using the styles in Automatic is that you cannot combine styles in one image. If you select multiple styles you will get a batch for each style you choose.
In the Fooocus UI you can combine styles for more variety in your art. You can also use the same model by manually adding it. I personally like to select one style and generate a prompt to see the results before adding another style or switching styles. The Fooocus Masterpiece style is my favorite so far.
I take my images, load them into Corel Painter or Corel Paintshop Pro and adjust them, make them fit an 8.5 X 11 sheet of paper or 11 X 17 sheet for printing. I also may completely alter the image because I don’t want photographs even though I want to study photography in order to write good prompts. At the moment I borrow other people’s prompts and modify them. I don’t want to be reliant on someone else’s prompts forever though.
It seems that the general consensus for most AI developers is they want photographic art. That’s all well and good but they are really limiting themselves in my opinion as SD is capable of so much more.
That is a raw image straight from Fooocus UI using 3 styles but the prompt itself also specifies a style which overrides apparently the selected styles.
I tried it in comfyUI local install and it just seems buggy, the nodes are anyway, and then I realized the good Dr had already put together a better implementation. I have the Gradio space working right now in a runpod
@Dr. Furkan Gözükara , when training a SDXL model with 30 training images of a subject for a person's resemblance, after how many epochs do you expect the images to start looking like the person?
Hello. I'm currently trying to create a face lora after watching your video: https://www.youtube.com/watch?v=TpuDOsuKIBo&t=721s. The results are somewhat disappointing, possibly due to dataset issues, but it's definitely improved since I started experimenting on my own after watching the video. However, I'm wondering if cropping a person's face or full body shot from the background and turning it into a white background would be significantly beneficial? I actually feel like it's ruining the train and result looks weird right now
#Kohya SS web GUI DreamBooth #LoRA training full tutorial. You don't need technical knowledge to follow this tutorial. In this tutorial I have explained how to generate professional photo studio quality portrait / self images for free with Stable Diffusion training.
They are called the Herculaneum Scrolls, and since we are so hyped by AI and that is the key to unlocking the text it might be fun to do it even if we lose.
Has anyone tried training a LoRA or Dreambooth using PNGs of a subject with a transparent background (essentially subject cut out and nothing behind)? I've tried with white background and backgrounds of random colors, both successful to remove associating the training with the background and focus more on the subject, however im curious if anyone has already tried this and could speak about the results EDIT: When using ohwx man within a prompt, the random color background works quite well. However I just tried a simple "(ohwx man:1.3)" prompt and it mostly worked well however I did receive one photo that was just a purple background, so thats worth noting
From what I've seen online it seems .dat upscalers are the best. However I can't personally speak to this since those are not yet supported by A1111. But looking at this, https://openmodeldb.info/models/4x-FaceUpDAT it seems exceptional