Hello everyone. I am Dr. Furkan Gözükara. PhD Computer Engineer. SECourses is a dedicated YouTube channel for the following topics : Tech, AI, News, Science, Robotics, Singularity, ComfyUI, SwarmUI, ML, Artificial Intelligence, Humanoid Robots, Wan 2.2, FLUX, Krea, Qwen Image, VLMs, Stable Diffusion
Hello Furkan, I hope you are well. As you may remember I have meticulously built both A1111 and Kohya_ss platforms according to your great tutorials. With my last LORA training I unfortunately get broken and corrupted images like you see in the enclosed examples. Can you please tell me what should I do to rectify this issue? Thank you.
The question i have with magic animate - if you use a photo of yourself and than the pre made dancing - would the resault's face looks like yours? or is it autogenerated and wouldnt look like it at all?
euler a does produce these, but it has to do with sampling step as well. in my case the last step creates these artificats. Also which lora output number are you using ? For me somewhere from 11-15 gives decent resuilts
DR.Siva thank you so much for your kind reply. I’ll implement your guiding points and will let you know of the outcome. What do you mean by lora output number? In my training parameters I used 6 epochs so I got 6 models, each of 147,571 KB.
Hello, after much frustration trying to work with a1111 and controlnets, using a1111 is an API backend service i come to you for help!
I just want someone who can link me json payload they send to a1111 running sdxl where the output makes sense compared to settings things up in the GUI: -json payload to a1111 for any sdxl checkpoint w. controlnet they use that works, img2img and or txt2img -also understanding in the api why i cant upload an image for controlNet for txt2img, but only in img2img
-> My problem is when using controlnet in my api call its like im doing regular img2img without the controlnet, and in the api call i cant upload my controlNet. In a1111 i can setup settings for perfect image generation for txt2img, w controlnet, but cant get the same results when running in --api mode, sending the json payload :S
a11 is horrible for API usage especially with third party plugins injecting, comfy UI albeit more messy to setup at first if youre not familiar, allows for much easier flow and has native API via dev mode
are there good guides for getting things started? Like heavy on the developer front? and i mean in general for cloud gpu / api shit / workflow everything. im looking for a hub/journy of learning which is why my question is maybe annoying, but thanks for input about comfy and api, usage. -> thoughts on where to start with comfy api mode, or i will just google and pray result number 1 is good. But thanks either way.
Anyway, last prayer for someone who uses a1111 with --api and can send me the payload they use for successfull image generation: I just want someone who can link me json payload they send to a1111 running sdxl where the output makes sense compared to settings things up in the GUI: -json payload to a1111 for any sdxl checkpoint w. controlnet they use that works, img2img and or txt2img -also understanding in the api why i cant upload an image for controlNet for txt2img, but only in img2img
-> feel free to delete, im just begging here with mouth open
Display the corresponding API payload after each generation on WebUI - GitHub - huchenlei/sd-webui-api-payload-display: Display the corresponding API payload after each generation on WebUI
magic animate seems to have a different color palette than the ones in the linked detectron repo also, or maybe it's just the fact the background is that shade of purple?
would need to also split a video into its images with ffmpeg in a previous step, then apply_net and then reassemble the output images it did looks like
I have installed and working. But there is literally 0 documentation and examples. in the same folder densepose_rcnn_R_101_FPN_DL_s1x.yaml located in the same folder Base-DensePose-RCNN-FPN.yaml lo...
Hello Furkan, DR. Siva replied the following to me: I am getting the same results , i was able to avoid by
check resolution 1024x1024 ( start with this res)
Decrease steps to 30
Use 2M karras sampler
Use some other models (ProtovisionXL,NightvisionXL) euler a does produce these, but it has to do with sampling step as well. in my case the last step creates these artificats. Also which lora output number are you using ? For me somewhere from 11-15 gives decent resuilts
The webui.zip is a binary distribution for people who can't install python and git. Everything is included - just double click run.bat to launch. No requirements apart from Windows 10. NVIDIA o...
im autist from euruope and i think this is sarcasm haha, but yes as a cs person programming shit is actually fucking difficult, like what the f is everything, i have no idea. docker, networkvolumes, gradio, api shit, errors browser. loool. pray for me. service layer.
wish my gpu was free right now, this seems promising and it has video input, maybe it does the apply_net step already since it mentions "show visualizations in an opecv window"
not sure if you went through that already however or if any of those models that are pretrained are the same one magic needs
3 days ago I was able to print 2048x2048 images with 4060 8gb vram, but now I can not print. If idle vram usage can help me in this case, thank you very much in advance. RuntimeError: Not enough memory, use lower resolution (max approx. 1600x1600). Need: 6.0GB free, Have:4.2GB free I'm dealing with this problem, do you know how to solve it.