Hello everyone. I am Dr. Furkan Gözükara. PhD Computer Engineer. SECourses is a dedicated YouTube channel for the following topics : Tech, AI, News, Science, Robotics, Singularity, ComfyUI, SwarmUI, ML, Artificial Intelligence, Humanoid Robots, Wan 2.2, FLUX, Krea, Qwen Image, VLMs, Stable Diffusion
Good morning 3 days ago I was able to print 2048x2048 images with 4060 8gb vram, but now I can not print. If idle vram usage can help me in this case, thank you very much in advance.
Sorry for taking so long to get back to you. Upon looking into it more online, it seems it is an error with if your first engine is not default. After youve made your first with default settings it will allegedly work. Ill edit this comment if it works. EDIT: This was the solution.
Any idea how to prevent this: UNSUPPORTED_STATESkipping tactic 0 due to insufficient memory on requested size of 7461533184 detected for tactic 0x00 Attempting to do 1536x1024. I have 12gb vram, do you think medvram could fix it?
Hello Furkan, I hope you are well. As you may remember I have meticulously built both A1111 and Kohya_ss platforms according to your great tutorials. With my last LORA training I unfortunately get broken and corrupted images like you see in the enclosed examples. Can you please tell me what should I do to rectify this issue? Thank you.
The question i have with magic animate - if you use a photo of yourself and than the pre made dancing - would the resault's face looks like yours? or is it autogenerated and wouldnt look like it at all?
euler a does produce these, but it has to do with sampling step as well. in my case the last step creates these artificats. Also which lora output number are you using ? For me somewhere from 11-15 gives decent resuilts
DR.Siva thank you so much for your kind reply. I’ll implement your guiding points and will let you know of the outcome. What do you mean by lora output number? In my training parameters I used 6 epochs so I got 6 models, each of 147,571 KB.
Hello, after much frustration trying to work with a1111 and controlnets, using a1111 is an API backend service i come to you for help!
I just want someone who can link me json payload they send to a1111 running sdxl where the output makes sense compared to settings things up in the GUI: -json payload to a1111 for any sdxl checkpoint w. controlnet they use that works, img2img and or txt2img -also understanding in the api why i cant upload an image for controlNet for txt2img, but only in img2img
-> My problem is when using controlnet in my api call its like im doing regular img2img without the controlnet, and in the api call i cant upload my controlNet. In a1111 i can setup settings for perfect image generation for txt2img, w controlnet, but cant get the same results when running in --api mode, sending the json payload :S
a11 is horrible for API usage especially with third party plugins injecting, comfy UI albeit more messy to setup at first if youre not familiar, allows for much easier flow and has native API via dev mode
are there good guides for getting things started? Like heavy on the developer front? and i mean in general for cloud gpu / api shit / workflow everything. im looking for a hub/journy of learning which is why my question is maybe annoying, but thanks for input about comfy and api, usage. -> thoughts on where to start with comfy api mode, or i will just google and pray result number 1 is good. But thanks either way.
Anyway, last prayer for someone who uses a1111 with --api and can send me the payload they use for successfull image generation: I just want someone who can link me json payload they send to a1111 running sdxl where the output makes sense compared to settings things up in the GUI: -json payload to a1111 for any sdxl checkpoint w. controlnet they use that works, img2img and or txt2img -also understanding in the api why i cant upload an image for controlNet for txt2img, but only in img2img
-> feel free to delete, im just begging here with mouth open
Display the corresponding API payload after each generation on WebUI - GitHub - huchenlei/sd-webui-api-payload-display: Display the corresponding API payload after each generation on WebUI
magic animate seems to have a different color palette than the ones in the linked detectron repo also, or maybe it's just the fact the background is that shade of purple?
would need to also split a video into its images with ffmpeg in a previous step, then apply_net and then reassemble the output images it did looks like
I have installed and working. But there is literally 0 documentation and examples. in the same folder densepose_rcnn_R_101_FPN_DL_s1x.yaml located in the same folder Base-DensePose-RCNN-FPN.yaml lo...
Hello Furkan, DR. Siva replied the following to me: I am getting the same results , i was able to avoid by
check resolution 1024x1024 ( start with this res)
Decrease steps to 30
Use 2M karras sampler
Use some other models (ProtovisionXL,NightvisionXL) euler a does produce these, but it has to do with sampling step as well. in my case the last step creates these artificats. Also which lora output number are you using ? For me somewhere from 11-15 gives decent resuilts