Hello everyone. I am Dr. Furkan Gözükara. PhD Computer Engineer. SECourses is a dedicated YouTube channel for the following topics : Tech, AI, News, Science, Robotics, Singularity, ComfyUI, SwarmUI, ML, Artificial Intelligence, Humanoid Robots, Wan 2.2, FLUX, Krea, Qwen Image, VLMs, Stable Diffusion
Oh, i see, yeah, let me do it as well. potential, means, when i run training, it doesnt take load on gpu it should take....usually, when the temperature gets higher, when the gpu is totally using it's power, but when i run training, with kohya, it only gets half of the temperature, compared to it usually takes.
Fixed to work tools/convert_diffusers20_original_sd.py. Thanks to Disty0! PR #1016 The issues in multi-GPU training are fixed. Thanks to Isotr0py! PR #989 and #1000
The only thing that broke, it seems, was Blip captioning, but as I recall, I was using some type of rollback version of something. I think it was transformers. Honestly most of these captions tools are lousy and if you are using a smaller data-set manual captioning is best
Can I do multi-gpu training in Koyha if the graphics cards are different models/different Vram? (example, RTX 3080 and RTX 3090). Sorry to ask before testing, it's that I need a new power supply if I want to use both cards.
Does anyone know the Syntax for the regional prompter well. I'm getting better with prompts but if I have two Character Loras what's the best way to set up the prompt?
I'm getting [virtualMemoryBuffer.cpp::nvinfer1::StdVirtualMemoryBufferImpl::resizePhysical::140] Error Code 2: OutofMemory when attempting to generate the TensorRT Default Engine
I have 12.3 installed, but it is not set in Path except that CUDA_PATH_V12_3 is set to that version. I installed 12.3 because it was the only way to check the samples with Visual Studio 2022.
Updating torch, torchvision, and xformers tossed another error saying torch could not access the GPU. It said I can set a flag, so it won't run the check, but I don't want to do that.
Your install program reinstalled cu118, so I'm assuming it must be that version which means I'm running the correct version of CUDA, so I don't think TensorRT is working at the moment.
So it is saying UNSUPPORTED_STATESkipping tactic...insufficient memory. Is there a launch setting such as --medvram I need to use? I haven't tried that yet, but that's next.
It is telling me what the problem is, but I'm not seeing it for some reason. Reading the error generated more carefully made it clear, but I'm not sure what to do about it. I set the launcher to --lowvram even though I have 12GB of it, but that still may not be enough. You recommended we choose a machine with at least 24GB of VRAM on RunPod if we are using it.
I was following your tutorial for Kohya, but if I can't even load Tensor then I highly doubt I can run that. The tutorial mentions it can be run on 12GB of VRAM, but I probably won't struggle with it to get it to work.