i got pretty impressive results with batch 2, which drove SDXL adamW8bit training by by 30% and bs 4

i got pretty impressive results with batch 2, which drove SDXL adamW8bit training by by 30% and bs 4 by 60%. For me it's somewhere in there. Need to check flexibility, but i am sure there are other experimental variables to test. Here's dataset of 12 images, vs xy plot with batch 2 up to 9 epochs (5000 steps), 5 hours vs 9 hours batch size 1
image.png
xyz_grid-0007-20231006232009-__lora_4du7_0103-lora_adamw8bit_40rep2bs-dim128-alpha64-000001_0.8___closeup_portrait_of_4du7_woman_green_studio_background_st.jpg
Was this page helpful?