I’m using the DirectML fork of A1111 because it’s the only one I can get to work.
For professional reasons, I have to have an AMD gpu. I had a 6650 and was able to upgrade to the 6700xt for the extra vram, but the change has made no difference in errors telling me I’m out of vram.
I am a fairly frustrated because it seems I’m locked out of a lot of really neat and powerful features. Generating a batch of 4 images at 512x512 already takes a couple minutes. Moving that resolution at all jumps up the time considerably. I can do very little with img2img and ControlNet is effectively useless for me.
So, now that I’m done whining, is there any news about AMD improvements that might bring performance up to even a decent level with comparable Nvidia cards?
I would guess it’s one of three things:
- You are using Windows instead of Linux for ROCm (I don’t know how much this affects performance, as I am only on Linux)
- You are generating the batch all at the same time, instead of just doing multiple generations. This can lead to out-of-memory issues, as it is a larger image being generated.
- You are not fully using the GPU. Does task manager say 100% utilization?