https://www.youtube.com/watch?v=QEbI6v2oPvQ
I had a lot of trouble setting up ROCm and Automatic1111. I tried first with Docker, then natively and failed many times. Then I found this video. It has a good overview for the setup and a couple of critical bits that really helped me. Those were the reinstallation of compatible version of PyTorch and how to test if ROCm and pytorch are working. I still had a few of those Python problems that crop up when updating A1111, but a quick search in A1111 bug reports gave work arounds for those. And a strange HIP hardware error came at startup, but a simple reboot solved that.
Also he says he couldn’t make it work with ROCm 5.7, but for me now 2 months later, ROCm 5.7 with 7900 XTX and Ubuntu 22.04 worked.
And coming from a Windows DirectML setup, the speed is heavenly.
Maybe he should switch to Fedora since ROCM is in the official repositories: https://i.imgur.com/wvdXZdl.png
And nope i didn’t have to install AMD drivers everything runs on Mesa inside a Distrobox Fedora container : https://i.imgur.com/5t3ucSu.png
Looking good!
How do you come out of the python dependency hell in fedora?
When i to set up the python3.10 env and try to install the requirements, torchsde always complain about a specific numpy, python or pip versions.
save yourself the trouble and go team green if you’re serious about AI.
u still need to install their crappy drivers that take your system down with them on a daily basis :>
It works out of the box using vladmantic sd.Next?
i am looking for a guide that is based on latest version of rocm&pytorch etc. available? Is it running good?
So far I’m pretty happy with it. I can do 1024 resolution gens at 2.7 it/s. If I try to put more of them in a batch, then I might run out of memory, but compared to Windows and DirectML this is quite a bit faster and has better memory management.
I also tried some animatediff for the first time on this, but only managed to render a 256 resolution gif. Even 512 resolution caused a crash.
I also managed to get ComfyUI setup to serve Krita as a Stable Diffusion backend, but I only just got it to work and don’t have the first clue about how to use it properly yet. I used this plugin: https://github.com/Acly/krita-ai-diffusion.
It should not be this involved. It is still a cluster of a process. But I hope some folks can get this to work.
It could be way easier with proper Docker images. That’s what I tend to do for all these projects.
ROCM team had the good idea to release Ubuntu image with the whole SDK & runtime pre-installed. But that’s simply not enough to conquer the market and gain trust. Ideally, they’d release images bundled with some of the most popular FLOSS ML tools ready to use and the latest stable ROCm version.
Are there reliable Docker images for Oobabooga, A1111, Silly Tavern etc?
However, you’ll often find repos with Docker scripts, some of them with ROCm.
For example.
Not that I’ve seen, at least no with ROCm pre-installed.
I did this setup for native use on Fedora 39 workstation about a week and a half ago, the amount of dicking about with python versions and venvs to get a compatible python+pytorch+rocm version together was a nightmare, 3 setups that pytorch site said were “supported” before it finally worked with rocm5.7. It was my first experience with setting it up natively, have ran a docker version in the past and will probably stick to that in the future.
Just use a Podman container with Distrobox