this post was submitted on 19 Nov 2023
2 points (100.0% liked)

AMD

26 readers
1 users here now

For all things AMD; come talk about Ryzen, Radeon, Threadripper, EPYC, rumors, reviews, news and more.

founded 1 year ago
MODERATORS
 

https://www.youtube.com/watch?v=QEbI6v2oPvQ

I had a lot of trouble setting up ROCm and Automatic1111. I tried first with Docker, then natively and failed many times. Then I found this video. It has a good overview for the setup and a couple of critical bits that really helped me. Those were the reinstallation of compatible version of PyTorch and how to test if ROCm and pytorch are working. I still had a few of those Python problems that crop up when updating A1111, but a quick search in A1111 bug reports gave work arounds for those. And a strange HIP hardware error came at startup, but a simple reboot solved that.

Also he says he couldn't make it work with ROCm 5.7, but for me now 2 months later, ROCm 5.7 with 7900 XTX and Ubuntu 22.04 worked.

And coming from a Windows DirectML setup, the speed is heavenly.

you are viewing a single comment's thread
view the rest of the comments
[โ€“] [email protected] 1 points 11 months ago (1 children)

I did this setup for native use on Fedora 39 workstation about a week and a half ago, the amount of dicking about with python versions and venvs to get a compatible python+pytorch+rocm version together was a nightmare, 3 setups that pytorch site said were "supported" before it finally worked with rocm5.7. It was my first experience with setting it up natively, have ran a docker version in the past and will probably stick to that in the future.

[โ€“] [email protected] 1 points 11 months ago

Just use a Podman container with Distrobox