I did this setup for native use on Fedora 39 workstation about a week and a half ago, the amount of dicking about with python versions and venvs to get a compatible python+pytorch+rocm version together was a nightmare, 3 setups that pytorch site said were "supported" before it finally worked with rocm5.7. It was my first experience with setting it up natively, have ran a docker version in the past and will probably stick to that in the future.
AMD
For all things AMD; come talk about Ryzen, Radeon, Threadripper, EPYC, rumors, reviews, news and more.
Just use a Podman container with Distrobox
It works out of the box using vladmantic sd.Next?
i am looking for a guide that is based on latest version of rocm&pytorch etc. available? Is it running good?
So far I'm pretty happy with it. I can do 1024 resolution gens at 2.7 it/s. If I try to put more of them in a batch, then I might run out of memory, but compared to Windows and DirectML this is quite a bit faster and has better memory management.
I also tried some animatediff for the first time on this, but only managed to render a 256 resolution gif. Even 512 resolution caused a crash.
I also managed to get ComfyUI setup to serve Krita as a Stable Diffusion backend, but I only just got it to work and don't have the first clue about how to use it properly yet. I used this plugin: https://github.com/Acly/krita-ai-diffusion.
It should not be this involved. It is still a cluster of a process. But I hope some folks can get this to work.
It could be way easier with proper Docker images. That's what I tend to do for all these projects.
ROCM team had the good idea to release Ubuntu image with the whole SDK & runtime pre-installed. But that's simply not enough to conquer the market and gain trust. Ideally, they'd release images bundled with some of the most popular FLOSS ML tools ready to use and the latest stable ROCm version.
Are there reliable Docker images for Oobabooga, A1111, Silly Tavern etc?
However, you'll often find repos with Docker scripts, some of them with ROCm.
For example.
Not that I've seen, at least no with ROCm pre-installed.
Maybe he should switch to Fedora since ROCM is in the official repositories: https://i.imgur.com/wvdXZdl.png
And nope i didn't have to install AMD drivers everything runs on Mesa inside a Distrobox Fedora container : https://i.imgur.com/5t3ucSu.png
Looking good!
How do you come out of the python dependency hell in fedora?
When i to set up the python3.10 env and try to install the requirements, torchsde always complain about a specific numpy, python or pip versions.
save yourself the trouble and go team green if you're serious about AI.
u still need to install their crappy drivers that take your system down with them on a daily basis :>