• 2 Posts
  • 54 Comments
Joined 7 months ago
cake
Cake day: April 27th, 2024

help-circle


  • As far as I was aware AMDGPU is used by default on most if not all distros so I still think its a bit confusing to decide which to use for your needs at first unless you check the video I linked. I noticed a performance increase after forcing RADV on NixOS so not really sure.

    Sidenote why cant AMDGPU and RADV combine their efforts to simplify and rename AMDGPU-PRO to AMDGPU-unfree because that itself is confusing since most people will be drawn to use the PRO version without realising the worse performance.
















  • Decent, for now until ROCM & ZLUDA improve. I use NixOS and run my AI stuff using docker containers as its the easiest way imo because of how fucked up the dependencies are for ROCM especially.

    Basically to get AMD working for this stuff right now is to make sure certain versions of ROCM for certain versions of projects interacting with certain versions of pytorch all like each other. The most dependency hell of all dependency hells.

    So most projects have a hell of a time supporting ROCM so you must use alternative forks mostly, and even if there is a ROCM version it is so hardly used that no one knows if it works or if it doesnt half the time. I will say you will have the EASIEST time by far if you use a 7900 XT because most things are built to support that card. Otherwise good luck. Get used to using environment variables such as:

    HSA_OVERRIDE_GFX_VERSION=11.0.0 (Or 10.3.0 if that one doesnt work. I use 11.0.1, these are codes for GPU’s supported by ROCM incase yours isnt supported)

    TL:DR - its all a big mess right now but it does work if you fuck with it a bunch, I got my 7800 XT to work nicely with Ollama + OpenWebUI for text generation. For stable diffusion its definitely a shit show atleast for my preferred UI Invoke AI. Doesnt work at all it only uses my CPU (also AMD so maybe some fuckery.) However I dont regret it as AMD is truly the best especially on Linux but definitely not for AI as it currently stands.



  • jaxiiruff@lemmy.ziptoLinux@lemmy.mlZLUDA's third life
    link
    fedilink
    arrow-up
    11
    arrow-down
    1
    ·
    1 month ago

    Oh hell yeah this will save us AMD users so much headache, I really dont understand why it was taken down recently.

    From what I understood it was AMD having a problem with ZLUDA but now suddenly its okay again? Whatever as long as ROCM is improved or benefitted by this im happy.