-
Notifications
You must be signed in to change notification settings - Fork 1.1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
GPU offloading not working on system with AMD 5900HX CPU #386
Comments
Hey I believe integrated GPU are not supported, probably better to run on CPU at this time, by passing -ngl 0 instead of 9999 Also I have seen a few open issues with the same error/warning at the start when using AMD, so am not sure if I should open a new issues. Perhaps there could be an Operating System check before looking for amdclang++ or clang++. |
OK. I was curious to know whether additional acceleration could be obtained by combining the iGPU with the CPU. In any case, I reckon the docs should explicitly state that AMD iGPUs are not supported (if they're not). Ideally, llamafile should also report this at runtime.
Yes. I copied clang++.exe to amdclang++.exe to overcome this. Both executables were in the $PATH, but the get_rocm_bin_path: note: amdclang++.exe not found on $PATH message still appeared. |
This bug still persists in 0.8.4 which makes AMD GPU offloading impossible on Windows right now :( |
You might want to refer to my experience on #188 . In short, iGPU could work with 0.8.4 as long as:
|
On 0.8.4 tinyBLAS works again on my RX6600 GPU which is great because I can run AI models with proper speed again! However the bug that llamafile is searching for amdclang++.exe still persists :( |
At least with the use of But we need to add the possibility to use this config with llamafile. UPDATE(13/08/2024) : with linux kernel 6.10+ look kernel change allow the use of GGT dynamic RAM and not only VRAM on some AMD APU. It is the case for my Ryzen 7940HS.) |
IIUC, GPU are not getting used on |
I'm running llamafile 0.8.1 on a Windows 10 mini PC with a AMD Ryzen 9 5900HX CPU
CPU Architecture: AMD Cezanne (Zen 3, Ryzen 5000)
GPU: AMD Radeon RX Vega 8
The mini PC has 64GB RAM installed.
When I enable llamafile GPU support with -ngl 9999, it exits with the error
My command line is:
llamafile-0.8.1.exe -ngl 9999 -m dolphin-2.9-llama3-8b-Q5_K_M.gguf
I have also tried re-running after installing the AMD HIP SDK but this made no difference.
Contrary to the runtime messages, amdclang++.exe was in my Windows PATH.
The text was updated successfully, but these errors were encountered: