Latest release builds not using AMD GPU on windows #9256
Unanswered
The-Lord-of-Owls
asked this question in
Q&A
Replies: 2 comments 5 replies
-
Beta Was this translation helpful? Give feedback.
5 replies
-
I didn't compile, I used the prebuilt stuff in the releases page. I'll try doing a compile in a couple hours or in the morning and see if that works |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
For my setup I'm using the RX 7600xt, and a uncensored Llama 3.1 model. I'm trying to use the llama-server.exe to load the model and run it on the GPU.
System specs:
CPU: 6 core Ryzen 5 with max 12 threads
GPU RX 7600xt w/ 16gb vram
System RAM: 64gb
Model: https://huggingface.co/LWDCLS/DarkIdol-Llama-3.1-8B-Instruct-1.2-Uncensored-GGUF-IQ-Imatrix-Request
cli being used:
llama-server -m DarkIdol-Llama-3.1-8B-Instruct-1.2-Uncensored-Q8_0-imat.gguf --port 8080
Note: The model file is located next to the llama-server.exe file
Current problem: I am able to start up llama-server with the model loading and the server allows me to connect in my browser just fine. When typing something in I get a response as intended. The actual PROBLEM is that its only running it on the CPU and completely ignoring the GPU and I don't see any setting in the browser page to specify to use the GPU explicitly, Nor did I notice a cli option to give it either for this.
I know that it isn't using the GPU as the utilization stays at 1% in task manager and the memory usage is no different than it was before starting llama-server.
The builds I've tried from the latest release
llama-b3647-bin-win-openblas-x64.zip - it works but only runs on cpu( current build being attempted before writing this post )
llama-b3647-bin-win-avx2-x64.zip - it works but only runs on cpu
llama-b3647-bin-win-cuda-cu12.2.0-x64.zip - Does not load at all and gives a window error for not having nvidia stuff for cuda installed
llama-b3647-bin-win-vulkan-x64.zip - Starts up llama-server but eventually fails to load the model( error provided below )
Error from vulkan build:
Beta Was this translation helpful? Give feedback.
All reactions