-
I have a The code tag is 5month old, but that shouldn be an issue i guess.
build instructions cmake -B /root/llama.cpp-build/ -S /root/llama.cpp/ \
-DGGML_CUDA=ON \
-DLLAMA_CUDA=ON \
-DLLAMA_BUILD_EXAMPLES=ON \
-DLLAMA_BUILD_TESTS=OFF \
&& cmake --build /root/llama.cpp-build/ --config Release --target install -- -j 10 |
Beta Was this translation helpful? Give feedback.
Replies: 2 comments
-
You will also need to specify the number of layers to be offloaded to the GPU by using |
Beta Was this translation helpful? Give feedback.
-
That worked perfectly. |
Beta Was this translation helpful? Give feedback.
You will also need to specify the number of layers to be offloaded to the GPU by using
--n-gpu-layers N
.More information can be found here.