Replies: 1 comment
-
Simply adding |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
I was setting below
CUDA_VISIBLE_DEVICES="3" LLAMA_ARG_N_GPU_LAYERS=100 /workspace/llama.cpp/llama-server --verbose --verbose-prompt --log-enable --log-file '/workspace/model_serving/llamacpp' -np 17 -m '/workspace/llama.cpp/models/llama3.1-8B-Instruct.Q4_K_M.gguf' --host '0.0.0.0' --port 20022 -t 12 -b 256 -c 20000 -fa
llamacpp created "/workspace/model_serving/llama.log", "/workspace/model_serving/llamacpp.log"
but I found out any context in "workspace/model_serving/llamacpp.log"
[1727229882] warming up the model with an empty run
Please tell me how to log.
Beta Was this translation helpful? Give feedback.
All reactions