Config file not being read #1139
Unanswered
gurbindersingh
asked this question in
Q&A
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
I've set up LocalAI on a Ubuntu 22.04 machine with a Nvidia GPU following the easy setup guide. But it looks like there is no GPU offloading.
I can see the following message in the logs
offloading 0/43 layers to the GPU
andVRAM used: 0MB
.It took a while to figure out that you need the
gpu_layers
parameter in the config file since that's not mentioned in the guide. But I've now created the config file in the same directory as the model (official Llama 2 model converted using Llama.cpp). This is what it looks like:But it doesn't seem like these configs are being read, even after restarting (and even rebuilding) the container. I am not sure if this is a bug or if I'm doing something wrong since the docs don't mention which of the properties in the configs are required and which are not.
Edit: I can run the model just fine with GPU offloading using Llama.cpp at 70 tokens/s. With LocalAI it's only running at 5-7 tokens/s.
Beta Was this translation helpful? Give feedback.
All reactions