You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
first of all: thank you for this great piece of software :)
I am trying to create embeddings with gguf models, such as Phi or Mistral. However, all my attempts to serve them via LocalAI and the LLama backend fail.
I am using the localai/localai:v2.5.1-cublas-cuda12 docker image
name: "phi-embeddings"
license: "Apache 2.0"
urls:
- https://huggingface.co/TheBloke/phi-2-GGUF
description: |
Phi model that can be used for embeddings
config_file: |
parameters:
model: phi-2.Q8_0.gguf
backend: llama
embeddings: true
files:
- filename: "phi-2.Q8_0.gguf"
sha256: "26a44c5a2bc22f33a1271cdf1accb689028141a6cb12e97671740a9803d23c63"
uri: "https://huggingface.co/TheBloke/phi-2-GGUF/resolve/main/phi-2.Q8_0.gguf"
Am i doing anything wrong or is this a bug? With the same models, I can create embeddings locally by using the llama-cpp-python bindings without problems. Any help solving this problem would be greatly appreciated.
reacted with thumbs up emoji reacted with thumbs down emoji reacted with laugh emoji reacted with hooray emoji reacted with confused emoji reacted with heart emoji reacted with rocket emoji reacted with eyes emoji
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Hi everyone,
first of all: thank you for this great piece of software :)
I am trying to create embeddings with gguf models, such as Phi or Mistral. However, all my attempts to serve them via LocalAI and the LLama backend fail.
I am using the localai/localai:v2.5.1-cublas-cuda12 docker image
Here is the http error response:
This is the debug output from the localai docker image:
This is the configuration file for the model.
https://raw.githubusercontent.com/fdewes/model_gallery/main/phi_embeddings.yaml
Am i doing anything wrong or is this a bug? With the same models, I can create embeddings locally by using the llama-cpp-python bindings without problems. Any help solving this problem would be greatly appreciated.
Beta Was this translation helpful? Give feedback.
All reactions