-
I followed the same steps as in #5801 In my case, the output of SentenseTransformer and embedding is different. What I am doing wrong? The output from the
llama.cpp
|
Beta Was this translation helpful? Give feedback.
Answered by
danbev
Oct 22, 2024
Replies: 2 comments 2 replies
-
Looking forward to some inputs. thanks |
Beta Was this translation helpful? Give feedback.
0 replies
-
I'm not familiar with the SentenceTransformer but the llama-embedding example uses Euclidean/L2 normalization by default: --embd-normalize N normalisation for embendings (default: 2) (-1=none, 0=max absolute
int16, 1=taxicab, 2=euclidean, >2=p-norm) Perhaps you can try to disable the normalization in ./llama-embedding -m paraphrase-MiniLM-L6-23M-v2-F32.gguf -p "What is your age?" --embd-normalize -1 --verbose-prompt |
Beta Was this translation helpful? Give feedback.
2 replies
Answer selected by
raulod
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
I'm not familiar with the SentenceTransformer but the llama-embedding example uses Euclidean/L2 normalization by default:
Perhaps you can try to disable the normalization in
llama-embedding
like this:./llama-embedding -m paraphrase-MiniLM-L6-23M-v2-F32.gguf -p "What is your age?" --embd-normalize -1 --verbose-prompt