embeddings
endpoint understanding
#443
Unanswered
wei-ann-Github
asked this question in
Q&A
Replies: 1 comment 3 replies
-
I think this is probably a bug. It should return a list of two embedding representation. |
Beta Was this translation helpful? Give feedback.
3 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Hi,
I am trying to understand the output of the embeddings endpoint, and how it relates to the requests.
Deployed model: Llama-2-7b-chat
Request platform: http://localhost:3000/#operations-Service_APIs-llm-llama-service__embeddings_v1
I am making the reuqests through the BentoServer UI.
In one example, I used
as the request body. The output I got is
When my request body is just a single string in the list:
[ "Hey Jude, welcome to the jungle!" ]
The size of the embeddings is much longer than the request consisting of 2 string:
How does one interprete the embeddings and use it if it gives different size?
I was expecting the output from using 2 input strings to consists of a list of 2 sets of embeddings, but this does not seem to be the case.
Beta Was this translation helpful? Give feedback.
All reactions