Possible FR: Server endpoint for metadata in loaded model #9341
BradHutchings
started this conversation in
Ideas
Replies: 3 comments 2 replies
-
I think I just found the right way to do this...
|
Beta Was this translation helpful? Give feedback.
0 replies
-
This works as expected.
|
Beta Was this translation helpful? Give feedback.
2 replies
-
Here's some code for to patch the
|
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
When the server starts up, it prints out metadata from the model it loads. For example:
I would like an endpoint that puts all the key/value pairs into a JSON block. This would be especially helpful for getting the correct English name the model developer intends into a client UI. Such attribution is requested by many of the popular models on Huggingface.
I am going to try to make this work. I need to create a new endpoint in server.cpp. I need to find the code that spits out the k/v pairs and repurpose it to spit out JSON. I'll update my implementation in comments.
I also want to make this work for the older version of llama.cpp that is in llamafile.
I'd appreciate any help on this in comments! When we get it working, I'll make an FR Issue.
-Brad
Brad Hutchings
brad@DemoMachine.net
Beta Was this translation helpful? Give feedback.
All reactions