Replies: 2 comments
-
That's an excellent question, and very much a research question at this point :) |
Beta Was this translation helpful? Give feedback.
0 replies
-
Implementing #479 may make it easier to answer this question. Note that you can already get the logits at each step by building the sequence generator with |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Hi everyone,
I am interested in the calibration of LLMs to ensure their trustworthiness. We saw with GPT4 paper that RLHF for instance destroyed calibrated answers.
I am interested in measuring calibration of models that were trained on general text generation and not classification specifically, e.g. GPT4.
For instance, imagine I want the model to do classification by using outlines to only output "Cat" or "Dog".
Can I use outline to provide an accurate measure of output probability without having to fine-tune the model for classification? Aka, does zero-ing all tokens that don't produce the class I am interested in, can be a good way to provide output probability similar to what I would get if I were to fine-tune a classification head?
Thanks for the help :)
Beta Was this translation helpful? Give feedback.
All reactions