ch06 - fine-tuning an LLM for binary classification task - add vs update output layer #192
-
I am following the example from ch06 for fine-tuning an LLM for classification task. When I run the following code from the example, it doesn't update the layer but add a new
|
Beta Was this translation helpful? Give feedback.
Replies: 2 comments 29 replies
-
Hi there, yes, you are right, there shouldn't be a second Instead of
you probably need to do
|
Beta Was this translation helpful? Give feedback.
-
So, I am trying to use And, then I implement the training loop:
I keep running into could you share how I can go about implementing |
Beta Was this translation helpful? Give feedback.
Hi there,
yes, you are right, there shouldn't be a second
lm_head
. My guess is that's because the base model has a weird implementation where they have 2 nested models.Instead of
you probably need to do