-
hi all, I have a doubt on the transfer learning part of the course where Feature extraction has to happen before we do the finetuning as it will be needing the same model weights to optimize further. Now what if I want to fine tune second time/third time? Which model to choose?
@mrdbourke , Please suggest which would be a correct approach. |
Beta Was this translation helpful? Give feedback.
Replies: 2 comments 3 replies
-
Theoretically, if you use an approach that can dynamically change the learning rate, there is no need for fine-tuning multiple times. But if you fine-tuning on fixed learning rate, then lowering the learning rate on every attempt might yield better results. Also you don't change or choose any different model, you simply recompile it. |
Beta Was this translation helpful? Give feedback.
-
Hey there, Your thoughts here would line up with how I would do it:
In essence:
etc... Though where the best results will come from is a little bit of guess and check. Jeremy Howard explains this methodology quite well in https://arxiv.org/abs/1801.06146 (see the ablation studies for different fine-tuning methods). @realnihal's explanation above could also get quite good results, see https://www.tensorflow.org/api_docs/python/tf/keras/callbacks/LearningRateScheduler to dynamically lower the learning rate during training. |
Beta Was this translation helpful? Give feedback.
Hey there,
Your thoughts here would line up with how I would do it:
In essence:
etc...
Though where the best results will come from is a little bit of guess and check.
Jeremy Howard explains this methodology quite well in https://arxiv.org/abs/1801.06146 (see the ablation studies for different fine-tuning methods).
@realnihal's explanation ab…