How to set up learning rate starting being decayed at the beginning of a training process #4875
nghiahuynh-ai
started this conversation in
General
Replies: 1 comment
-
Hi @nghiahuynh-ai just to clear up you want to continue the LR schedule when you restart training? it seems you're using
This will ensure the step etc are reloaded. Any chance you can use the example scripts provided, such as this? This makes it even easier to enabling checkpoint resume from the cmdline, just passing these flags:
With these flags, NeMo will automatically resume if a checkpoint is found in the directory, handling all the state management for you! |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
I have an issue that I want to train a model (Conformer-Transducer) in 2 stages. First stage, the learning rate raises only (just warm up). Second stage, I want the learning rate to be decayed. I used load_from_checkpoint method to load the previous checkpoint and set up all parts needed. I found that the learning rate have been warm up again. So, how to set up the learning rate starting being decayed at the beginning of second stage.
Thank for regard.
Beta Was this translation helpful? Give feedback.
All reactions