-
Hi, I have trained two NeMo ASR model
According to my understanding the model for which the Could anyone give me any kind of a clue why the model behaves like this? Thanks |
Beta Was this translation helpful? Give feedback.
Replies: 2 comments 2 replies
-
For ASR, since we use heavy spec augment, lower loss does not correspond to lower wer always, since your mask can differ and the model is not trained for filling in text but to align sequences. If a seq is masked the loss is not fully correlated with the wer. |
Beta Was this translation helpful? Give feedback.
-
@titu1994, If we turn off the spec augmentation then the loss will be more correlated with the WER? |
Beta Was this translation helpful? Give feedback.
For ASR, since we use heavy spec augment, lower loss does not correspond to lower wer always, since your mask can differ and the model is not trained for filling in text but to align sequences.
If a seq is masked the loss is not fully correlated with the wer.