Adding KenLM degrades performance #3384
Unanswered
errorfixrepeat
asked this question in
Q&A
Replies: 1 comment 4 replies
-
Seems like your LM is not performing well. Maybe the data distribution of the LM training text is very different than the one you are trying to measure accuracy. You can check the perplexity of your LM on the test data. |
Beta Was this translation helpful? Give feedback.
4 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
I am trying to add a language model to a Quartznet 15x5 which I have finetuned to a new language. I am using the same manifest JSON I used to train the acoustic model so am expecting only a small benefit. However the introduction of the language model causes a significant drop in performance.
When I switch from 'greedy' to 'beamsearch' I see a small WER improvement of 0.32, however using 'beamsearch_ngram' I always see a performance drop of 6+ WER. I have tried fine-tuning alpha and beta, however the performance always decreases. Interestingly this still occurs when alpha and beta are 0, where the language model should have no effect.
Does anything stand out from the training and inference runs below?
Training command
Training output (minus model load for brevity)
Beamsearch command
Beamsearch output
Beamsearch_ngram command
Beamsearch_ngram output
Beta Was this translation helpful? Give feedback.
All reactions