Hi,
There is a problem about training a conformer+RNN-T model.
How about the cer and wer with one GPU?
I'm train the model on one RTX TITAN GPU, training the conformer(encoder layers 16, encoder dim 144, decoder layer 1, decoder dim 320) on Librispeech 960h., after 50 epoch training the CER is about 27 and don't reduce anymore.
could you tell me why?