Skip to content
Discussion options

You must be logged in to vote

So, in these cases what is the appropriate way to handle these words in our training/test data set?

You would leave them as is most of the time. Acoustically, these are common speech patterns and the model should learn that the text should be represented as such. Of course, pronunciation of speech depends on the dataset, optimizing for just American English and expecting it to work with say British/Scottish English is unrealistic. The model will only be robust to accents it was trained on.

We train on some 12000 hours of speech, and a lot of it is not formal speech, we do normalization for numerics and remove punctuation and capitalization, but keep the rest and it seems to work quite w…

Replies: 1 comment 3 replies

Comment options

You must be logged in to vote
3 replies
@ahkarami
Comment options

@titu1994
Comment options

@ahkarami
Comment options

Answer selected by ahkarami
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
2 participants