Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Making LJSpeech Tacotron2 model model say 'video'. #883

Open
bduvenhage opened this issue Mar 23, 2021 · 0 comments
Open

Making LJSpeech Tacotron2 model model say 'video'. #883

bduvenhage opened this issue Mar 23, 2021 · 0 comments

Comments

@bduvenhage
Copy link

bduvenhage commented Mar 23, 2021

Hi

This is possibly a question on TTS in general, but I'm fine tuning PyTorch Tacotron2 and I found that even the pre-trained LJSpeech model has trouble saying 'video'. I see the training and inference code supports arpabet, but I assume that the pre-trained models were trained on the plain text of LJSpeech and not phonetic sequences? I tried doing inference using arpabet sequences in curly brackets, but it doesn't work.

Do you have any recommendations for making the model say modern words that might not have been part of the LJSpeech corpus?

Thanks,

@nv-kkudrynski nv-kkudrynski assigned ghost Jun 15, 2021
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant