Replies: 13 comments
-
>>> erogol |
Beta Was this translation helpful? Give feedback.
-
>>> carlfm01 |
Beta Was this translation helpful? Give feedback.
-
>>> carlfm01 |
Beta Was this translation helpful? Give feedback.
-
>>> josh_meyer |
Beta Was this translation helpful? Give feedback.
-
>>> erogol |
Beta Was this translation helpful? Give feedback.
-
>>> erogol |
Beta Was this translation helpful? Give feedback.
-
>>> carlfm01 |
Beta Was this translation helpful? Give feedback.
-
>>> josh_meyer |
Beta Was this translation helpful? Give feedback.
-
>>> josh_meyer |
Beta Was this translation helpful? Give feedback.
-
>>> carlfm01 |
Beta Was this translation helpful? Give feedback.
-
>>> erogol |
Beta Was this translation helpful? Give feedback.
-
>>> josh_meyer |
Beta Was this translation helpful? Give feedback.
-
>>> erogol |
Beta Was this translation helpful? Give feedback.
-
>>> josh_meyer
[August 2, 2019, 6:26pm]
Hey All!
Has anyone tried fine-tuning a pre-trained Tacotron2 model to new,
smaller dataset? What results are you getting? What do's and don't's
have you discovered?
This approach is also called adapting the model, speaker adaptation, or
transfer learning.
I'm currently fine-tuning a pre-trained Tacotron2 to a 15-hour in-house
dataset. After 10K iterations, I can hear the new voice (not the
original LJSpeech), but the model can only produce the first few words
of a sentence. Nevertheless, the amount of progress in just 10K
iterations is promising.
I'm thinking that learning rate will be very important here... and
possibly 'freezing' certain model layers. Does anyone have experience on
this?
Best, slash
Josh
[This is an archived TTS discussion thread from discourse.mozilla.org/t/fine-tuning-trained-model-to-new-dataset]
Beta Was this translation helpful? Give feedback.
All reactions