mirror of https://github.com/coqui-ai/TTS.git
* Implement most similar ref training approach * Use non-enhanced hifigan for test samples * Add Perceiver * Update GPT Trainer for perceiver support * Update XTTS docs * Bug fix masking with XTTS perceiver * Bug fix on gpt forward * Bug Fix on XTTS v2.0 training * Add XTTS v2.0 unit tests * Add XTTS v2.0 inference unit tests * Bug Fix on diffusion inference * Add XTTS v2.0 training recipe * Placeholder model entry * Add cloning params to config * Make prompt embedding configurable * Make cloning configurable * Cheap fix for a cheaper fix * Prevent resampling * Update model entry * Update docs * Update requirements * Code linting * Add xtts v2 to sep tests * Bug fix on XTTS get_gpt_cond_latents * Bug fix on rebase * Make style * Bug fix in Japenese tokenizer * Add num2words to deps * Remove unused kwarg and added num_beams=1 as default --------- Co-authored-by: Eren G??lge <egolge@coqui.ai> |
||
---|---|---|
.. | ||
align_tts | ||
delightful_tts | ||
fast_pitch | ||
fast_speech | ||
fastspeech2 | ||
glow_tts | ||
hifigan | ||
multiband_melgan | ||
neuralhmm_tts | ||
overflow | ||
speedy_speech | ||
tacotron2-Capacitron | ||
tacotron2-DCA | ||
tacotron2-DDC | ||
univnet | ||
vits_tts | ||
wavegrad | ||
wavernn | ||
xtts_v1 | ||
xtts_v2 | ||
README.md | ||
download_ljspeech.sh |
README.md
🐸💬 TTS LJspeech Recipes
For running the recipes
-
Download the LJSpeech dataset here either manually from its official website or using
download_ljspeech.sh
. -
Go to your desired model folder and run the training.
Running Python files. (Choose the desired GPU ID for your run and set
CUDA_VISIBLE_DEVICES
)CUDA_VISIBLE_DEVICES="0" python train_modelX.py
Running bash scripts.
bash run.sh
💡 Note that these runs are just templates to help you start training your first model. They are not optimized for the best result. Double-check the configurations and feel free to share your experiments to find better parameters together 💪.