* Implement most similar ref training approach
* Use non-enhanced hifigan for test samples
* Add Perceiver
* Update GPT Trainer for perceiver support
* Update XTTS docs
* Bug fix masking with XTTS perceiver
* Bug fix on gpt forward
* Bug Fix on XTTS v2.0 training
* Add XTTS v2.0 unit tests
* Add XTTS v2.0 inference unit tests
* Bug Fix on diffusion inference
* Add XTTS v2.0 training recipe
* Placeholder model entry
* Add cloning params to config
* Make prompt embedding configurable
* Make cloning configurable
* Cheap fix for a cheaper fix
* Prevent resampling
* Update model entry
* Update docs
* Update requirements
* Code linting
* Add xtts v2 to sep tests
* Bug fix on XTTS get_gpt_cond_latents
* Bug fix on rebase
* Make style
* Bug fix in Japenese tokenizer
* Add num2words to deps
* Remove unused kwarg and added num_beams=1 as default
---------
Co-authored-by: Eren G??lge <egolge@coqui.ai>
* Adding neural HMM TTS
* Adding tests
* Adding neural hmm on readme
* renaming training recipe
* Removing overflow\s decoder parameters from the config
* Update the Trainer requirement version for a compatible one (#2276)
* Bump up to v0.10.2
* Adding neural HMM TTS
* Adding tests
* Adding neural hmm on readme
* renaming training recipe
* Removing overflow\s decoder parameters from the config
* fixing documentation
Co-authored-by: Edresson Casanova <edresson1@gmail.com>
Co-authored-by: Eren Gölge <erogol@hotmail.com>
* Fixed bug related to yourtts speaker embeddings issue
* Reverted code for base_tts
* Bug fix on VITS d_vector_file type
* Ignore the test speakers on YourTTS recipe
* Add speaker encoder model and config on YourTTS recipe to easily do zero-shot inference
* Update YourTTS config file
* Update ModelManager._update_path to deal with list attributes
* Fix lint checks
* Remove unused code
* Fix unit tests
* Reset name_to_id to get the right speaker ids on load_embeddings_from_list_of_files
* Set weighted_sampler_multipliers as an empty dict to prevent users' mistakes
Co-authored-by: Edresson Casanova <edresson1@gmail.com>
* Add YourTTS VCTK recipe
* Fix lint
* Add compute_embeddings and resample_files functions to be able to reuse it
* Add automatic download and speaker embedding computation for YourTTS VCTK recipe
* Add parameter for eval metadata file on compute embeddings function
* Adding encoder
* currently modifying hmm
* Adding hmm
* Adding overflow
* Adding overflow setting up flat start
* Removing runs
* adding normalization parameters
* Fixing models on same device
* Training overflow and plotting evaluations
* Adding inference
* At the end of epoch the test sentences are coming on cpu instead of gpu
* Adding figures from model during training to monitor
* reverting tacotron2 training recipe
* fixing inference on gpu for test sentences on config
* moving helpers and texts within overflows source code
* renaming to overflow
* moving loss to the model file
* Fixing the rename
* Model training but not plotting the test config sentences's audios
* Formatting logs
* Changing model name to camelcase
* Fixing test log
* Fixing plotting bug
* Adding some tests
* Adding more tests to overflow
* Adding all tests for overflow
* making changes to camel case in config
* Adding information about parameters and docstring
* removing compute_mel_statistics moved statistic computation to the model instead
* Added overflow in readme
* Adding more test cases, now it doesn't saves transition_p like tensor and can be dumped as json
* Update BaseDatasetConfig
- Add dataset_name
- Chane name to formatter_name
* Update compute_embedding
- Allow entering dataset by args
- Use released model by default
- Use the new key format
* Update loading
* Update recipes
* Update other dep code
* Update tests
* Fixup
* Load multiple embedding files
* Fix argument names in dep code
* Update docs
* Fix argument name
* Fix linter
- updated the recipes activating more losses for more stable training
- re-enabling guided attention loss
- fixed a bug about not the correct lr fetched for logging