Commit Graph

2125 Commits

Author SHA1 Message Date
Enno Hermann 63625e79af refactor: import get_last_checkpoint from trainer.io 2024-11-29 13:59:43 +01:00
Enno Hermann 170d3dae92 refactor: remove duplicate to_camel 2024-11-24 19:57:14 +01:00
Enno Hermann 7330ad8854 refactor: move duplicate alignment functions into helpers 2024-11-24 19:57:14 +01:00
Enno Hermann 76df6421de refactor: move more audio processing into torch_transforms 2024-11-24 19:57:14 +01:00
Enno Hermann b1ac884e07 refactor: move shared function into dataset.py 2024-11-24 19:57:14 +01:00
Enno Hermann 54f4228a46 refactor(xtts): use existing cleaners 2024-11-24 19:57:14 +01:00
Enno Hermann b45a7a4220 refactor: move exists() and default() into generic_utils 2024-11-24 19:57:14 +01:00
Enno Hermann fa844e0fb7 refactor(tacotron): remove duplicate function 2024-11-24 19:57:14 +01:00
Enno Hermann 0f69d31f70 refactor(vocoder): remove duplicate function 2024-11-24 19:57:14 +01:00
Enno Hermann 6ecf47312c refactor(xtts): use tortoise conditioning encoder 2024-11-24 19:57:14 +01:00
Enno Hermann 69a599d403 refactor(freevc): remove duplicate code 2024-11-24 19:57:14 +01:00
Enno Hermann 2e5f68df6a refactor(wavernn): remove duplicate Stretch2d
I checked that the implementations are the same
2024-11-23 01:04:17 +01:00
Enno Hermann e63962c226 refactor(losses): move shared losses into losses.py 2024-11-23 01:04:17 +01:00
Enno Hermann 6f25c2b904 refactor(delightful_tts): remove unused classes 2024-11-23 01:04:17 +01:00
Enno Hermann 7cdfde226b refactor: move amp_to_db/db_to_amp into torch_transforms 2024-11-23 01:04:17 +01:00
Enno Hermann 33ac0d6ee1 refactor(xtts): use build_hf_gpt_transformer from tortoise 2024-11-21 15:44:42 +01:00
Enno Hermann 490c973371 refactor(xtts): use position embedding from tortoise 2024-11-21 15:44:42 +01:00
Enno Hermann 5ffc0543b7 refactor(bark): remove custom layer norm
Pytorch LayerNorm supports bias=False since version 2.1
2024-11-21 15:44:42 +01:00
Enno Hermann 705551c60c refactor(tortoise): remove unused do_checkpoint arguments
These are assigned but not used for anything.
2024-11-21 15:44:42 +01:00
Enno Hermann 4ba83f42ab chore(tortoise): remove unused AudioMiniEncoder
There's one in tortoise.classifier that's actually used
2024-11-21 15:44:42 +01:00
Enno Hermann 66701e1e51 refactor(xtts): reuse functions/classes from tortoise 2024-11-21 12:51:42 +01:00
Enno Hermann 1f27f994a1 refactor(utils): remove duplicate set_partial_state_dict 2024-11-21 11:53:35 +01:00
Enno Hermann 1b6d3ebd33 refactor(xtts): remove duplicate hifigan generator 2024-11-21 11:53:35 +01:00
Enno Hermann fbbae5ac6a refactor(audio): remove duplicate rms_volume_norm function 2024-11-18 10:01:52 +01:00
Enno Hermann 8ba3233ec6 refactor(audio): remove duplicate save_wav code 2024-11-18 10:01:50 +01:00
Enno Hermann 5784f6705a refactor(audio): improve type hints, address lint issues 2024-11-18 10:00:55 +01:00
Enno Hermann 48f5be2ccb feat(audio): automatically convert audio to mono 2024-11-18 01:11:43 +01:00
Enno Hermann 627bbe4150 fix(xtts): more helpful error message when vocab.json not found 2024-11-17 00:30:32 +01:00
Enno Hermann e81f8d079c fix: more helpful error message when formatter is not found 2024-11-17 00:29:49 +01:00
Enno Hermann 993da778b4 chore: use original instead of scarf urls
These allowed Coqui to get download stats, which we don't need anymore
2024-11-10 22:19:25 +01:00
Enno Hermann 2df9bfa78e
refactor: handle deprecation of torch.cuda.amp.autocast (#144)
torch.cuda.amp.autocast(args...) and torch.cpu.amp.autocast(args...) will be
deprecated. Please use torch.autocast("cuda", args...) or torch.autocast("cpu",
args...) instead.

https://pytorch.org/docs/stable/amp.html
2024-11-09 18:37:08 +01:00
Shavit 540e8d6cf2
fix(bin.synthesize): return speakers names only (#147) 2024-11-09 18:35:54 +01:00
Enno Hermann 0971bc236e refactor: use external package for monotonic alignment 2024-11-08 12:35:04 +01:00
Enno Hermann 6314032fd7
Merge pull request #113 from idiap/pytorch
fix: only enable load with weights_only in pytorch>=2.4
2024-11-04 22:14:42 +01:00
Enno Hermann 8e66be2c32 fix: only enable load with weights_only in pytorch>=2.4
Allows moving the minimum Pytorch version back to 2.1
2024-11-04 18:39:51 +01:00
Enno Hermann 47ad0bf190
fix(text.characters): add nasal diacritic (#127) 2024-11-04 17:40:23 +01:00
Enno Hermann 964b813235 fix(gpt): set attention mask and address other warnings 2024-10-25 18:49:04 +02:00
Johnny Street 018d4ba1db fix(xtts): support transformers>=4.43.0 in streaming inference 2024-10-20 15:20:26 +02:00
Shavit 36611a7192
feat: normalize unicode characters in text cleaners (#85)
* Add normalizer type C to text cleaners

* Linter recommendations

* Add unicode normalize to every cleaner

* Format test_text_cleaners.py
2024-10-02 17:01:19 +02:00
Enno Hermann e5dd06b3bb
Merge pull request #77 from shavit/71-torch-load
Load weights only in torch.load
2024-09-12 23:28:57 +01:00
Enno Hermann 659b4852ba chore(bark): remove manual download of hubert model
Bark was previously adapted to download Hubert from HuggingFace, so the manual
download is superfluous.
2024-09-12 23:37:19 +02:00
Enno Hermann 86b58fb6d9 fix: define torch safe globals for torch.load
Required for loading some models using torch.load(..., weights_only=True). This
is only available from Pytorch 2.4
2024-09-12 23:37:19 +02:00
shavit 17ca24c3d6 fix: load weights only in torch.load 2024-09-12 23:37:19 +02:00
Enno Hermann 1920328822
feat(xtts): support hindi in tokenizer (#64)
Added proper tokenizer support for Hindi Language which would prevent crash while fine tuning Hindi language.

Co-authored-by: Akshat Bhardwaj <157223825+akshatrocky@users.noreply.github.com>
2024-09-12 21:29:21 +02:00
Enno Hermann 9c604c1de0 chore(dataset): address lint issues 2024-07-31 15:47:27 +02:00
Enno Hermann 8c460d0cd0 fix(dataset): skip files where audio length can't be computed
Avoids hard failures when the audio can't be decoded.
2024-07-31 15:20:56 +02:00
Daniel Walmsley 20bbb411c2
fix(xtts): update streaming for transformers>=4.42.0 (#59)
* Fix Stream Generator on MacOS

* Make it work on mps

* Implement custom tensor.isin

* Fix for latest TF

* Comment out hack for now

* Remove unused code

* build: increase minimum transformers version

* style: fix

---------

Co-authored-by: Enno Hermann <Eginhard@users.noreply.github.com>
2024-07-25 16:24:10 +02:00
Enno Hermann 20583a496e
Merge pull request #57 from idiap/xtts-vocab
fix(xtts): load tokenizer file based on config as last resort
2024-07-25 13:26:28 +01:00
Enno Hermann de35920317
Merge pull request #50 from idiap/umap
build: move umap-learn into optional notebook dependencies
2024-07-25 13:26:09 +01:00
Enno Hermann 9192ef1aa6 fix(xtts): load tokenizer file based on config as last resort 2024-07-05 13:52:01 +02:00