Enno Hermann
63625e79af
refactor: import get_last_checkpoint from trainer.io
2024-11-29 13:59:43 +01:00
Enno Hermann
170d3dae92
refactor: remove duplicate to_camel
2024-11-24 19:57:14 +01:00
Enno Hermann
7330ad8854
refactor: move duplicate alignment functions into helpers
2024-11-24 19:57:14 +01:00
Enno Hermann
76df6421de
refactor: move more audio processing into torch_transforms
2024-11-24 19:57:14 +01:00
Enno Hermann
b1ac884e07
refactor: move shared function into dataset.py
2024-11-24 19:57:14 +01:00
Enno Hermann
54f4228a46
refactor(xtts): use existing cleaners
2024-11-24 19:57:14 +01:00
Enno Hermann
b45a7a4220
refactor: move exists() and default() into generic_utils
2024-11-24 19:57:14 +01:00
Enno Hermann
fa844e0fb7
refactor(tacotron): remove duplicate function
2024-11-24 19:57:14 +01:00
Enno Hermann
0f69d31f70
refactor(vocoder): remove duplicate function
2024-11-24 19:57:14 +01:00
Enno Hermann
6ecf47312c
refactor(xtts): use tortoise conditioning encoder
2024-11-24 19:57:14 +01:00
Enno Hermann
69a599d403
refactor(freevc): remove duplicate code
2024-11-24 19:57:14 +01:00
Enno Hermann
2e5f68df6a
refactor(wavernn): remove duplicate Stretch2d
...
I checked that the implementations are the same
2024-11-23 01:04:17 +01:00
Enno Hermann
e63962c226
refactor(losses): move shared losses into losses.py
2024-11-23 01:04:17 +01:00
Enno Hermann
6f25c2b904
refactor(delightful_tts): remove unused classes
2024-11-23 01:04:17 +01:00
Enno Hermann
7cdfde226b
refactor: move amp_to_db/db_to_amp into torch_transforms
2024-11-23 01:04:17 +01:00
Enno Hermann
33ac0d6ee1
refactor(xtts): use build_hf_gpt_transformer from tortoise
2024-11-21 15:44:42 +01:00
Enno Hermann
490c973371
refactor(xtts): use position embedding from tortoise
2024-11-21 15:44:42 +01:00
Enno Hermann
5ffc0543b7
refactor(bark): remove custom layer norm
...
Pytorch LayerNorm supports bias=False since version 2.1
2024-11-21 15:44:42 +01:00
Enno Hermann
705551c60c
refactor(tortoise): remove unused do_checkpoint arguments
...
These are assigned but not used for anything.
2024-11-21 15:44:42 +01:00
Enno Hermann
4ba83f42ab
chore(tortoise): remove unused AudioMiniEncoder
...
There's one in tortoise.classifier that's actually used
2024-11-21 15:44:42 +01:00
Enno Hermann
66701e1e51
refactor(xtts): reuse functions/classes from tortoise
2024-11-21 12:51:42 +01:00
Enno Hermann
1f27f994a1
refactor(utils): remove duplicate set_partial_state_dict
2024-11-21 11:53:35 +01:00
Enno Hermann
1b6d3ebd33
refactor(xtts): remove duplicate hifigan generator
2024-11-21 11:53:35 +01:00
Enno Hermann
fbbae5ac6a
refactor(audio): remove duplicate rms_volume_norm function
2024-11-18 10:01:52 +01:00
Enno Hermann
8ba3233ec6
refactor(audio): remove duplicate save_wav code
2024-11-18 10:01:50 +01:00
Enno Hermann
5784f6705a
refactor(audio): improve type hints, address lint issues
2024-11-18 10:00:55 +01:00
Enno Hermann
48f5be2ccb
feat(audio): automatically convert audio to mono
2024-11-18 01:11:43 +01:00
Enno Hermann
627bbe4150
fix(xtts): more helpful error message when vocab.json not found
2024-11-17 00:30:32 +01:00
Enno Hermann
e81f8d079c
fix: more helpful error message when formatter is not found
2024-11-17 00:29:49 +01:00
Enno Hermann
993da778b4
chore: use original instead of scarf urls
...
These allowed Coqui to get download stats, which we don't need anymore
2024-11-10 22:19:25 +01:00
Enno Hermann
2df9bfa78e
refactor: handle deprecation of torch.cuda.amp.autocast ( #144 )
...
torch.cuda.amp.autocast(args...) and torch.cpu.amp.autocast(args...) will be
deprecated. Please use torch.autocast("cuda", args...) or torch.autocast("cpu",
args...) instead.
https://pytorch.org/docs/stable/amp.html
2024-11-09 18:37:08 +01:00
Shavit
540e8d6cf2
fix(bin.synthesize): return speakers names only ( #147 )
2024-11-09 18:35:54 +01:00
Enno Hermann
0971bc236e
refactor: use external package for monotonic alignment
2024-11-08 12:35:04 +01:00
Enno Hermann
6314032fd7
Merge pull request #113 from idiap/pytorch
...
fix: only enable load with weights_only in pytorch>=2.4
2024-11-04 22:14:42 +01:00
Enno Hermann
8e66be2c32
fix: only enable load with weights_only in pytorch>=2.4
...
Allows moving the minimum Pytorch version back to 2.1
2024-11-04 18:39:51 +01:00
Enno Hermann
47ad0bf190
fix(text.characters): add nasal diacritic ( #127 )
2024-11-04 17:40:23 +01:00
Enno Hermann
964b813235
fix(gpt): set attention mask and address other warnings
2024-10-25 18:49:04 +02:00
Johnny Street
018d4ba1db
fix(xtts): support transformers>=4.43.0 in streaming inference
2024-10-20 15:20:26 +02:00
Shavit
36611a7192
feat: normalize unicode characters in text cleaners ( #85 )
...
* Add normalizer type C to text cleaners
* Linter recommendations
* Add unicode normalize to every cleaner
* Format test_text_cleaners.py
2024-10-02 17:01:19 +02:00
Enno Hermann
e5dd06b3bb
Merge pull request #77 from shavit/71-torch-load
...
Load weights only in torch.load
2024-09-12 23:28:57 +01:00
Enno Hermann
659b4852ba
chore(bark): remove manual download of hubert model
...
Bark was previously adapted to download Hubert from HuggingFace, so the manual
download is superfluous.
2024-09-12 23:37:19 +02:00
Enno Hermann
86b58fb6d9
fix: define torch safe globals for torch.load
...
Required for loading some models using torch.load(..., weights_only=True). This
is only available from Pytorch 2.4
2024-09-12 23:37:19 +02:00
shavit
17ca24c3d6
fix: load weights only in torch.load
2024-09-12 23:37:19 +02:00
Enno Hermann
1920328822
feat(xtts): support hindi in tokenizer ( #64 )
...
Added proper tokenizer support for Hindi Language which would prevent crash while fine tuning Hindi language.
Co-authored-by: Akshat Bhardwaj <157223825+akshatrocky@users.noreply.github.com>
2024-09-12 21:29:21 +02:00
Enno Hermann
9c604c1de0
chore(dataset): address lint issues
2024-07-31 15:47:27 +02:00
Enno Hermann
8c460d0cd0
fix(dataset): skip files where audio length can't be computed
...
Avoids hard failures when the audio can't be decoded.
2024-07-31 15:20:56 +02:00
Daniel Walmsley
20bbb411c2
fix(xtts): update streaming for transformers>=4.42.0 ( #59 )
...
* Fix Stream Generator on MacOS
* Make it work on mps
* Implement custom tensor.isin
* Fix for latest TF
* Comment out hack for now
* Remove unused code
* build: increase minimum transformers version
* style: fix
---------
Co-authored-by: Enno Hermann <Eginhard@users.noreply.github.com>
2024-07-25 16:24:10 +02:00
Enno Hermann
20583a496e
Merge pull request #57 from idiap/xtts-vocab
...
fix(xtts): load tokenizer file based on config as last resort
2024-07-25 13:26:28 +01:00
Enno Hermann
de35920317
Merge pull request #50 from idiap/umap
...
build: move umap-learn into optional notebook dependencies
2024-07-25 13:26:09 +01:00
Enno Hermann
9192ef1aa6
fix(xtts): load tokenizer file based on config as last resort
2024-07-05 13:52:01 +02:00