mirror of https://github.com/coqui-ai/TTS.git
Better config for training
This commit is contained in:
parent
2eca96bfbb
commit
9af37e43a4
|
@ -12,7 +12,7 @@
|
|||
"text_cleaner": "english_cleaners",
|
||||
|
||||
"epochs": 2000,
|
||||
"lr": 0.01,
|
||||
"lr": 0.001,
|
||||
"lr_patience": 2,
|
||||
"lr_decay": 0.5,
|
||||
"batch_size": 256,
|
||||
|
@ -20,6 +20,8 @@
|
|||
"power": 1.5,
|
||||
"r": 5,
|
||||
|
||||
"num_loader_workers": 32,
|
||||
|
||||
"save_step": 200,
|
||||
"data_path": "/data/shared/KeithIto/LJSpeech-1.0",
|
||||
"output_path": "result",
|
||||
|
|
2
train.py
2
train.py
|
@ -106,7 +106,7 @@ def main(args):
|
|||
|
||||
dataloader = DataLoader(dataset, batch_size=c.batch_size,
|
||||
shuffle=True, collate_fn=dataset.collate_fn,
|
||||
drop_last=True, num_workers=32)
|
||||
drop_last=True, num_workers=c.num_loader_workers)
|
||||
print("\n | > Epoch {}/{}".format(epoch, c.epochs))
|
||||
progbar = Progbar(len(dataset) / c.batch_size)
|
||||
|
||||
|
|
Loading…
Reference in New Issue