Better config for training

This commit is contained in:
Eren Golge 2018-01-26 02:53:01 -08:00
parent 2eca96bfbb
commit 9af37e43a4
2 changed files with 4 additions and 2 deletions

View File

@ -12,7 +12,7 @@
"text_cleaner": "english_cleaners",
"epochs": 2000,
"lr": 0.01,
"lr": 0.001,
"lr_patience": 2,
"lr_decay": 0.5,
"batch_size": 256,
@ -20,6 +20,8 @@
"power": 1.5,
"r": 5,
"num_loader_workers": 32,
"save_step": 200,
"data_path": "/data/shared/KeithIto/LJSpeech-1.0",
"output_path": "result",

View File

@ -106,7 +106,7 @@ def main(args):
dataloader = DataLoader(dataset, batch_size=c.batch_size,
shuffle=True, collate_fn=dataset.collate_fn,
drop_last=True, num_workers=32)
drop_last=True, num_workers=c.num_loader_workers)
print("\n | > Epoch {}/{}".format(epoch, c.epochs))
progbar = Progbar(len(dataset) / c.batch_size)