diff --git a/train.py b/train.py index ff587d0e..7011b024 100644 --- a/train.py +++ b/train.py @@ -130,7 +130,7 @@ def train(model, criterion, data_loader, optimizer, epoch): mel_lengths_var) print(M.shape) print(alignments.shape) - attention_loss = criterion(M, alignments, mel_lengths_var) + attention_loss = criterion(alignments, M, mel_lengths_var) loss = mel_loss + linear_loss + 0.2 * attention_loss # backpass and check the grad norm