diff --git a/layers/tacotron.py b/layers/tacotron.py index 5256d215..b8210b78 100644 --- a/layers/tacotron.py +++ b/layers/tacotron.py @@ -286,7 +286,7 @@ class Decoder(nn.Module): processed_memory = self.prenet(memory_input) # Attention RNN attention_cat = torch.cat((attention.unsqueeze(1), - attention_cum.unsqueeze(1) / (t + 1)), + attention_cum.unsqueeze(1)), dim=1) attention_rnn_hidden, current_context_vec, attention = self.attention_rnn( processed_memory, current_context_vec, attention_rnn_hidden, inputs, attention_cat)