From 013ec2f1689f63bbc5333cf60b98b04457506710 Mon Sep 17 00:00:00 2001 From: Eren Golge Date: Tue, 28 May 2019 14:29:17 +0200 Subject: [PATCH] bug fix for tacotron adapting it to new common layers --- layers/tacotron.py | 2 +- layers/tacotron2.py | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/layers/tacotron.py b/layers/tacotron.py index 125d56c7..8d762cc8 100644 --- a/layers/tacotron.py +++ b/layers/tacotron.py @@ -454,7 +454,7 @@ class Decoder(nn.Module): if t > 0: new_memory = outputs[-1] self._update_memory_queue(new_memory) - output, stop_token, attention = self.decode(inputs, t, None) + output, stop_token, attention = self.decode(inputs, None) stop_token = torch.sigmoid(stop_token.data) outputs += [output] attentions += [attention] diff --git a/layers/tacotron2.py b/layers/tacotron2.py index 8cdc9c16..0277734c 100644 --- a/layers/tacotron2.py +++ b/layers/tacotron2.py @@ -113,7 +113,7 @@ class Decoder(nn.Module): self.prenet = Prenet(self.mel_channels * r, prenet_type, prenet_dropout, - [self.prenet_dim, self.prenet_dim]) + [self.prenet_dim, self.prenet_dim], bias=False) self.attention_rnn = nn.LSTMCell(self.prenet_dim + in_features, self.attention_rnn_dim)