diff --git a/layers/tacotron.py b/layers/tacotron.py index 81328585..aa8531ca 100644 --- a/layers/tacotron.py +++ b/layers/tacotron.py @@ -263,15 +263,15 @@ class Decoder(nn.Module): T_decoder = memory.size(1) # go frame - 0 frames tarting the sequence initial_memory = \ - inputs.data.new(B, self.memory_dim * self.r).zero_().requires_grad_() + inputs.data.new(B, self.memory_dim * self.r).zero_() # Init decoder states attention_rnn_hidden = \ - inputs.data.new(B, 256).zero_().requires_grad_() + inputs.data.new(B, 256).zero_() decoder_rnn_hiddens = [ - inputs.data.new(B, 256).zero_().requires_grad_() + inputs.data.new(B, 256).zero_() for _ in range(len(self.decoder_rnns))] current_context_vec = \ - inputs.data.new(B, 256).zero_().requires_grad_() + inputs.data.new(B, 256).zero_() # Time first (T_decoder, B, memory_dim) if memory is not None: memory = memory.transpose(0, 1) diff --git a/train.py b/train.py index fb0c47fc..75a69081 100644 --- a/train.py +++ b/train.py @@ -93,12 +93,6 @@ def train(model, criterion, data_loader, optimizer, epoch): optimizer.zero_grad() - # convert inputs to variables - text_input.requires_grad_() - mel_spec.requires_grad_() - # mel_lengths.requires_grad_() - # linear_spec.requires_grad_() - # dispatch data to GPU if use_cuda: text_input = text_input.cuda()