diff --git a/layers/losses.py b/layers/losses.py index 3fbd3141..2135ee89 100644 --- a/layers/losses.py +++ b/layers/losses.py @@ -1,6 +1,5 @@ import torch from torch.nn import functional -from torch.autograd import Variable from torch import nn @@ -11,7 +10,6 @@ def _sequence_mask(sequence_length, max_len=None): batch_size = sequence_length.size(0) seq_range = torch.arange(0, max_len).long() seq_range_expand = seq_range.unsqueeze(0).expand(batch_size, max_len) - seq_range_expand = Variable(seq_range_expand) if sequence_length.is_cuda: seq_range_expand = seq_range_expand.cuda() seq_length_expand = (sequence_length.unsqueeze(1)