From e4ce66192d4a8ddc7c581ed1ab0d7ee12dbdfa41 Mon Sep 17 00:00:00 2001 From: Eren Golge Date: Wed, 23 May 2018 06:20:04 -0700 Subject: [PATCH] Add a missing class variable to attention class --- layers/attention.py | 1 + 1 file changed, 1 insertion(+) diff --git a/layers/attention.py b/layers/attention.py index 6b9ee47b..f9a36e66 100644 --- a/layers/attention.py +++ b/layers/attention.py @@ -80,6 +80,7 @@ class AttentionRNN(nn.Module): align_model (str): 'b' for Bahdanau, 'ls' Location Sensitive alignment. """ super(AttentionRNN, self).__init__() + self.align_model = align_model self.rnn_cell = nn.GRUCell(out_dim + memory_dim, out_dim) # pick bahdanau or location sensitive attention if align_model == 'b':