diff --git a/layers/attention.py b/layers/attention.py
index d64f6894..ee18386e 100644
--- a/layers/attention.py
+++ b/layers/attention.py
@@ -118,8 +118,8 @@ class AttentionRNNCell(nn.Module):
         self.rnn_cell = nn.GRUCell(annot_dim + memory_dim, rnn_dim)
         self.windowing = windowing
         if self.windowing:
-            self.win_back = 1
-            self.win_front = 3
+            self.win_back = 3
+            self.win_front = 6
             self.win_idx = None
         # pick bahdanau or location sensitive attention
         if align_model == 'b':