From 84814db73fadd46a296281f21153abd799d7957c Mon Sep 17 00:00:00 2001 From: Eren Golge Date: Wed, 2 Jan 2019 12:52:17 +0100 Subject: [PATCH] reduce dropout ratio --- layers/tacotron.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/layers/tacotron.py b/layers/tacotron.py index 6b6cadb0..d538b274 100644 --- a/layers/tacotron.py +++ b/layers/tacotron.py @@ -22,7 +22,7 @@ class Prenet(nn.Module): for (in_size, out_size) in zip(in_features, out_features) ]) self.relu = nn.ReLU() - self.dropout = nn.Dropout(0.5) + self.dropout = nn.Dropout(0.1) # self.init_layers() def init_layers(self): @@ -455,7 +455,7 @@ class StopNet(nn.Module): def __init__(self, in_features): super(StopNet, self).__init__() - self.dropout = nn.Dropout(0.5) + self.dropout = nn.Dropout(0.1) self.linear = nn.Linear(in_features, 1) self.sigmoid = nn.Sigmoid() torch.nn.init.xavier_uniform_(