mirror of https://github.com/coqui-ai/TTS.git
tf lstm does not match torch lstm wrt bias vectors. So I avoid bias in LSTM as an easy solution.
This commit is contained in:
parent
d282222553
commit
736f169cc9
|
@ -61,6 +61,7 @@ class Encoder(nn.Module):
|
||||||
int(output_input_dim / 2),
|
int(output_input_dim / 2),
|
||||||
num_layers=1,
|
num_layers=1,
|
||||||
batch_first=True,
|
batch_first=True,
|
||||||
|
bias=False,
|
||||||
bidirectional=True)
|
bidirectional=True)
|
||||||
self.rnn_state = None
|
self.rnn_state = None
|
||||||
|
|
||||||
|
@ -121,7 +122,8 @@ class Decoder(nn.Module):
|
||||||
bias=False)
|
bias=False)
|
||||||
|
|
||||||
self.attention_rnn = nn.LSTMCell(self.prenet_dim + input_dim,
|
self.attention_rnn = nn.LSTMCell(self.prenet_dim + input_dim,
|
||||||
self.query_dim)
|
self.query_dim,
|
||||||
|
bias=False)
|
||||||
|
|
||||||
self.attention = init_attn(attn_type=attn_type,
|
self.attention = init_attn(attn_type=attn_type,
|
||||||
query_dim=self.query_dim,
|
query_dim=self.query_dim,
|
||||||
|
|
Loading…
Reference in New Issue