mirror of https://github.com/coqui-ai/TTS.git
more docstring
This commit is contained in:
parent
d95b1458e8
commit
cf44624df8
|
@ -198,6 +198,17 @@ class HifiganGenerator(torch.nn.Module):
|
||||||
self.conv_post = weight_norm(Conv1d(ch, out_channels, 7, 1, padding=3))
|
self.conv_post = weight_norm(Conv1d(ch, out_channels, 7, 1, padding=3))
|
||||||
|
|
||||||
def forward(self, x):
|
def forward(self, x):
|
||||||
|
"""
|
||||||
|
Args:
|
||||||
|
x (Tensor): conditioning input tensor.
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
Tensor: output waveform.
|
||||||
|
|
||||||
|
Shapes:
|
||||||
|
x: [B, C, T]
|
||||||
|
Tensor: [B, 1, T]
|
||||||
|
"""
|
||||||
o = self.conv_pre(x)
|
o = self.conv_pre(x)
|
||||||
for i in range(self.num_upsamples):
|
for i in range(self.num_upsamples):
|
||||||
o = F.leaky_relu(o, LRELU_SLOPE)
|
o = F.leaky_relu(o, LRELU_SLOPE)
|
||||||
|
@ -212,10 +223,22 @@ class HifiganGenerator(torch.nn.Module):
|
||||||
o = F.leaky_relu(o)
|
o = F.leaky_relu(o)
|
||||||
o = self.conv_post(o)
|
o = self.conv_post(o)
|
||||||
o = torch.tanh(o)
|
o = torch.tanh(o)
|
||||||
|
breakpoint()
|
||||||
return o
|
return o
|
||||||
|
|
||||||
@torch.no_grad()
|
@torch.no_grad()
|
||||||
def inference(self, c):
|
def inference(self, c):
|
||||||
|
"""
|
||||||
|
Args:
|
||||||
|
x (Tensor): conditioning input tensor.
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
Tensor: output waveform.
|
||||||
|
|
||||||
|
Shapes:
|
||||||
|
x: [B, C, T]
|
||||||
|
Tensor: [B, 1, T]
|
||||||
|
"""
|
||||||
c = c.to(self.conv_pre.weight.device)
|
c = c.to(self.conv_pre.weight.device)
|
||||||
c = torch.nn.functional.pad(
|
c = torch.nn.functional.pad(
|
||||||
c, (self.inference_padding, self.inference_padding), 'replicate')
|
c, (self.inference_padding, self.inference_padding), 'replicate')
|
||||||
|
|
Loading…
Reference in New Issue