Skip to content

Commit 0b89c10

Browse files
committed
2 parents 2967e8c + d2df554 commit 0b89c10

File tree

3 files changed

+15
-15
lines changed

3 files changed

+15
-15
lines changed

deepvoice3_pytorch/deepvoice3.py

Lines changed: 9 additions & 9 deletions
Original file line numberDiff line numberDiff line change
@@ -305,13 +305,13 @@ def forward(self, encoder_out, inputs=None,
305305
w = self.key_position_rate
306306
# TODO: may be useful to have projection per attention layer
307307
if self.speaker_proj1 is not None:
308-
w = w * F.sigmoid(self.speaker_proj1(speaker_embed)).view(-1)
308+
w = w * torch.sigmoid(self.speaker_proj1(speaker_embed)).view(-1)
309309
text_pos_embed = self.embed_keys_positions(text_positions, w)
310310
keys = keys + text_pos_embed
311311
if frame_positions is not None:
312312
w = self.query_position_rate
313313
if self.speaker_proj2 is not None:
314-
w = w * F.sigmoid(self.speaker_proj2(speaker_embed)).view(-1)
314+
w = w * torch.sigmoid(self.speaker_proj2(speaker_embed)).view(-1)
315315
frame_pos_embed = self.embed_query_positions(frame_positions, w)
316316

317317
# transpose only once to speed up attention layers
@@ -357,10 +357,10 @@ def forward(self, encoder_out, inputs=None,
357357
x = x.transpose(1, 2)
358358

359359
# project to mel-spectorgram
360-
outputs = F.sigmoid(x)
360+
outputs = torch.sigmoid(x)
361361

362362
# Done flag
363-
done = F.sigmoid(self.fc(x))
363+
done = torch.sigmoid(self.fc(x))
364364

365365
return outputs, torch.stack(alignments), done, decoder_states
366366

@@ -373,7 +373,7 @@ def incremental_forward(self, encoder_out, text_positions, speaker_embed=None,
373373
w = self.key_position_rate
374374
# TODO: may be useful to have projection per attention layer
375375
if self.speaker_proj1 is not None:
376-
w = w * F.sigmoid(self.speaker_proj1(speaker_embed)).view(-1)
376+
w = w * torch.sigmoid(self.speaker_proj1(speaker_embed)).view(-1)
377377
text_pos_embed = self.embed_keys_positions(text_positions, w)
378378
keys = keys + text_pos_embed
379379

@@ -399,7 +399,7 @@ def incremental_forward(self, encoder_out, text_positions, speaker_embed=None,
399399
frame_pos = keys.data.new(B, 1).fill_(t + 1).long()
400400
w = self.query_position_rate
401401
if self.speaker_proj2 is not None:
402-
w = w * F.sigmoid(self.speaker_proj2(speaker_embed)).view(-1)
402+
w = w * torch.sigmoid(self.speaker_proj2(speaker_embed)).view(-1)
403403
frame_pos_embed = self.embed_query_positions(frame_pos, w)
404404

405405
if test_inputs is not None:
@@ -457,8 +457,8 @@ def incremental_forward(self, encoder_out, text_positions, speaker_embed=None,
457457
ave_alignment = ave_alignment.div_(num_attention_layers)
458458

459459
# Ooutput & done flag predictions
460-
output = F.sigmoid(x)
461-
done = F.sigmoid(self.fc(x))
460+
output = torch.sigmoid(x)
461+
done = torch.sigmoid(self.fc(x))
462462

463463
decoder_states += [decoder_state]
464464
outputs += [output]
@@ -601,4 +601,4 @@ def forward(self, x, speaker_embed=None):
601601
# Back to B x T x C
602602
x = x.transpose(1, 2)
603603

604-
return F.sigmoid(x)
604+
return torch.sigmoid(x)

deepvoice3_pytorch/modules.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -160,7 +160,7 @@ def _forward(self, x, speaker_embed, is_incremental):
160160
# Since conv layer assumes BCT, we need to transpose
161161
softsign = softsign if is_incremental else softsign.transpose(1, 2)
162162
a = a + softsign
163-
x = a * F.sigmoid(b)
163+
x = a * torch.sigmoid(b)
164164
return (x + residual) * math.sqrt(0.5) if self.residual else x
165165

166166
def clear_buffer(self):
@@ -222,7 +222,7 @@ def _forward(self, x, is_incremental):
222222
return (x + residual) * math.sqrt(0.5)
223223
else:
224224
a, b = x.split(x.size(splitdim) // 2, dim=splitdim)
225-
T = F.sigmoid(b)
225+
T = torch.sigmoid(b)
226226
return (T * a + (1 - T) * residual)
227227

228228
def clear_buffer(self):

deepvoice3_pytorch/nyanko.py

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -237,10 +237,10 @@ def forward(self, encoder_out, inputs=None,
237237
x = x.transpose(1, 2)
238238

239239
# Mel
240-
outputs = F.sigmoid(x)
240+
outputs = torch.sigmoid(x)
241241

242242
# Done prediction
243-
done = F.sigmoid(self.fc(x))
243+
done = torch.sigmoid(self.fc(x))
244244

245245
# Adding extra dim for convenient
246246
alignments = alignments.unsqueeze(0)
@@ -310,8 +310,8 @@ def incremental_forward(self, encoder_out, text_positions,
310310
x = self.last_conv.incremental_forward(x)
311311

312312
# Ooutput & done flag predictions
313-
output = F.sigmoid(x)
314-
done = F.sigmoid(self.fc(x))
313+
output = torch.sigmoid(x)
314+
done = torch.sigmoid(self.fc(x))
315315

316316
decoder_states += [decoder_state]
317317
outputs += [output]

0 commit comments

Comments
 (0)