mirror of
https://github.com/malarinv/tacotron2
synced 2026-03-09 02:02:33 +00:00
Compare commits
3 Commits
mask-utils
...
padding-pa
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
b20765a3dc | ||
|
|
2a394f4aaa | ||
|
|
2c545ac800 |
2
model.py
2
model.py
@@ -470,8 +470,8 @@ class Tacotron2(nn.Module):
|
|||||||
text_padded, input_lengths, mel_padded, gate_padded, \
|
text_padded, input_lengths, mel_padded, gate_padded, \
|
||||||
output_lengths = batch
|
output_lengths = batch
|
||||||
text_padded = to_gpu(text_padded).long()
|
text_padded = to_gpu(text_padded).long()
|
||||||
|
max_len = int(torch.max(input_lengths.data).numpy())
|
||||||
input_lengths = to_gpu(input_lengths).long()
|
input_lengths = to_gpu(input_lengths).long()
|
||||||
max_len = torch.max(input_lengths.data)
|
|
||||||
mel_padded = to_gpu(mel_padded).float()
|
mel_padded = to_gpu(mel_padded).float()
|
||||||
gate_padded = to_gpu(gate_padded).float()
|
gate_padded = to_gpu(gate_padded).float()
|
||||||
output_lengths = to_gpu(output_lengths).long()
|
output_lengths = to_gpu(output_lengths).long()
|
||||||
|
|||||||
Reference in New Issue
Block a user