1
0
mirror of https://github.com/malarinv/tacotron2 synced 2026-03-08 09:42:34 +00:00

8 Commits

Author SHA1 Message Date
Raul Puri
b20765a3dc 0.4 scalar tensor padding update 2018-05-04 12:12:08 -07:00
Raul Puri
2a394f4aaa integer maxlen for padding 2018-05-04 11:11:14 -07:00
Rafael Valle
2c545ac800 Merge pull request #4 from NVIDIA/mask-utils-0.4
mask utils update for 0.4 cuda
2018-05-04 11:02:30 -07:00
Raul Puri
6fbba8ef0f mask utils update for 0.4 cuda 2018-05-04 10:14:30 -07:00
Rafael Valle
c141726a96 requirements.txt: updating tensorflow requirements 2018-05-04 09:44:14 -07:00
Rafael Valle
535042a584 README.md: updating readme to include docker setup 2018-05-04 09:42:11 -07:00
Rafael Valle
a72160b8cb Dockerfile: adding dockerfile 2018-05-04 09:39:34 -07:00
Rafael Valle
d750fcf395 Merge pull request #2 from NVIDIA/single-gpu-and-0.4
train.py single gpu and 0.4 update
2018-05-04 09:12:13 -07:00
5 changed files with 7 additions and 6 deletions

View File

@@ -1,3 +1,4 @@
FROM pytorch/pytorch:0.4_cuda9_cudnn7
RUN pip install numpy scipy matplotlib librosa==0.6.0 tensorflow tensorboardX inflect==0.2.5 Unidecode==1.0.22
RUN pip install numpy scipy matplotlib librosa==0.6.0 tensorflow tensorboardX
inflect==0.2.5 Unidecode==1.0.22

View File

@@ -21,10 +21,10 @@ Distributed and FP16 support relies on work by Christian Sarofeen and NVIDIA's
3. CD into this repo: `cd tacotron2`
4. Update .wav paths: `sed -i -- 's,DUMMY,ljs_dataset_folder/wavs,g' filelists/*.txt`
5. Install [pytorch 0.4](https://github.com/pytorch/pytorch)
6. Install python requirements or use docker container (tbd)
6. Install python requirements or build docker image
- Install python requirements: `pip install requirements.txt`
- **OR**
- Docker container `(tbd)`
- Build docker image: `docker build --tag tacotron2 .`
## Training
1. `python train.py --output_directory=outdir --log_directory=logdir`

View File

@@ -470,8 +470,8 @@ class Tacotron2(nn.Module):
text_padded, input_lengths, mel_padded, gate_padded, \
output_lengths = batch
text_padded = to_gpu(text_padded).long()
max_len = int(torch.max(input_lengths.data).numpy())
input_lengths = to_gpu(input_lengths).long()
max_len = torch.max(input_lengths.data)
mel_padded = to_gpu(mel_padded).float()
gate_padded = to_gpu(gate_padded).float()
output_lengths = to_gpu(output_lengths).long()

View File

@@ -1,6 +1,6 @@
torch==0.2.0.post3
matplotlib==2.1.0
tensorflow==1.5.0
tensorflow
numpy==1.13.3
inflect==0.2.5
librosa==0.6.0

View File

@@ -5,7 +5,7 @@ import torch
def get_mask_from_lengths(lengths):
max_len = torch.max(lengths)
ids = torch.arange(0, max_len, out=torch.LongTensor(max_len)).cuda()
ids = torch.arange(0, max_len).long().cuda()
mask = (ids < lengths.unsqueeze(1)).byte()
return mask