11import torch
22from torch .utils .data import DataLoader
3- from synthesizer .hparams import hparams_debug_string
43from synthesizer .synthesizer_dataset import SynthesizerDataset , collate_synthesizer
54from synthesizer .models .tacotron import Tacotron
65from synthesizer .utils .text import text_to_sequence
76from synthesizer .utils .symbols import symbols
87import numpy as np
98from pathlib import Path
109from tqdm import tqdm
10+ import sys
1111
1212
1313def run_synthesis (in_dir , out_dir , model_dir , hparams ):
1414 # This generates ground truth-aligned mels for vocoder training
1515 synth_dir = Path (out_dir ).joinpath ("mels_gta" )
16- synth_dir .mkdir (exist_ok = True )
17- print (hparams_debug_string (hparams ))
16+ synth_dir .mkdir (parents = True , exist_ok = True )
17+ print (str (hparams ))
1818
1919 # Check for GPU
2020 if torch .cuda .is_available ():
@@ -59,12 +59,12 @@ def run_synthesis(in_dir, out_dir, model_dir, hparams):
5959 metadata_fpath = in_dir .joinpath ("train.txt" )
6060 mel_dir = in_dir .joinpath ("mels" )
6161 embed_dir = in_dir .joinpath ("embeds" )
62-
62+ num_workers = 0 if sys . platform . startswith ( "win" ) else 2 ;
6363 dataset = SynthesizerDataset (metadata_fpath , mel_dir , embed_dir , hparams )
6464 data_loader = DataLoader (dataset ,
65- collate_fn = lambda batch : collate_synthesizer (batch , r ),
65+ collate_fn = lambda batch : collate_synthesizer (batch ),
6666 batch_size = hparams .synthesis_batch_size ,
67- num_workers = 2 ,
67+ num_workers = num_workers ,
6868 shuffle = False ,
6969 pin_memory = True )
7070
@@ -78,9 +78,9 @@ def run_synthesis(in_dir, out_dir, model_dir, hparams):
7878
7979 # Parallelize model onto GPUS using workaround due to python bug
8080 if device .type == "cuda" and torch .cuda .device_count () > 1 :
81- _ , mels_out , _ = data_parallel_workaround (model , texts , mels , embeds )
81+ _ , mels_out , _ , _ = data_parallel_workaround (model , texts , mels , embeds )
8282 else :
83- _ , mels_out , _ = model (texts , mels , embeds )
83+ _ , mels_out , _ , _ = model (texts , mels , embeds )
8484
8585 for j , k in enumerate (idx ):
8686 # Note: outputs mel-spectrogram files and target ones have same names, just different folders
0 commit comments