Explorar el Código

[Tacotron2/PyT] rename common->tacotron2_common

Grzegorz Karch hace 4 años
padre
commit
5a8521ee05

BIN
PyTorch/SpeechSynthesis/Tacotron2/data/mel.pt


+ 1 - 1
PyTorch/SpeechSynthesis/Tacotron2/exports/export_tacotron2_onnx.py

@@ -35,7 +35,7 @@ sys.path.append('./')
 
 import models
 from inference import checkpoint_from_distributed, unwrap_distributed, load_and_setup_model, prepare_input_sequence
-from common.utils import to_gpu, get_mask_from_lengths
+from tacotron2_common.utils import to_gpu, get_mask_from_lengths
 
 def parse_args(parser):
     """

+ 1 - 1
PyTorch/SpeechSynthesis/Tacotron2/preprocess_audio2mel.py

@@ -2,7 +2,7 @@ import argparse
 import torch
 
 from tacotron2.data_function import TextMelLoader
-from common.utils import load_filepaths_and_text
+from tacotron2_common.utils import load_filepaths_and_text
 
 def parse_args(parser):
     """

+ 2 - 2
PyTorch/SpeechSynthesis/Tacotron2/tacotron2/data_function.py

@@ -30,8 +30,8 @@ import numpy as np
 import torch
 import torch.utils.data
 
-import common.layers as layers
-from common.utils import load_wav_to_torch, load_filepaths_and_text, to_gpu
+import tacotron2_common.layers as layers
+from tacotron2_common.utils import load_wav_to_torch, load_filepaths_and_text, to_gpu
 from tacotron2.text import text_to_sequence
 
 class TextMelLoader(torch.utils.data.Dataset):

+ 2 - 2
PyTorch/SpeechSynthesis/Tacotron2/tacotron2/model.py

@@ -33,8 +33,8 @@ import sys
 from os.path import abspath, dirname
 # enabling modules discovery from global entrypoint
 sys.path.append(abspath(dirname(__file__)+'/../'))
-from common.layers import ConvNorm, LinearNorm
-from common.utils import to_gpu, get_mask_from_lengths
+from tacotron2_common.layers import ConvNorm, LinearNorm
+from tacotron2_common.utils import to_gpu, get_mask_from_lengths
 
 
 class LocationLayer(nn.Module):

+ 0 - 0
PyTorch/SpeechSynthesis/Tacotron2/common/audio_processing.py → PyTorch/SpeechSynthesis/Tacotron2/tacotron2_common/audio_processing.py


+ 2 - 2
PyTorch/SpeechSynthesis/Tacotron2/common/layers.py → PyTorch/SpeechSynthesis/Tacotron2/tacotron2_common/layers.py

@@ -27,8 +27,8 @@
 
 import torch
 from librosa.filters import mel as librosa_mel_fn
-from common.audio_processing import dynamic_range_compression, dynamic_range_decompression
-from common.stft import STFT
+from tacotron2_common.audio_processing import dynamic_range_compression, dynamic_range_decompression
+from tacotron2_common.stft import STFT
 
 
 class LinearNorm(torch.nn.Module):

+ 1 - 1
PyTorch/SpeechSynthesis/Tacotron2/common/stft.py → PyTorch/SpeechSynthesis/Tacotron2/tacotron2_common/stft.py

@@ -36,7 +36,7 @@ import torch.nn.functional as F
 from torch.autograd import Variable
 from scipy.signal import get_window
 from librosa.util import pad_center, tiny
-from common.audio_processing import window_sumsquare
+from tacotron2_common.audio_processing import window_sumsquare
 
 
 class STFT(torch.nn.Module):

+ 0 - 0
PyTorch/SpeechSynthesis/Tacotron2/common/utils.py → PyTorch/SpeechSynthesis/Tacotron2/tacotron2_common/utils.py


+ 1 - 1
PyTorch/SpeechSynthesis/Tacotron2/tensorrt/convert_tacotron22onnx.py

@@ -35,7 +35,7 @@ sys.path.append('./')
 
 import models
 from inference import checkpoint_from_distributed, unwrap_distributed, load_and_setup_model, prepare_input_sequence
-from common.utils import to_gpu, get_mask_from_lengths
+from tacotron2_common.utils import to_gpu, get_mask_from_lengths
 
 def parse_args(parser):
     """

+ 1 - 1
PyTorch/SpeechSynthesis/Tacotron2/tensorrt/convert_waveglow2onnx.py

@@ -31,7 +31,7 @@ import os
 import sys
 sys.path.append('./')
 
-from common.utils import ParseFromConfigFile
+from tacotron2_common.utils import ParseFromConfigFile
 from inference import load_and_setup_model
 
 def parse_args(parser):

+ 1 - 1
PyTorch/SpeechSynthesis/Tacotron2/tensorrt/inference_trt.py

@@ -35,7 +35,7 @@ import argparse
 import sys
 sys.path.append('./')
 
-from common.utils import to_gpu, get_mask_from_lengths
+from tacotron2_common.utils import to_gpu, get_mask_from_lengths
 from tacotron2.text import text_to_sequence
 from inference import MeasureTime, prepare_input_sequence, load_and_setup_model
 import dllogger as DLLogger

+ 1 - 1
PyTorch/SpeechSynthesis/Tacotron2/train.py

@@ -44,7 +44,7 @@ from apex.parallel import DistributedDataParallel as DDP
 import models
 import loss_functions
 import data_functions
-from common.utils import ParseFromConfigFile
+from tacotron2_common.utils import ParseFromConfigFile
 
 import dllogger as DLLogger
 from dllogger import StdOutBackend, JSONStreamBackend, Verbosity

+ 2 - 2
PyTorch/SpeechSynthesis/Tacotron2/waveglow/data_function.py

@@ -27,8 +27,8 @@
 
 import torch
 import random
-import common.layers as layers
-from common.utils import load_wav_to_torch, load_filepaths_and_text, to_gpu
+import tacotron2_common.layers as layers
+from tacotron2_common.utils import load_wav_to_torch, load_filepaths_and_text, to_gpu
 
 
 class MelAudioLoader(torch.utils.data.Dataset):

+ 1 - 1
PyTorch/SpeechSynthesis/Tacotron2/waveglow/denoiser.py

@@ -28,7 +28,7 @@
 import sys
 sys.path.append('tacotron2')
 import torch
-from common.layers import STFT
+from tacotron2_common.layers import STFT
 
 
 class Denoiser(torch.nn.Module):