rakhlin commited on
Commit
d3cbff2
·
1 Parent(s): 12e2807

Upload 25 files

Browse files
Files changed (25) hide show
  1. tts/tts_models--en--ljspeech--fast_pitch/config.json +163 -0
  2. tts/tts_models--en--ljspeech--fast_pitch/model_file.pth +3 -0
  3. tts/tts_models--en--ljspeech--glow-tts/config.json +150 -0
  4. tts/tts_models--en--ljspeech--glow-tts/model_file.pth +3 -0
  5. tts/tts_models--en--ljspeech--neural_hmm/._config.json +0 -0
  6. tts/tts_models--en--ljspeech--neural_hmm/._model_file.pth +3 -0
  7. tts/tts_models--en--ljspeech--neural_hmm/._tts_models--en--ljspeech--neural_hmm +0 -0
  8. tts/tts_models--en--ljspeech--neural_hmm/__MACOSX/._tts_models--en--ljspeech--neural_hmm +0 -0
  9. tts/tts_models--en--ljspeech--neural_hmm/__MACOSX/tts_models--en--ljspeech--neural_hmm/._config.json +0 -0
  10. tts/tts_models--en--ljspeech--neural_hmm/__MACOSX/tts_models--en--ljspeech--neural_hmm/._model_file.pth +3 -0
  11. tts/tts_models--en--ljspeech--neural_hmm/config.json +172 -0
  12. tts/tts_models--en--ljspeech--neural_hmm/model_file.pth +3 -0
  13. tts/tts_models--en--ljspeech--overflow/._config.json +0 -0
  14. tts/tts_models--en--ljspeech--overflow/._model_file.pth +3 -0
  15. tts/tts_models--en--ljspeech--overflow/._tts_models--en--ljspeech--overflow +0 -0
  16. tts/tts_models--en--ljspeech--overflow/__MACOSX/._tts_models--en--ljspeech--overflow +0 -0
  17. tts/tts_models--en--ljspeech--overflow/__MACOSX/tts_models--en--ljspeech--overflow/._config.json +0 -0
  18. tts/tts_models--en--ljspeech--overflow/__MACOSX/tts_models--en--ljspeech--overflow/._model_file.pth +3 -0
  19. tts/tts_models--en--ljspeech--overflow/config.json +185 -0
  20. tts/tts_models--en--ljspeech--overflow/model_file.pth +3 -0
  21. tts/tts_models--en--ljspeech--tacotron2-DCA/config.json +204 -0
  22. tts/tts_models--en--ljspeech--tacotron2-DCA/model_file.pth +3 -0
  23. tts/tts_models--en--ljspeech--tacotron2-DCA/scale_stats.npy +3 -0
  24. tts/tts_models--en--ljspeech--vits/config.json +203 -0
  25. tts/tts_models--en--ljspeech--vits/model_file.pth +3 -0
tts/tts_models--en--ljspeech--fast_pitch/config.json ADDED
@@ -0,0 +1,163 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model": "fast_pitch",
3
+ "run_name": "fast_pitch_ljspeech",
4
+ "run_description": "",
5
+ "epochs": 1000,
6
+ "batch_size": 32,
7
+ "eval_batch_size": 16,
8
+ "mixed_precision": false,
9
+ "scheduler_after_epoch": false,
10
+ "run_eval": true,
11
+ "test_delay_epochs": -1,
12
+ "print_eval": false,
13
+ "dashboard_logger": "tensorboard",
14
+ "print_step": 50,
15
+ "plot_step": 100,
16
+ "model_param_stats": false,
17
+ "project_name": null,
18
+ "log_model_step": null,
19
+ "wandb_entity": null,
20
+ "save_step": 10000,
21
+ "checkpoint": true,
22
+ "keep_all_best": false,
23
+ "keep_after": 10000,
24
+ "num_loader_workers": 0,
25
+ "num_eval_loader_workers": 4,
26
+ "use_noise_augment": false,
27
+ "output_path": "/home/ubuntu/TTS/recipes/ljspeech/fast_pitch",
28
+ "distributed_backend": "nccl",
29
+ "distributed_url": "tcp://localhost:54321",
30
+ "audio": {
31
+ "fft_size": 1024,
32
+ "win_length": 1024,
33
+ "hop_length": 256,
34
+ "frame_shift_ms": null,
35
+ "frame_length_ms": null,
36
+ "stft_pad_mode": "reflect",
37
+ "sample_rate": 22050,
38
+ "resample": false,
39
+ "preemphasis": 0.0,
40
+ "ref_level_db": 20,
41
+ "do_sound_norm": false,
42
+ "log_func": "np.log",
43
+ "do_trim_silence": true,
44
+ "trim_db": 60.0,
45
+ "power": 1.5,
46
+ "griffin_lim_iters": 60,
47
+ "num_mels": 80,
48
+ "mel_fmin": 0.0,
49
+ "mel_fmax": 8000,
50
+ "spec_gain": 1.0,
51
+ "do_amp_to_db_linear": true,
52
+ "do_amp_to_db_mel": true,
53
+ "signal_norm": false,
54
+ "min_level_db": -100,
55
+ "symmetric_norm": true,
56
+ "max_norm": 4.0,
57
+ "clip_norm": true,
58
+ "stats_path": null
59
+ },
60
+ "use_phonemes": true,
61
+ "phonemizer": "gruut",
62
+ "phoneme_language": "en-us",
63
+ "compute_input_seq_cache": true,
64
+ "text_cleaner": "english_cleaners",
65
+ "enable_eos_bos_chars": false,
66
+ "test_sentences_file": "",
67
+ "phoneme_cache_path": "/home/ubuntu/TTS/recipes/ljspeech/fast_pitch/phoneme_cache",
68
+ "characters": {
69
+ "pad": "_",
70
+ "eos": "~",
71
+ "bos": "^",
72
+ "characters": "ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz!'(),-.:;? ",
73
+ "punctuations": "!'(),-.:;? ",
74
+ "phonemes": "iy\u0268\u0289\u026fu\u026a\u028f\u028ae\u00f8\u0258\u0259\u0275\u0264o\u025b\u0153\u025c\u025e\u028c\u0254\u00e6\u0250a\u0276\u0251\u0252\u1d7b\u0298\u0253\u01c0\u0257\u01c3\u0284\u01c2\u0260\u01c1\u029bpbtd\u0288\u0256c\u025fk\u0261q\u0262\u0294\u0274\u014b\u0272\u0273n\u0271m\u0299r\u0280\u2c71\u027e\u027d\u0278\u03b2fv\u03b8\u00f0sz\u0283\u0292\u0282\u0290\u00e7\u029dx\u0263\u03c7\u0281\u0127\u0295h\u0266\u026c\u026e\u028b\u0279\u027bj\u0270l\u026d\u028e\u029f\u02c8\u02cc\u02d0\u02d1\u028dw\u0265\u029c\u02a2\u02a1\u0255\u0291\u027a\u0267\u02b2\u025a\u02de\u026b"
75
+ },
76
+ "batch_group_size": 0,
77
+ "loss_masking": null,
78
+ "sort_by_audio_len": true,
79
+ "min_seq_len": 13,
80
+ "max_seq_len": 500000,
81
+ "compute_f0": true,
82
+ "compute_linear_spec": false,
83
+ "add_blank": false,
84
+ "datasets": [
85
+ {
86
+ "name": "ljspeech",
87
+ "path": "/home/ubuntu/TTS/recipes/ljspeech/fast_pitch/../LJSpeech-1.1/",
88
+ "meta_file_train": "metadata.csv",
89
+ "ununsed_speakers": null,
90
+ "meta_file_val": "",
91
+ "meta_file_attn_mask": ""
92
+ }
93
+ ],
94
+ "optimizer": "Adam",
95
+ "optimizer_params": {
96
+ "betas": [
97
+ 0.9,
98
+ 0.998
99
+ ],
100
+ "weight_decay": 1e-06
101
+ },
102
+ "lr_scheduler": "NoamLR",
103
+ "lr_scheduler_params": {
104
+ "warmup_steps": 4000
105
+ },
106
+ "test_sentences": [
107
+ "It took me quite a long time to develop a voice, and now that I have it I'm not going to be silent.",
108
+ "Be a voice, not an echo.",
109
+ "I'm sorry Dave. I'm afraid I can't do that.",
110
+ "This cake is great. It's so delicious and moist.",
111
+ "Prior to November 22, 1963."
112
+ ],
113
+ "use_speaker_embedding": false,
114
+ "use_d_vector_file": false,
115
+ "d_vector_dim": 0,
116
+ "model_args": {
117
+ "num_chars": null,
118
+ "out_channels": 80,
119
+ "hidden_channels": 384,
120
+ "num_speakers": 0,
121
+ "duration_predictor_hidden_channels": 256,
122
+ "duration_predictor_kernel_size": 3,
123
+ "duration_predictor_dropout_p": 0.1,
124
+ "pitch_predictor_hidden_channels": 256,
125
+ "pitch_predictor_kernel_size": 3,
126
+ "pitch_predictor_dropout_p": 0.1,
127
+ "pitch_embedding_kernel_size": 3,
128
+ "positional_encoding": true,
129
+ "poisitonal_encoding_use_scale": true,
130
+ "length_scale": 1,
131
+ "encoder_type": "fftransformer",
132
+ "encoder_params": {
133
+ "hidden_channels_ffn": 1024,
134
+ "num_heads": 1,
135
+ "num_layers": 6,
136
+ "dropout_p": 0.1
137
+ },
138
+ "decoder_type": "fftransformer",
139
+ "decoder_params": {
140
+ "hidden_channels_ffn": 1024,
141
+ "num_heads": 1,
142
+ "num_layers": 6,
143
+ "dropout_p": 0.1
144
+ },
145
+ "use_d_vector": false,
146
+ "d_vector_dim": 0,
147
+ "detach_duration_predictor": false,
148
+ "max_duration": 75,
149
+ "use_aligner": true
150
+ },
151
+ "d_vector_file": false,
152
+ "lr": 0.0001,
153
+ "grad_clip": 5.0,
154
+ "ssim_loss_alpha": 1.0,
155
+ "dur_loss_alpha": 1.0,
156
+ "spec_loss_alpha": 1.0,
157
+ "pitch_loss_alpha": 1.0,
158
+ "aligner_loss_alpha": 1.0,
159
+ "binary_align_loss_alpha": 1.0,
160
+ "binary_align_loss_start_step": 20000,
161
+ "r": 1,
162
+ "f0_cache_path": "/home/ubuntu/TTS/recipes/ljspeech/fast_pitch/f0_cache"
163
+ }
tts/tts_models--en--ljspeech--fast_pitch/model_file.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1779ef4ef9f9f3c016efee5925c0742393eb7c7183f6daae1928b88cbef294b8
3
+ size 458065485
tts/tts_models--en--ljspeech--glow-tts/config.json ADDED
@@ -0,0 +1,150 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model": "glow_tts",
3
+ "run_name": "glow-tts-rel_pos_transformer",
4
+ "run_description": "glow-tts model training with relation position transformer encoder.",
5
+
6
+ // AUDIO PARAMETERS
7
+ "audio":{
8
+ "fft_size": 1024, // number of stft frequency levels. Size of the linear spectogram frame.
9
+ "win_length": 1024, // stft window length in ms.
10
+ "hop_length": 256, // stft window hop-lengh in ms.
11
+ "frame_length_ms": null, // stft window length in ms.If null, 'win_length' is used.
12
+ "frame_shift_ms": null, // stft window hop-lengh in ms. If null, 'hop_length' is used.
13
+
14
+ // Audio processing parameters
15
+ "sample_rate": 22050, // DATASET-RELATED: wav sample-rate. If different than the original data, it is resampled.
16
+ "preemphasis": 0.0, // pre-emphasis to reduce spec noise and make it more structured. If 0.0, no -pre-emphasis.
17
+ "ref_level_db": 0, // reference level db, theoretically 20db is the sound of air.
18
+
19
+ // Griffin-Lim
20
+ "power": 1.1, // value to sharpen wav signals after GL algorithm.
21
+ "griffin_lim_iters": 60,// #griffin-lim iterations. 30-60 is a good range. Larger the value, slower the generation.
22
+
23
+ // Silence trimming
24
+ "do_trim_silence": true,// enable trimming of slience of audio as you load it. LJspeech (false), TWEB (false), Nancy (true)
25
+ "trim_db": 60, // threshold for timming silence. Set this according to your dataset.
26
+
27
+ // MelSpectrogram parameters
28
+ "num_mels": 80, // size of the mel spec frame.
29
+ "mel_fmin": 50.0, // minimum freq level for mel-spec. ~50 for male and ~95 for female voices. Tune for dataset!!
30
+ "mel_fmax": 7600.0, // maximum freq level for mel-spec. Tune for dataset!!
31
+ "spec_gain": 1.0, // scaler value appplied after log transform of spectrogram.00
32
+
33
+ // Normalization parameters
34
+ "signal_norm": false, // normalize spec values. Mean-Var normalization if 'stats_path' is defined otherwise range normalization defined by the other params.
35
+ "min_level_db": -100, // lower bound for normalization
36
+ "symmetric_norm": true, // move normalization to range [-1, 1]
37
+ "max_norm": 1.0, // scale normalization to range [-max_norm, max_norm] or [0, max_norm]
38
+ "clip_norm": true, // clip normalized values into the range.
39
+ "stats_path": null // DO NOT USE WITH MULTI_SPEAKER MODEL. scaler stats file computed by 'compute_statistics.py'. If it is defined, mean-std based notmalization is used and other normalization params are ignored
40
+ },
41
+
42
+ // VOCABULARY PARAMETERS
43
+ // if custom character set is not defined,
44
+ // default set in symbols.py is used
45
+ "characters": {
46
+ "pad": "_",
47
+ "eos": "~",
48
+ "bos": "^",
49
+ "characters": "ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz!'(),-.:;? ",
50
+ "punctuations": "!'(),-.:;? ",
51
+ "phonemes": "abcdefhijklmnopqrstuvwxyz\u00e6\u00e7\u00f0\u00f8\u0127\u014b\u0153\u01c0\u01c1\u01c2\u01c3\u0250\u0251\u0252\u0253\u0254\u0255\u0256\u0257\u0258\u0259\u025a\u025b\u025c\u025e\u025f\u0260\u0261\u0262\u0263\u0264\u0265\u0266\u0267\u0268\u026a\u026b\u026c\u026d\u026e\u026f\u0270\u0271\u0272\u0273\u0274\u0275\u0276\u0278\u0279\u027a\u027b\u027d\u027e\u0280\u0281\u0282\u0283\u0284\u0288\u0289\u028a\u028b\u028c\u028d\u028e\u028f\u0290\u0291\u0292\u0294\u0295\u0298\u0299\u029b\u029c\u029d\u029f\u02a1\u02a2\u02c8\u02cc\u02d0\u02d1\u02de\u03b2\u03b8\u03c7\u1d7b\u2c71"
52
+ },
53
+
54
+ "add_blank": false, // if true add a new token after each token of the sentence. This increases the size of the input sequence, but has considerably improved the prosody of the GlowTTS model.
55
+
56
+ // DISTRIBUTED TRAINING
57
+ "distributed":{
58
+ "backend": "nccl",
59
+ "url": "tcp:\/\/localhost:54321"
60
+ },
61
+
62
+ "reinit_layers": [], // give a list of layer names to restore from the given checkpoint. If not defined, it reloads all heuristically matching layers.
63
+
64
+ // MODEL PARAMETERS
65
+ // "use_mas": false, // use Monotonic Alignment Search if true. Otherwise use pre-computed attention alignments.
66
+ "hidden_channels_encoder": 192,
67
+ "hidden_channels_decoder": 192,
68
+ "hidden_channels_duration_predictor": 256,
69
+ "use_encoder_prenet": true,
70
+ "encoder_type": "rel_pos_transformer",
71
+ "encoder_params": {
72
+ "kernel_size":3,
73
+ "dropout_p": 0.1,
74
+ "num_layers": 6,
75
+ "num_heads": 2,
76
+ "hidden_channels_ffn": 768,
77
+ "input_length": null
78
+ },
79
+
80
+ // TRAINING
81
+ "batch_size": 32, // Batch size for training. Lower values than 32 might cause hard to learn attention. It is overwritten by 'gradual_training'.
82
+ "eval_batch_size":16,
83
+ "r": 1, // Number of decoder frames to predict per iteration. Set the initial values if gradual training is enabled.
84
+ "loss_masking": true, // enable / disable loss masking against the sequence padding.
85
+ "mixed_precision": true,
86
+ "data_dep_init_iter": 10,
87
+
88
+ // VALIDATION
89
+ "run_eval": true,
90
+ "test_delay_epochs": 0, //Until attention is aligned, testing only wastes computation time.
91
+ "test_sentences_file": null, // set a file to load sentences to be used for testing. If it is null then we use default english sentences.
92
+
93
+ // OPTIMIZER
94
+ "noam_schedule": true, // use noam warmup and lr schedule.
95
+ "grad_clip": 5.0, // upper limit for gradients for clipping.
96
+ "epochs": 10000, // total number of epochs to train.
97
+ "lr": 1e-3, // Initial learning rate. If Noam decay is active, maximum learning rate.
98
+ "wd": 0.000001, // Weight decay weight.
99
+ "warmup_steps": 4000, // Noam decay steps to increase the learning rate from 0 to "lr"
100
+ "seq_len_norm": false, // Normalize eash sample loss with its length to alleviate imbalanced datasets. Use it if your dataset is small or has skewed distribution of sequence lengths.
101
+
102
+ // TENSORBOARD and LOGGING
103
+ "print_step": 25, // Number of steps to log training on console.
104
+ "tb_plot_step": 100, // Number of steps to plot TB training figures.
105
+ "print_eval": false, // If True, it prints intermediate loss values in evalulation.
106
+ "save_step": 5000, // Number of training steps expected to save traninpg stats and checkpoints.
107
+ "checkpoint": true, // If true, it saves checkpoints per "save_step"
108
+ "tb_model_param_stats": false, // true, plots param stats per layer on tensorboard. Might be memory consuming, but good for debugging.
109
+
110
+ // DATA LOADING
111
+ "text_cleaner": "phoneme_cleaners",
112
+ "enable_eos_bos_chars": false, // enable/disable beginning of sentence and end of sentence chars.
113
+ "num_loader_workers": 4, // number of training data loader processes. Don't set it too big. 4-8 are good values.
114
+ "num_val_loader_workers": 4, // number of evaluation data loader processes.
115
+ "batch_group_size": 0, //Number of batches to shuffle after bucketing.
116
+ "min_seq_len": 3, // DATASET-RELATED: minimum text length to use in training
117
+ "max_seq_len": 500, // DATASET-RELATED: maximum text length
118
+ "compute_f0": false, // compute f0 values in data-loader
119
+ "use_noise_augment": true, //add a random noise to audio signal for augmentation at training .
120
+ "compute_input_seq_cache": true,
121
+
122
+ // PATHS
123
+ "output_path": "/home/erogol/Models/LJSpeech/",
124
+
125
+ // PHONEMES
126
+ "phoneme_cache_path": "/home/erogol/Models/phoneme_cache/", // phoneme computation is slow, therefore, it caches results in the given folder.
127
+ "use_phonemes": true, // use phonemes instead of raw characters. It is suggested for better pronounciation.
128
+ "phoneme_language": "en-us", // depending on your target language, pick one from https://github.com/bootphon/phonemizer#languages
129
+ "phonemizer": "gruut",
130
+
131
+ // MULTI-SPEAKER and GST
132
+ "use_speaker_embedding": false, // use speaker embedding to enable multi-speaker learning.
133
+ "use_external_speaker_embedding_file": false,
134
+ "style_wav_for_test": null, // path to style wav file to be used in TacotronGST inference.
135
+ "use_gst": false, // TACOTRON ONLY: use global style tokens
136
+
137
+ // DATASETS
138
+ "datasets": // List of datasets. They all merged and they get different speaker_ids.
139
+ [
140
+ {
141
+ "name": "ljspeech",
142
+ "path": "/home/erogol/Data/LJSpeech-1.1/",
143
+ "meta_file_train": "metadata.csv",
144
+ "meta_file_val": null
145
+ // "path_for_attn": "/home/erogol/Data/LJSpeech-1.1/alignments/"
146
+ }
147
+ ]
148
+ }
149
+
150
+
tts/tts_models--en--ljspeech--glow-tts/model_file.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:132f8c4d4dcb0307f04f6243139da5eeb170e29b69bc22167680241d8e6f97c8
3
+ size 343840040
tts/tts_models--en--ljspeech--neural_hmm/._config.json ADDED
Binary file (212 Bytes). View file
 
tts/tts_models--en--ljspeech--neural_hmm/._model_file.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2e4a10a7e6d1dc0d52951cb6746ee984a7e4809bfbf84eb56f34afe1d8aeb3e9
3
+ size 212
tts/tts_models--en--ljspeech--neural_hmm/._tts_models--en--ljspeech--neural_hmm ADDED
Binary file (312 Bytes). View file
 
tts/tts_models--en--ljspeech--neural_hmm/__MACOSX/._tts_models--en--ljspeech--neural_hmm ADDED
Binary file (312 Bytes). View file
 
tts/tts_models--en--ljspeech--neural_hmm/__MACOSX/tts_models--en--ljspeech--neural_hmm/._config.json ADDED
Binary file (212 Bytes). View file
 
tts/tts_models--en--ljspeech--neural_hmm/__MACOSX/tts_models--en--ljspeech--neural_hmm/._model_file.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2e4a10a7e6d1dc0d52951cb6746ee984a7e4809bfbf84eb56f34afe1d8aeb3e9
3
+ size 212
tts/tts_models--en--ljspeech--neural_hmm/config.json ADDED
@@ -0,0 +1,172 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "output_path": "/home/smehta/Projects/TTS/recipes/ljspeech/neuralhmm_tts",
3
+ "logger_uri": null,
4
+ "run_name": "neuralhmmtts_ljspeech",
5
+ "project_name": null,
6
+ "run_description": "\ud83d\udc38Coqui trainer run.",
7
+ "print_step": 1,
8
+ "plot_step": 1,
9
+ "model_param_stats": false,
10
+ "wandb_entity": null,
11
+ "dashboard_logger": "tensorboard",
12
+ "log_model_step": null,
13
+ "save_step": 500,
14
+ "save_n_checkpoints": 5000000,
15
+ "save_checkpoints": true,
16
+ "save_all_best": false,
17
+ "save_best_after": 10000,
18
+ "target_loss": null,
19
+ "print_eval": true,
20
+ "test_delay_epochs": -1,
21
+ "run_eval": true,
22
+ "run_eval_steps": 100,
23
+ "distributed_backend": "nccl",
24
+ "distributed_url": "tcp://localhost:54321",
25
+ "mixed_precision": true,
26
+ "epochs": 1000,
27
+ "batch_size": 28,
28
+ "eval_batch_size": 16,
29
+ "grad_clip": 40000.0,
30
+ "scheduler_after_epoch": true,
31
+ "lr": 0.001,
32
+ "optimizer": "Adam",
33
+ "optimizer_params": {
34
+ "weight_decay": 1e-06
35
+ },
36
+ "lr_scheduler": null,
37
+ "lr_scheduler_params": {},
38
+ "use_grad_scaler": false,
39
+ "cudnn_enable": true,
40
+ "cudnn_deterministic": false,
41
+ "cudnn_benchmark": false,
42
+ "training_seed": 54321,
43
+ "model": "NeuralHMM_TTS",
44
+ "num_loader_workers": 4,
45
+ "num_eval_loader_workers": 4,
46
+ "use_noise_augment": false,
47
+ "audio": {
48
+ "fft_size": 1024,
49
+ "win_length": 1024,
50
+ "hop_length": 256,
51
+ "frame_shift_ms": null,
52
+ "frame_length_ms": null,
53
+ "stft_pad_mode": "reflect",
54
+ "sample_rate": 22050,
55
+ "resample": false,
56
+ "preemphasis": 0.0,
57
+ "ref_level_db": 20,
58
+ "do_sound_norm": false,
59
+ "log_func": "np.log",
60
+ "do_trim_silence": true,
61
+ "trim_db": 60,
62
+ "do_rms_norm": false,
63
+ "db_level": null,
64
+ "power": 1.5,
65
+ "griffin_lim_iters": 60,
66
+ "num_mels": 80,
67
+ "mel_fmin": 0.0,
68
+ "mel_fmax": 8000.0,
69
+ "spec_gain": 1,
70
+ "do_amp_to_db_linear": true,
71
+ "do_amp_to_db_mel": true,
72
+ "pitch_fmax": 640.0,
73
+ "pitch_fmin": 1.0,
74
+ "signal_norm": false,
75
+ "min_level_db": -100,
76
+ "symmetric_norm": true,
77
+ "max_norm": 4.0,
78
+ "clip_norm": true,
79
+ "stats_path": null
80
+ },
81
+ "use_phonemes": true,
82
+ "phonemizer": "espeak",
83
+ "phoneme_language": "en-us",
84
+ "compute_input_seq_cache": false,
85
+ "text_cleaner": "phoneme_cleaners",
86
+ "enable_eos_bos_chars": false,
87
+ "test_sentences_file": "",
88
+ "phoneme_cache_path": "/home/smehta/Projects/TTS/recipes/ljspeech/neuralhmm_tts/phoneme_cache",
89
+ "characters": {
90
+ "characters_class": "TTS.tts.utils.text.characters.IPAPhonemes",
91
+ "vocab_dict": null,
92
+ "pad": "<PAD>",
93
+ "eos": "<EOS>",
94
+ "bos": "<BOS>",
95
+ "blank": "<BLNK>",
96
+ "characters": "iy\u0268\u0289\u026fu\u026a\u028f\u028ae\u00f8\u0258\u0259\u0275\u0264o\u025b\u0153\u025c\u025e\u028c\u0254\u00e6\u0250a\u0276\u0251\u0252\u1d7b\u0298\u0253\u01c0\u0257\u01c3\u0284\u01c2\u0260\u01c1\u029bpbtd\u0288\u0256c\u025fk\u0261q\u0262\u0294\u0274\u014b\u0272\u0273n\u0271m\u0299r\u0280\u2c71\u027e\u027d\u0278\u03b2fv\u03b8\u00f0sz\u0283\u0292\u0282\u0290\u00e7\u029dx\u0263\u03c7\u0281\u0127\u0295h\u0266\u026c\u026e\u028b\u0279\u027bj\u0270l\u026d\u028e\u029f\u02c8\u02cc\u02d0\u02d1\u028dw\u0265\u029c\u02a2\u02a1\u0255\u0291\u027a\u0267\u02b2\u025a\u02de\u026b",
97
+ "punctuations": "!'(),-.:;? ",
98
+ "phonemes": null,
99
+ "is_unique": false,
100
+ "is_sorted": true
101
+ },
102
+ "add_blank": false,
103
+ "batch_group_size": 0,
104
+ "loss_masking": null,
105
+ "min_audio_len": 512,
106
+ "max_audio_len": Infinity,
107
+ "min_text_len": 10,
108
+ "max_text_len": 500,
109
+ "compute_f0": false,
110
+ "compute_linear_spec": false,
111
+ "precompute_num_workers": 8,
112
+ "start_by_longest": false,
113
+ "shuffle": false,
114
+ "drop_last": false,
115
+ "datasets": [
116
+ {
117
+ "formatter": "ljspeech",
118
+ "dataset_name": "",
119
+ "path": "data/LJSpeech-1.1/",
120
+ "meta_file_train": "metadata.csv",
121
+ "ignored_speakers": null,
122
+ "language": "",
123
+ "phonemizer": "",
124
+ "meta_file_val": "",
125
+ "meta_file_attn_mask": ""
126
+ }
127
+ ],
128
+ "test_sentences": [
129
+ "Be a voice, not an echo."
130
+ ],
131
+ "eval_split_max_size": null,
132
+ "eval_split_size": 0.01,
133
+ "use_speaker_weighted_sampler": false,
134
+ "speaker_weighted_sampler_alpha": 1.0,
135
+ "use_language_weighted_sampler": false,
136
+ "language_weighted_sampler_alpha": 1.0,
137
+ "use_length_weighted_sampler": false,
138
+ "length_weighted_sampler_alpha": 1.0,
139
+ "force_generate_statistics": false,
140
+ "mel_statistics_parameter_path": "/home/smehta/Projects/TTS/recipes/ljspeech/neuralhmm_tts/lj_parameters.pt",
141
+ "num_chars": 131,
142
+ "state_per_phone": 2,
143
+ "encoder_in_out_features": 512,
144
+ "encoder_n_convolutions": 3,
145
+ "out_channels": 80,
146
+ "ar_order": 1,
147
+ "sampling_temp": 0.0,
148
+ "deterministic_transition": true,
149
+ "duration_threshold": 0.55,
150
+ "use_grad_checkpointing": true,
151
+ "max_sampling_time": 1000,
152
+ "prenet_type": "original",
153
+ "prenet_dim": 256,
154
+ "prenet_n_layers": 2,
155
+ "prenet_dropout": 0.5,
156
+ "prenet_dropout_at_inference": true,
157
+ "memory_rnn_dim": 1024,
158
+ "outputnet_size": [
159
+ 1024
160
+ ],
161
+ "flat_start_params": {
162
+ "mean": 0.0,
163
+ "std": 1.0,
164
+ "transition_p": 0.14
165
+ },
166
+ "std_floor": 0.001,
167
+ "r": 1,
168
+ "use_d_vector_file": false,
169
+ "use_speaker_embedding": false,
170
+ "restore_path": "/home/smehta/Projects/TTS/recipes/ljspeech/neuralhmm_tts/neuralhmmtts_ljspeech-January-09-2023_01+44PM-c56dce6b/checkpoint_158500.pth",
171
+ "github_branch": "* rebuttal"
172
+ }
tts/tts_models--en--ljspeech--neural_hmm/model_file.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0287896bd01519d1d8fa4958680b96ea132d0e630bff5d77bc232eac28515df5
3
+ size 183446015
tts/tts_models--en--ljspeech--overflow/._config.json ADDED
Binary file (212 Bytes). View file
 
tts/tts_models--en--ljspeech--overflow/._model_file.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:93ee90b5a99e2b350b2bfc92daf2518c067aea942f159689542d4a9d5c47288b
3
+ size 212
tts/tts_models--en--ljspeech--overflow/._tts_models--en--ljspeech--overflow ADDED
Binary file (312 Bytes). View file
 
tts/tts_models--en--ljspeech--overflow/__MACOSX/._tts_models--en--ljspeech--overflow ADDED
Binary file (312 Bytes). View file
 
tts/tts_models--en--ljspeech--overflow/__MACOSX/tts_models--en--ljspeech--overflow/._config.json ADDED
Binary file (212 Bytes). View file
 
tts/tts_models--en--ljspeech--overflow/__MACOSX/tts_models--en--ljspeech--overflow/._model_file.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:93ee90b5a99e2b350b2bfc92daf2518c067aea942f159689542d4a9d5c47288b
3
+ size 212
tts/tts_models--en--ljspeech--overflow/config.json ADDED
@@ -0,0 +1,185 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "output_path": "/data/TTS/recipes/ljspeech/overflow",
3
+ "logger_uri": null,
4
+ "run_name": "overflow_ljspeech",
5
+ "project_name": null,
6
+ "run_description": "\ud83d\udc38Coqui trainer run.",
7
+ "print_step": 1,
8
+ "plot_step": 1,
9
+ "model_param_stats": false,
10
+ "wandb_entity": null,
11
+ "dashboard_logger": "tensorboard",
12
+ "log_model_step": null,
13
+ "save_step": 500,
14
+ "save_n_checkpoints": 5,
15
+ "save_checkpoints": true,
16
+ "save_all_best": false,
17
+ "save_best_after": 10000,
18
+ "target_loss": null,
19
+ "print_eval": true,
20
+ "test_delay_epochs": -1,
21
+ "run_eval": true,
22
+ "run_eval_steps": 100,
23
+ "distributed_backend": "nccl",
24
+ "distributed_url": "tcp://localhost:54321",
25
+ "mixed_precision": true,
26
+ "epochs": 1000,
27
+ "batch_size": 30,
28
+ "eval_batch_size": 16,
29
+ "grad_clip": 40000.0,
30
+ "scheduler_after_epoch": true,
31
+ "lr": 0.001,
32
+ "optimizer": "Adam",
33
+ "optimizer_params": {
34
+ "weight_decay": 1e-06
35
+ },
36
+ "lr_scheduler": null,
37
+ "lr_scheduler_params": {},
38
+ "use_grad_scaler": false,
39
+ "cudnn_enable": true,
40
+ "cudnn_deterministic": false,
41
+ "cudnn_benchmark": false,
42
+ "training_seed": 54321,
43
+ "model": "OverFlow",
44
+ "num_loader_workers": 4,
45
+ "num_eval_loader_workers": 4,
46
+ "use_noise_augment": false,
47
+ "audio": {
48
+ "fft_size": 1024,
49
+ "win_length": 1024,
50
+ "hop_length": 256,
51
+ "frame_shift_ms": null,
52
+ "frame_length_ms": null,
53
+ "stft_pad_mode": "reflect",
54
+ "sample_rate": 22050,
55
+ "resample": false,
56
+ "preemphasis": 0.0,
57
+ "ref_level_db": 20,
58
+ "do_sound_norm": false,
59
+ "log_func": "np.log",
60
+ "do_trim_silence": true,
61
+ "trim_db": 60.0,
62
+ "do_rms_norm": false,
63
+ "db_level": null,
64
+ "power": 1.5,
65
+ "griffin_lim_iters": 60,
66
+ "num_mels": 80,
67
+ "mel_fmin": 0.0,
68
+ "mel_fmax": 8000,
69
+ "spec_gain": 1.0,
70
+ "do_amp_to_db_linear": true,
71
+ "do_amp_to_db_mel": true,
72
+ "pitch_fmax": 640.0,
73
+ "pitch_fmin": 1.0,
74
+ "signal_norm": false,
75
+ "min_level_db": -100,
76
+ "symmetric_norm": true,
77
+ "max_norm": 4.0,
78
+ "clip_norm": true,
79
+ "stats_path": null
80
+ },
81
+ "use_phonemes": true,
82
+ "phonemizer": "espeak",
83
+ "phoneme_language": "en-us",
84
+ "compute_input_seq_cache": false,
85
+ "text_cleaner": "phoneme_cleaners",
86
+ "enable_eos_bos_chars": false,
87
+ "test_sentences_file": "",
88
+ "phoneme_cache_path": "/data/TTS/recipes/ljspeech/overflow/phoneme_cache",
89
+ "characters": {
90
+ "characters_class": "TTS.tts.utils.text.characters.IPAPhonemes",
91
+ "vocab_dict": null,
92
+ "pad": "<PAD>",
93
+ "eos": "<EOS>",
94
+ "bos": "<BOS>",
95
+ "blank": "<BLNK>",
96
+ "characters": "iy\u0268\u0289\u026fu\u026a\u028f\u028ae\u00f8\u0258\u0259\u0275\u0264o\u025b\u0153\u025c\u025e\u028c\u0254\u00e6\u0250a\u0276\u0251\u0252\u1d7b\u0298\u0253\u01c0\u0257\u01c3\u0284\u01c2\u0260\u01c1\u029bpbtd\u0288\u0256c\u025fk\u0261q\u0262\u0294\u0274\u014b\u0272\u0273n\u0271m\u0299r\u0280\u2c71\u027e\u027d\u0278\u03b2fv\u03b8\u00f0sz\u0283\u0292\u0282\u0290\u00e7\u029dx\u0263\u03c7\u0281\u0127\u0295h\u0266\u026c\u026e\u028b\u0279\u027bj\u0270l\u026d\u028e\u029f\u02c8\u02cc\u02d0\u02d1\u028dw\u0265\u029c\u02a2\u02a1\u0255\u0291\u027a\u0267\u02b2\u025a\u02de\u026b",
97
+ "punctuations": "!'(),-.:;? ",
98
+ "phonemes": null,
99
+ "is_unique": false,
100
+ "is_sorted": true
101
+ },
102
+ "add_blank": false,
103
+ "batch_group_size": 0,
104
+ "loss_masking": null,
105
+ "min_audio_len": 1,
106
+ "max_audio_len": Infinity,
107
+ "min_text_len": 1,
108
+ "max_text_len": Infinity,
109
+ "compute_f0": false,
110
+ "compute_linear_spec": false,
111
+ "precompute_num_workers": 8,
112
+ "start_by_longest": false,
113
+ "datasets": [
114
+ {
115
+ "formatter": "ljspeech",
116
+ "dataset_name": "",
117
+ "path": "data/LJSpeech-1.1/LJSpeech-1.1",
118
+ "meta_file_train": "metadata.csv",
119
+ "ignored_speakers": null,
120
+ "language": "",
121
+ "meta_file_val": "",
122
+ "meta_file_attn_mask": ""
123
+ }
124
+ ],
125
+ "test_sentences": [
126
+ "It took me quite a long time to develop a voice, and now that I have it I'm not going to be silent.",
127
+ "Be a voice, not an echo.",
128
+ "I'm sorry Dave. I'm afraid I can't do that.",
129
+ "This cake is great. It's so delicious and moist.",
130
+ "Prior to November 22, 1963."
131
+ ],
132
+ "eval_split_max_size": null,
133
+ "eval_split_size": 0.01,
134
+ "use_speaker_weighted_sampler": false,
135
+ "speaker_weighted_sampler_alpha": 1.0,
136
+ "use_language_weighted_sampler": false,
137
+ "language_weighted_sampler_alpha": 1.0,
138
+ "use_length_weighted_sampler": false,
139
+ "length_weighted_sampler_alpha": 1.0,
140
+ "force_generate_statistics": false,
141
+ "mel_statistics_parameter_path": "/data/TTS/recipes/ljspeech/overflow/lj_parameters.pt",
142
+ "num_chars": 131,
143
+ "state_per_phone": 2,
144
+ "encoder_in_out_features": 512,
145
+ "encoder_n_convolutions": 3,
146
+ "out_channels": 80,
147
+ "ar_order": 1,
148
+ "sampling_temp": 0.334,
149
+ "deterministic_transition": true,
150
+ "duration_threshold": 0.55,
151
+ "use_grad_checkpointing": true,
152
+ "max_sampling_time": 1000,
153
+ "prenet_type": "original",
154
+ "prenet_dim": 256,
155
+ "prenet_n_layers": 2,
156
+ "prenet_dropout": 0.5,
157
+ "prenet_dropout_at_inference": false,
158
+ "memory_rnn_dim": 1024,
159
+ "outputnet_size": [
160
+ 256,
161
+ 256
162
+ ],
163
+ "flat_start_params": {
164
+ "mean": 0.0,
165
+ "std": 1.0,
166
+ "transition_p": 0.14
167
+ },
168
+ "std_floor": 0.01,
169
+ "hidden_channels_dec": 150,
170
+ "kernel_size_dec": 5,
171
+ "dilation_rate": 1,
172
+ "num_flow_blocks_dec": 12,
173
+ "num_block_layers": 4,
174
+ "dropout_p_dec": 0.05,
175
+ "num_splits": 4,
176
+ "num_squeeze": 2,
177
+ "sigmoid_scale": false,
178
+ "c_in_channels": 0,
179
+ "min_seq_len": 3,
180
+ "max_seq_len": 500,
181
+ "r": 1,
182
+ "use_d_vector_file": false,
183
+ "use_speaker_embedding": false,
184
+ "github_branch": "* pr/shivammehta25/2183"
185
+ }
tts/tts_models--en--ljspeech--overflow/model_file.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:444ed46bc352d550974506d0b7f888d930d2af3adce46dc14ba2b074e8de8750
3
+ size 327250447
tts/tts_models--en--ljspeech--tacotron2-DCA/config.json ADDED
@@ -0,0 +1,204 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "output_path": "/data/rw/home/Trainings/LJSpeech/",
3
+ "logger_uri": null,
4
+ "run_name": "ljspeech-dcattn",
5
+ "project_name": null,
6
+ "run_description": "tacotron2 with dynamic convolution attention.",
7
+ "print_step": 25,
8
+ "plot_step": 100,
9
+ "model_param_stats": false,
10
+ "wandb_entity": null,
11
+ "dashboard_logger": "tensorboard",
12
+ "log_model_step": null,
13
+ "save_step": 10000,
14
+ "save_n_checkpoints": 5,
15
+ "save_checkpoints": true,
16
+ "save_all_best": false,
17
+ "save_best_after": 10000,
18
+ "target_loss": null,
19
+ "print_eval": false,
20
+ "test_delay_epochs": 10,
21
+ "run_eval": true,
22
+ "run_eval_steps": null,
23
+ "distributed_backend": "nccl",
24
+ "distributed_url": "tcp://localhost:54321",
25
+ "mixed_precision": true,
26
+ "epochs": 1000,
27
+ "batch_size": 64,
28
+ "eval_batch_size": 16,
29
+ "grad_clip": 0.05,
30
+ "scheduler_after_epoch": true,
31
+ "lr": 0.001,
32
+ "optimizer": "RAdam",
33
+ "optimizer_params": {
34
+ "betas": [
35
+ 0.9,
36
+ 0.998
37
+ ],
38
+ "weight_decay": 1e-06
39
+ },
40
+ "lr_scheduler": "NoamLR",
41
+ "lr_scheduler_params": {
42
+ "warmup_steps": 4000
43
+ },
44
+ "use_grad_scaler": false,
45
+ "cudnn_enable": true,
46
+ "cudnn_deterministic": false,
47
+ "cudnn_benchmark": true,
48
+ "training_seed": 54321,
49
+ "model": "Tacotron2",
50
+ "num_loader_workers": 4,
51
+ "num_eval_loader_workers": 0,
52
+ "use_noise_augment": false,
53
+ "audio": {
54
+ "fft_size": 1024,
55
+ "win_length": 1024,
56
+ "hop_length": 256,
57
+ "frame_shift_ms": null,
58
+ "frame_length_ms": null,
59
+ "stft_pad_mode": "reflect",
60
+ "sample_rate": 22050,
61
+ "resample": false,
62
+ "preemphasis": 0.0,
63
+ "ref_level_db": 20,
64
+ "do_sound_norm": false,
65
+ "log_func": "np.log10",
66
+ "do_trim_silence": true,
67
+ "trim_db": 60,
68
+ "do_rms_norm": false,
69
+ "db_level": null,
70
+ "power": 1.5,
71
+ "griffin_lim_iters": 60,
72
+ "num_mels": 80,
73
+ "mel_fmin": 50.0,
74
+ "mel_fmax": 7600.0,
75
+ "spec_gain": 1,
76
+ "do_amp_to_db_linear": true,
77
+ "do_amp_to_db_mel": true,
78
+ "pitch_fmax": 640.0,
79
+ "pitch_fmin": 0.0,
80
+ "signal_norm": true,
81
+ "min_level_db": -100,
82
+ "symmetric_norm": true,
83
+ "max_norm": 4.0,
84
+ "clip_norm": true,
85
+ "stats_path": "C:/Users/Torch/AppData/Local\\tts\\tts_models--en--ljspeech--tacotron2-DCA\\scale_stats.npy"
86
+ },
87
+ "use_phonemes": true,
88
+ "phonemizer": "gruut",
89
+ "phoneme_language": "en-us",
90
+ "compute_input_seq_cache": true,
91
+ "text_cleaner": "phoneme_cleaners",
92
+ "enable_eos_bos_chars": false,
93
+ "test_sentences_file": null,
94
+ "phoneme_cache_path": "~/phoneme_cache/",
95
+ "characters": {
96
+ "characters_class": null,
97
+ "vocab_dict": null,
98
+ "pad": "_",
99
+ "eos": "~",
100
+ "bos": "^",
101
+ "blank": null,
102
+ "characters": "ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz!'(),-.:;? ",
103
+ "punctuations": "!'(),-.:;? ",
104
+ "phonemes": "abcdefhijklmnopqrstuvwxyz\u00e6\u00e7\u00f0\u00f8\u0127\u014b\u0153\u01c0\u01c1\u01c2\u01c3\u0250\u0251\u0252\u0253\u0254\u0255\u0256\u0257\u0258\u0259\u025a\u025b\u025c\u025e\u025f\u0260\u0261\u0262\u0263\u0264\u0265\u0266\u0267\u0268\u026a\u026b\u026c\u026d\u026e\u026f\u0270\u0271\u0272\u0273\u0274\u0275\u0276\u0278\u0279\u027a\u027b\u027d\u027e\u0280\u0281\u0282\u0283\u0284\u0288\u0289\u028a\u028b\u028c\u028d\u028e\u028f\u0290\u0291\u0292\u0294\u0295\u0298\u0299\u029b\u029c\u029d\u029f\u02a1\u02a2\u02c8\u02cc\u02d0\u02d1\u02de\u03b2\u03b8\u03c7\u1d7b\u2c71",
105
+ "is_unique": true,
106
+ "is_sorted": true
107
+ },
108
+ "add_blank": false,
109
+ "batch_group_size": 4,
110
+ "loss_masking": true,
111
+ "min_audio_len": 1,
112
+ "max_audio_len": Infinity,
113
+ "min_text_len": 1,
114
+ "max_text_len": Infinity,
115
+ "compute_f0": false,
116
+ "compute_energy": false,
117
+ "compute_linear_spec": false,
118
+ "precompute_num_workers": 0,
119
+ "start_by_longest": false,
120
+ "shuffle": false,
121
+ "drop_last": false,
122
+ "datasets": [
123
+ {
124
+ "formatter": "",
125
+ "dataset_name": "",
126
+ "path": "/root/LJSpeech-1.1/",
127
+ "meta_file_train": "metadata.csv",
128
+ "ignored_speakers": null,
129
+ "language": "",
130
+ "phonemizer": "",
131
+ "meta_file_val": null,
132
+ "meta_file_attn_mask": ""
133
+ }
134
+ ],
135
+ "test_sentences": [
136
+ "It took me quite a long time to develop a voice, and now that I have it I'm not going to be silent.",
137
+ "Be a voice, not an echo.",
138
+ "I'm sorry Dave. I'm afraid I can't do that.",
139
+ "This cake is great. It's so delicious and moist.",
140
+ "Prior to November 22, 1963."
141
+ ],
142
+ "eval_split_max_size": null,
143
+ "eval_split_size": 0.01,
144
+ "use_speaker_weighted_sampler": false,
145
+ "speaker_weighted_sampler_alpha": 1.0,
146
+ "use_language_weighted_sampler": false,
147
+ "language_weighted_sampler_alpha": 1.0,
148
+ "use_length_weighted_sampler": false,
149
+ "length_weighted_sampler_alpha": 1.0,
150
+ "use_gst": false,
151
+ "gst": {
152
+ "gst_style_input_wav": null,
153
+ "gst_style_input_weights": null,
154
+ "gst_embedding_dim": 512,
155
+ "gst_use_speaker_embedding": false,
156
+ "gst_num_heads": 4,
157
+ "gst_num_style_tokens": 10
158
+ },
159
+ "gst_style_input": null,
160
+ "use_capacitron_vae": false,
161
+ "capacitron_vae": null,
162
+ "num_speakers": 1,
163
+ "num_chars": 0,
164
+ "r": 2,
165
+ "gradual_training": null,
166
+ "memory_size": -1,
167
+ "prenet_type": "original",
168
+ "prenet_dropout": true,
169
+ "prenet_dropout_at_inference": false,
170
+ "stopnet": true,
171
+ "separate_stopnet": true,
172
+ "stopnet_pos_weight": 15.0,
173
+ "max_decoder_steps": 500,
174
+ "encoder_in_features": 512,
175
+ "decoder_in_features": 512,
176
+ "decoder_output_dim": 80,
177
+ "out_channels": 80,
178
+ "attention_type": "dynamic_convolution",
179
+ "attention_heads": 4,
180
+ "attention_norm": "softmax",
181
+ "attention_win": false,
182
+ "windowing": false,
183
+ "use_forward_attn": false,
184
+ "forward_attn_mask": false,
185
+ "transition_agent": false,
186
+ "location_attn": false,
187
+ "bidirectional_decoder": false,
188
+ "double_decoder_consistency": false,
189
+ "ddc_r": 7,
190
+ "speakers_file": null,
191
+ "use_speaker_embedding": false,
192
+ "speaker_embedding_dim": 512,
193
+ "use_d_vector_file": false,
194
+ "d_vector_file": false,
195
+ "d_vector_dim": null,
196
+ "seq_len_norm": false,
197
+ "decoder_loss_alpha": 0.25,
198
+ "postnet_loss_alpha": 0.25,
199
+ "postnet_diff_spec_alpha": 0.0,
200
+ "decoder_diff_spec_alpha": 0.0,
201
+ "decoder_ssim_alpha": 0.25,
202
+ "postnet_ssim_alpha": 0.25,
203
+ "ga_alpha": 0.0
204
+ }
tts/tts_models--en--ljspeech--tacotron2-DCA/model_file.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dfe9e59413636dc20955df27f171732996f57c7b060c05818ef61aeaa16910ad
3
+ size 339206517
tts/tts_models--en--ljspeech--tacotron2-DCA/scale_stats.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8c4a45b935563157509ddbff09f59e4ffea35e1d07f3bbf87ec21484cb275c4a
3
+ size 10491
tts/tts_models--en--ljspeech--vits/config.json ADDED
@@ -0,0 +1,203 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model": "vits",
3
+ "run_name": "",
4
+ "run_description": "",
5
+ "epochs": 10000,
6
+ "batch_size": null,
7
+ "eval_batch_size": null,
8
+ "mixed_precision": false,
9
+ "scheduler_after_epoch": true,
10
+ "run_eval": true,
11
+ "test_delay_epochs": 0,
12
+ "print_eval": false,
13
+ "print_step": 25,
14
+ "tb_plot_step": 100,
15
+ "tb_model_param_stats": false,
16
+ "save_step": 10000,
17
+ "checkpoint": true,
18
+ "keep_all_best": false,
19
+ "keep_after": 10000,
20
+ "num_loader_workers": null,
21
+ "num_eval_loader_workers": 0,
22
+ "use_noise_augment": false,
23
+ "output_path": null,
24
+ "distributed_backend": "nccl",
25
+ "distributed_url": "tcp://localhost:54321",
26
+ "audio": {
27
+ "fft_size": 1024,
28
+ "win_length": 1024,
29
+ "hop_length": 256,
30
+ "frame_shift_ms": null,
31
+ "frame_length_ms": null,
32
+ "stft_pad_mode": "reflect",
33
+ "sample_rate": 22050,
34
+ "resample": false,
35
+ "preemphasis": 0.0,
36
+ "ref_level_db": 20,
37
+ "do_sound_norm": false,
38
+ "log_func": "np.log10",
39
+ "do_trim_silence": true,
40
+ "trim_db": 45,
41
+ "power": 1.5,
42
+ "griffin_lim_iters": 60,
43
+ "num_mels": 80,
44
+ "mel_fmin": 0.0,
45
+ "mel_fmax": null,
46
+ "spec_gain": 20,
47
+ "do_amp_to_db_linear": true,
48
+ "do_amp_to_db_mel": true,
49
+ "signal_norm": true,
50
+ "min_level_db": -100,
51
+ "symmetric_norm": true,
52
+ "max_norm": 4.0,
53
+ "clip_norm": true,
54
+ "stats_path": null
55
+ },
56
+ "use_phonemes": true,
57
+ "phoneme_language": "en",
58
+ "compute_input_seq_cache": false,
59
+ "text_cleaner": "phoneme_cleaners",
60
+ "phonemizer": "espeak",
61
+ "enable_eos_bos_chars": false,
62
+ "test_sentences_file": "",
63
+ "phoneme_cache_path": null,
64
+ "characters":{
65
+ "characters_class": "TTS.tts.models.vits.VitsCharacters",
66
+ "pad": "_",
67
+ "eos": "",
68
+ "bos": "",
69
+ "characters": "ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz",
70
+ "punctuations":";:,.!?¡¿—…\"«»“” ",
71
+ "phonemes":"ɑɐɒæɓʙβɔɕçɗɖðʤəɘɚɛɜɝɞɟʄɡɠɢʛɦɧħɥʜɨɪʝɭɬɫɮʟɱɯɰŋɳɲɴøɵɸθœɶʘɹɺɾɻʀʁɽʂʃʈʧʉʊʋⱱʌɣɤʍχʎʏʑʐʒʔʡʕʢǀǁǂǃˈˌːˑʼʴʰʱʲʷˠˤ˞↓↑→↗↘'̩'ᵻ"
72
+ },
73
+ "batch_group_size": 0,
74
+ "loss_masking": null,
75
+ "min_seq_len": 13,
76
+ "max_seq_len": 200,
77
+ "compute_f0": false,
78
+ "compute_linear_spec": true,
79
+ "add_blank": true,
80
+ "datasets": [
81
+ {
82
+ "name": "",
83
+ "path": "",
84
+ "meta_file_train": "",
85
+ "ununsed_speakers": null,
86
+ "meta_file_val": "",
87
+ "meta_file_attn_mask": ""
88
+ }
89
+ ],
90
+ "optimizer": "AdamW",
91
+ "optimizer_params": {
92
+ "betas": [
93
+ 0.8,
94
+ 0.99
95
+ ],
96
+ "eps": 1e-09,
97
+ "weight_decay": 0.01
98
+ },
99
+ "lr_scheduler": "",
100
+ "lr_scheduler_params": {},
101
+ "test_sentences": [
102
+ "It took me quite a long time to develop a voice, and now that I have it I'm not going to be silent.",
103
+ "Be a voice, not an echo.",
104
+ "I'm sorry Dave. I'm afraid I can't do that.",
105
+ "This cake is great. It's so delicious and moist.",
106
+ "Prior to November 22, 1963."
107
+ ],
108
+ "use_speaker_embedding": false,
109
+ "use_d_vector_file": false,
110
+ "d_vector_dim": 0,
111
+ "model_args": {
112
+ "num_chars": 179,
113
+ "out_channels": 513,
114
+ "spec_segment_size": 32,
115
+ "hidden_channels": 192,
116
+ "hidden_channels_ffn_text_encoder": 768,
117
+ "num_heads_text_encoder": 2,
118
+ "num_layers_text_encoder": 6,
119
+ "kernel_size_text_encoder": 3,
120
+ "dropout_p_text_encoder": 0.1,
121
+ "kernel_size_posterior_encoder": 5,
122
+ "dilation_rate_posterior_encoder": 1,
123
+ "num_layers_posterior_encoder": 16,
124
+ "kernel_size_flow": 5,
125
+ "dilation_rate_flow": 1,
126
+ "num_layers_flow": 4,
127
+ "resblock_type_decoder": "1",
128
+ "resblock_kernel_sizes_decoder": [
129
+ 3,
130
+ 7,
131
+ 11
132
+ ],
133
+ "resblock_dilation_sizes_decoder": [
134
+ [
135
+ 1,
136
+ 3,
137
+ 5
138
+ ],
139
+ [
140
+ 1,
141
+ 3,
142
+ 5
143
+ ],
144
+ [
145
+ 1,
146
+ 3,
147
+ 5
148
+ ]
149
+ ],
150
+ "upsample_rates_decoder": [
151
+ 8,
152
+ 8,
153
+ 2,
154
+ 2
155
+ ],
156
+ "upsample_initial_channel_decoder": 512,
157
+ "upsample_kernel_sizes_decoder": [
158
+ 16,
159
+ 16,
160
+ 4,
161
+ 4
162
+ ],
163
+ "use_sdp": true,
164
+ "noise_scale": 1.0,
165
+ "inference_noise_scale": 0.667,
166
+ "length_scale": 1,
167
+ "noise_scale_dp": 1.0,
168
+ "inference_noise_scale_dp": 0.8,
169
+ "max_inference_len": null,
170
+ "init_discriminator": false,
171
+ "use_spectral_norm_disriminator": false,
172
+ "use_speaker_embedding": false,
173
+ "num_speakers": 0,
174
+ "speakers_file": null,
175
+ "speaker_embedding_channels": 256,
176
+ "use_d_vector_file": false,
177
+ "d_vector_dim": 0,
178
+ "detach_dp_input": true
179
+ },
180
+ "grad_clip": [
181
+ 5,
182
+ 5
183
+ ],
184
+ "lr_gen": 0.0002,
185
+ "lr_disc": 0.0002,
186
+ "lr_scheduler_gen": "ExponentialLR",
187
+ "lr_scheduler_gen_params": {
188
+ "gamma": 0.999875,
189
+ "last_epoch": -1
190
+ },
191
+ "lr_scheduler_disc": "ExponentialLR",
192
+ "lr_scheduler_disc_params": {
193
+ "gamma": 0.999875,
194
+ "last_epoch": -1
195
+ },
196
+ "kl_loss_alpha": 1.0,
197
+ "disc_loss_alpha": 1.0,
198
+ "gen_loss_alpha": 1.0,
199
+ "feat_loss_alpha": 1.0,
200
+ "mel_loss_alpha": 45.0,
201
+ "return_wav": true,
202
+ "r": 1
203
+ }
tts/tts_models--en--ljspeech--vits/model_file.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3b71d7bea7bd4f6745b87a695174ec486f633ad6c4fecdcff7513d76d1fbafe6
3
+ size 145588597