From a12c56d7e7577597a0351815b1894688beeb84c2 Mon Sep 17 00:00:00 2001 From: xwaeaewcrhomesysplug <67897598+xwaeaewcrhomesysplug@users.noreply.github.com> Date: Thu, 3 Dec 2020 01:49:36 +0900 Subject: [PATCH] just add audio setup info --- README.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/README.md b/README.md index 4c63381f7..2190abc83 100644 --- a/README.md +++ b/README.md @@ -59,7 +59,6 @@ Details of each folder are as in follows: |VCTK_V3|V3|VCTK|No| |UNIVERSAL_V1|V1|Universal|No| -We provide the universal model with discriminator weights that can be used as a base for transfer learning to other datasets. ## Fine-Tuning 1. Generate mel-spectrograms in numpy format using [Tacotron2](https://github.com/NVIDIA/tacotron2) with teacher-forcing.
@@ -86,6 +85,7 @@ Example: Generated wav files are saved in `generated_files` by default.
You can change the path by adding `--output_dir` option. +Extra Note for user given audio:wav,is mono,22050khz. ## Inference for end-to-end speech synthesis 1. Make `test_mel_files` directory and copy generated mel-spectrogram files into the directory.