From a12c56d7e7577597a0351815b1894688beeb84c2 Mon Sep 17 00:00:00 2001
From: xwaeaewcrhomesysplug
<67897598+xwaeaewcrhomesysplug@users.noreply.github.com>
Date: Thu, 3 Dec 2020 01:49:36 +0900
Subject: [PATCH] just add audio setup info
---
README.md | 2 +-
1 file changed, 1 insertion(+), 1 deletion(-)
diff --git a/README.md b/README.md
index 4c63381f7..2190abc83 100644
--- a/README.md
+++ b/README.md
@@ -59,7 +59,6 @@ Details of each folder are as in follows:
|VCTK_V3|V3|VCTK|No|
|UNIVERSAL_V1|V1|Universal|No|
-We provide the universal model with discriminator weights that can be used as a base for transfer learning to other datasets.
## Fine-Tuning
1. Generate mel-spectrograms in numpy format using [Tacotron2](https://github.com/NVIDIA/tacotron2) with teacher-forcing.
@@ -86,6 +85,7 @@ Example:
Generated wav files are saved in `generated_files` by default.
You can change the path by adding `--output_dir` option.
+Extra Note for user given audio:wav,is mono,22050khz.
## Inference for end-to-end speech synthesis
1. Make `test_mel_files` directory and copy generated mel-spectrogram files into the directory.