Update note to avoid using wrong models

This commit is contained in:
Weijia Chen 2021-08-11 09:51:02 +08:00 committed by GitHub
parent 67d301fc5b
commit c21d2c11dd
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23

View File

@ -27,9 +27,10 @@
* Install [ffmpeg](https://ffmpeg.org/download.html#get-packages). * Install [ffmpeg](https://ffmpeg.org/download.html#get-packages).
* Run `pip install -r requirements.txt` to install the remaining necessary packages. * Run `pip install -r requirements.txt` to install the remaining necessary packages.
### 2. reuse the pretrained encoder/vocoder ### 2. Reuse the pretrained encoder/vocoder
* Download the following models and extract to the root directory of this project. * Download the following models and extract to the root directory of this project. Don't use the synthesizer
https://github.com/CorentinJ/Real-Time-Voice-Cloning/wiki/Pretrained-models https://github.com/CorentinJ/Real-Time-Voice-Cloning/wiki/Pretrained-models
> Note that we need to specify the newly trained synthesizer model, since the original model is incompatible with the Chinese sympols. It means the demo_cli is not working at this moment.
### 3. Train synthesizer with aidatatang_200zh ### 3. Train synthesizer with aidatatang_200zh
* Download aidatatang_200zh dataset and unzip: make sure you can access all .wav in *train* folder * Download aidatatang_200zh dataset and unzip: make sure you can access all .wav in *train* folder
* Preprocess with the audios and the mel spectrograms: * Preprocess with the audios and the mel spectrograms: