This is the implementation of the paper "Converting anyone's emotion: steps towards speaker-independent emotional voice conversion". Please kindly cite our paper if you are using our codes.
- Ubuntu 16.04
- Python 3.6
- Tensorflow-gpu 1.5.0
- PyWorld
- librosa
- soundfile
- numpy 1.14.0
- sklearn
- glob
- sprocket-vc
- pycwt
- scipy
- Activate your virtual enviroment.
source activate [your env]
- Train VAW-GAN for prosody.
./train_f0.sh
- Train VAW-GAN for spectrum.
./train_sp.sh
- Generate the converted emotional speech.
./convert.sh
Note: The codes are based on VAW-GAN Voice Conversion: https://github.com/JeremyCCHsu/vae-npvc/tree/vawgan