Holistic open-domain automatic dialogue evaluation: [https://github.com/e0397123/D-score]
Unified framework for speaker and utterance verification: [https://github.com/sn1ff1918/SUV]
Multi-level adaptive speech activity detector: [https://github.com/bidishasharma/MultiSAD/]
PESnQ: Perceptual evaluation of singing quality: [https://github.com/chitralekha18/PESnQ_APSIPA2017] [Paper]
Automatic sung-lyrics data annotation: [https://github.com/chitralekha18/AutomaticSungLyricsAnnotation_ISMIR2018.git] [Paper]
NUS AutoLyrixAlign: [https://github.com/chitralekha18/AutoLyrixAlign.git]
Emotional voice conversion and/or speaker identity conversion with non-parallel training data: [https://github.com/KunZhou9646/emotional-voice-conversion-with-CycleGAN-and-CWT-for-Spectrum-and-F0]
Speaker-independent emotional voice conversion based on conditional VAW-GAN and CWT: [https://github.com/KunZhou9646/Speaker-independent-emotional-voice-conversion-based-on-conditional-VAW-GAN-and-CWT]
Singing voice conversion with conditional VAW-GAN: [https://github.com/KunZhou9646/Singing-Voice-Conversion-with-conditional-VAW-GAN]
Transformer-based dialect identification: [https://github.com/LIN-WANQIU/ADI17]
- Multi-modal target speaker extraction with visual cues: [https://github.com/zexupan/MuSE]