Human Language Technology

Resources

Code

  1. Unified framework for speaker and utterance verification: [https://github.com/sn1ff1918/SUV]
  2. Multi-level Adaptive Speech Activity Detector: [https://github.com/bidishasharma/MultiSAD/]
  3. PESnQ: Perceptual Evaluation of Singing Quality: [https://github.com/chitralekha18/PESnQ_APSIPA2017] [Paper]
  4. Automatic Sung-Lyrics Data Annotation: [https://github.com/chitralekha18/AutomaticSungLyricsAnnotation_ISMIR2018.git] [Paper]
  5. NUS AutoLyrixAlign: [https://github.com/chitralekha18/AutoLyrixAlign.git]
  6. Emotional voice conversion and/or speaker identity conversion with non-parallel training data: [https://github.com/KunZhou9646/emotional-voice-conversion-with-CycleGAN-and-CWT-for-Spectrum-and-F0]
  7. Speaker-independent emotional voice conversion based on conditional VAW-GAN and CWT: [https://github.com/KunZhou9646/Speaker-independent-emotional-voice-conversion-based-on-conditional-VAW-GAN-and-CWT]
  8. Singing voice conversion with conditional VAW-GAN: [https://github.com/KunZhou9646/Singing-Voice-Conversion-with-conditional-VAW-GAN]

Data Set

  1. NHSS: A Speech and Singing Parallel Database: [https://hltnus.github.io/NHSSDatabase/index.html]
  2. Solo singing damp dataset with aligned lyrics: [https://github.com/chitralekha18/lyrics-aligned-solo-singing-dataset]
  3. Pronunciation evaluation in singing: [https://github.com/chitralekha18/Dataset-for-pronunciation-evaluation-in-singing]
  4. RSL2019: A Realistic Speech Localization Corpus: [https://bidishasharma.github.io/RSL2019/]

Demo

  1. Robust Sound Recognition: A Neuromorphic Approach: [https://youtu.be/MIVvNb0sWOM]
  2. Speak-to-Sing: [https://speak-to-sing.hltnus.org/] [Poster]
  3. MuSigPro: Automatic Leaderboard Generation of Singers using Reference-Independent Singing Quality Evaluation Methods: [https://youtu.be/IAlsECqd9IE]
  4. AutoLyrixAlign: Automatic lyrics-to-audio alignment system for polyphonic music audio