Search Results for author: Zongyang Du

Found 6 papers, 0 papers with code

Exploring speech style spaces with language models: Emotional TTS without emotion labels

no code implementations18 May 2024 Shreeram Suresh Chandra, Zongyang Du, Berrak Sisman

Many frameworks for emotional text-to-speech (E-TTS) rely on human-annotated emotion labels that are often inaccurate and difficult to obtain.

Transfer Learning

Revealing Emotional Clusters in Speaker Embeddings: A Contrastive Learning Strategy for Speech Emotion Recognition

no code implementations19 Jan 2024 Ismail Rasim Ulgen, Zongyang Du, Carlos Busso, Berrak Sisman

In order to leverage this information, we introduce a novel contrastive pretraining approach applied to emotion-unlabeled data for speech emotion recognition.

Contrastive Learning Speech Emotion Recognition

Disentanglement of Emotional Style and Speaker Identity for Expressive Voice Conversion

no code implementations20 Oct 2021 Zongyang Du, Berrak Sisman, Kun Zhou, Haizhou Li

Expressive voice conversion performs identity conversion for emotional speakers by jointly converting speaker identity and emotional style.

Disentanglement Voice Conversion

Spectrum and Prosody Conversion for Cross-lingual Voice Conversion with CycleGAN

no code implementations11 Aug 2020 Zongyang Du, Kun Zhou, Berrak Sisman, Haizhou Li

It relies on non-parallel training data from two different languages, hence, is more challenging than mono-lingual voice conversion.

Voice Conversion

Cannot find the paper you are looking for? You can Submit a new open access paper.