1 code implementation • 14 Apr 2024 • Jiang Li, Xiangdong Su, Yeyun Gong, Guanglai Gao
Recent studies have highlighted the effectiveness of tensor decomposition methods in the Temporal Knowledge Graphs Embedding (TKGE) task.
1 code implementation • 10 Mar 2024 • Qiuyu Liang, Weihua Wang, Feilong Bao, Guanglai Gao
Specifically, we map the learned features of graph nodes into hyperbolic space, and then perform a Lorentzian linear feature transformation to capture the underlying tree-like structure of data.
1 code implementation • 21 Sep 2023 • Qi Fan, Haolin Zuo, Rui Liu, Zheng Lian, Guanglai Gao
This approach includes two pivotal components: firstly, a noise scheduler that adjusts the type and level of noise in the data to emulate various realistic incomplete situations.
2 code implementations • 26 Jun 2023 • Jiang Li, Xiangdong Su, Fujun Zhang, Guanglai Gao
This paper presents a translation-based knowledge geraph embedding method via efficient relation rotation (TransERR), a straightforward yet effective alternative to traditional translation-based knowledge graph embedding models.
Ranked #16 on Link Property Prediction on ogbl-wikikg2
1 code implementation • 25 May 2023 • Rui Liu, Jinhua Zhang, Guanglai Gao, Haizhou Li
In this paper, we propose a novel ADD model, termed as M2S-ADD, that attempts to discover audio authenticity cues during the mono-to-stereo conversion process.
1 code implementation • 11 Dec 2022 • Kailin Liang, Bin Liu, Yifan Hu, Rui Liu, Feilong Bao, Guanglai Gao
Text-to-Speech (TTS) synthesis for low-resource languages is an attractive research issue in academia and industry nowadays.
1 code implementation • 27 Oct 2022 • Haolin Zuo, Rui Liu, Jinming Zhao, Guanglai Gao, Haizhou Li
Multimodal emotion recognition leverages complementary information across modalities to gain performance.
no code implementations • 27 Oct 2022 • Rui Liu, Haolin Zuo, De Hu, Guanglai Gao, Haizhou Li
Accented text-to-speech (TTS) synthesis seeks to generate speech with an accent (L2) as a variant of the standard version (L1).
1 code implementation • 27 Oct 2022 • Yifan Hu, Rui Liu, Guanglai Gao, Haizhou Li
Therefore, we propose a novel expressive conversational TTS model, termed as FCTalker, that learn the fine and coarse grained context dependency at the same time during speech generation.
no code implementations • 24 Sep 2022 • Muhan Na, Rui Liu, Feilong, Guanglai Gao
To answer this question, this paper investigates the utility of these two powerful techniques for CTMBC task combined with agglutinative characteristics of Mongolian language.
1 code implementation • 22 Sep 2022 • Yifan Hu, Pengkai Yin, Rui Liu, Feilong Bao, Guanglai Gao
This paper introduces a high-quality open-source text-to-speech (TTS) synthesis dataset for Mongolian, a low-resource language spoken by over 10 million people worldwide.
no code implementations • 22 Sep 2022 • Rui Liu, Berrak Sisman, Guanglai Gao, Haizhou Li
Accented TTS synthesis is challenging as L2 is different from L1 in both in terms of phonetic rendering and prosody pattern.
1 code implementation • 15 Jun 2022 • Rui Liu, Berrak Sisman, Björn Schuller, Guanglai Gao, Haizhou Li
In this paper, we propose a data-driven deep learning model, i. e. StrengthNet, to improve the generalization of emotion strength assessment for seen and unseen speech.
no code implementations • 26 Mar 2021 • Hao Li, Xueliang Zhang, Guanglai Gao
Another way is to use an anchor speech, a short speech of the target speaker, to model the speaker identity.
no code implementations • COLING 2020 • Na Liu, Xiangdong Su, Haoran Zhang, Guanglai Gao, Feilong Bao
The inner-word encoder uses the self-attention mechanisms to capture the inner-word features of the target word.
no code implementations • 11 Aug 2020 • Rui Liu, Berrak Sisman, Feilong Bao, Guanglai Gao, Haizhou Li
We propose a multi-task learning scheme for Tacotron training, that optimizes the system to predict both Mel spectrum and phrase breaks.
no code implementations • 11 Jun 2020 • Huali Xu, Xiangdong Su, Meng Wang, Xiang Hao, Guanglai Gao
The mask shrinking strategy is employed in the image completion model to track the areas to be repaired.
no code implementations • 29 May 2020 • Xiang Hao, Shixue Wen, Xiangdong Su, Yun Liu, Guanglai Gao, Xiaofei Li
In single-channel speech enhancement, methods based on full-band spectral features have been widely studied.
no code implementations • 29 May 2020 • Xiang Hao, Xiangdong Su, Zhiyu Wang, Qiang Zhang, Huali Xu, Guanglai Gao
Specifically, this method consists of multiple teacher models and a student model.
no code implementations • 2 Feb 2020 • Rui Liu, Berrak Sisman, Feilong Bao, Guanglai Gao, Haizhou Li
To address this problem, we propose a new training scheme for Tacotron-based TTS, referred to as WaveTTS, that has 2 loss functions: 1) time-domain loss, denoted as the waveform loss, that measures the distortion between the natural and generated waveform; and 2) frequency-domain loss, that measures the Mel-scale acoustic feature loss between the natural and generated acoustic features.
no code implementations • 7 Nov 2019 • Rui Liu, Berrak Sisman, Jingdong Li, Feilong Bao, Guanglai Gao, Haizhou Li
We first train a Tacotron2-based TTS model by always providing natural speech frames to the decoder, that serves as a teacher model.
no code implementations • COLING 2018 • Rui Liu, Feilong Bao, Guanglai Gao, HUI ZHANG, Yonghe Wang
In this paper, we first utilize the word embedding that focuses on sub-word units to the Mongolian Phrase Break (PB) prediction task by using Long-Short-Term-Memory (LSTM) model.
no code implementations • COLING 2016 • Weihua Wang, Feilong Bao, Guanglai Gao
The system based on segmenting suffixes with all proposed features yields benchmark result of F-measure=84. 65 on this corpus.