CMGAN: Conformer-Based Metric-GAN for Monaural Speech Enhancement

22 Sep 2022  ยท  Sherif Abdulatif, Ruizhe Cao, Bin Yang ยท

In this work, we further develop the conformer-based metric generative adversarial network (CMGAN) model for speech enhancement (SE) in the time-frequency (TF) domain. This paper builds on our previous work but takes a more in-depth look by conducting extensive ablation studies on model inputs and architectural design choices. We rigorously tested the generalization ability of the model to unseen noise types and distortions. We have fortified our claims through DNS-MOS measurements and listening tests. Rather than focusing exclusively on the speech denoising task, we extend this work to address the dereverberation and super-resolution tasks. This necessitated exploring various architectural changes, specifically metric discriminator scores and masking techniques. It is essential to highlight that this is among the earliest works that attempted complex TF-domain super-resolution. Our findings show that CMGAN outperforms existing state-of-the-art methods in the three major speech enhancement tasks: denoising, dereverberation, and super-resolution. For example, in the denoising task using the Voice Bank+DEMAND dataset, CMGAN notably exceeded the performance of prior models, attaining a PESQ score of 3.41 and an SSNR of 11.10 dB. Audio samples and CMGAN implementations are available online.

PDF Abstract
Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Audio Super-Resolution VCTK Multi-Speaker CMGAN Log-Spectral Distance 0.76 # 1
Speech Enhancement VoiceBank + DEMAND CMGAN PESQ 3.41 # 5
CSIG 4.63 # 4
CBAK 3.94 # 3
COVL 4.12 # 4
STOI 96 # 1
SSNR 11.1 # 1

Methods


No methods listed for this paper. Add relevant methods here