An overview of deep-learning-based audio-visual speech enhancement and separation

D Michelsanti, ZH Tan, SX Zhang, Y Xu… - … on Audio, Speech …, 2021 - ieeexplore.ieee.org
Speech enhancement and speech separation are two related tasks, whose purpose is to
extract either one or more target speech signals, respectively, from a mixture of sounds …

Survey of deep learning paradigms for speech processing

KB Bhangale, M Kothandaraman - Wireless Personal Communications, 2022 - Springer
Over the past decades, a particular focus is given to research on machine learning
techniques for speech processing applications. However, in the past few years, research …

Dual-path rnn: efficient long sequence modeling for time-domain single-channel speech separation

Y Luo, Z Chen, T Yoshioka - ICASSP 2020-2020 IEEE …, 2020 - ieeexplore.ieee.org
Recent studies in deep learning-based speech separation have proven the superiority of
time-domain approaches to conventional time-frequency-based methods. Unlike the time …

Dual-path transformer network: Direct context-aware modeling for end-to-end monaural speech separation

J Chen, Q Mao, D Liu - arxiv preprint arxiv:2007.13975, 2020 - arxiv.org
The dominant speech separation models are based on complex recurrent or convolution
neural network that model speech sequences indirectly conditioning on context, such as …

SDR–half-baked or well done?

J Le Roux, S Wisdom, H Erdogan… - ICASSP 2019-2019 …, 2019 - ieeexplore.ieee.org
In speech enhancement and source separation, signal-to-noise ratio is a ubiquitous
objective measure of denoising/separation quality. A decade ago, the BSS_eval toolkit was …

Conv-tasnet: Surpassing ideal time–frequency magnitude masking for speech separation

Y Luo, N Mesgarani - IEEE/ACM transactions on audio, speech …, 2019 - ieeexplore.ieee.org
Single-channel, speaker-independent speech separation methods have recently seen great
progress. However, the accuracy, latency, and computational cost of such methods remain …

Tasnet: time-domain audio separation network for real-time, single-channel speech separation

Y Luo, N Mesgarani - 2018 IEEE International Conference on …, 2018 - ieeexplore.ieee.org
Robust speech processing in multi-talker environments requires effective speech
separation. Recent deep learning systems have made significant progress toward solving …

Wavesplit: End-to-end speech separation by speaker clustering

N Zeghidour, D Grangier - IEEE/ACM Transactions on Audio …, 2021 - ieeexplore.ieee.org
We introduce Wavesplit, an end-to-end source separation system. From a single mixture, the
model infers a representation for each source and then estimates each source signal given …

Ablation studies in artificial neural networks

R Meyes, M Lu, CW de Puiseau, T Meisen - arxiv preprint arxiv …, 2019 - arxiv.org
Ablation studies have been widely used in the field of neuroscience to tackle complex
biological systems such as the extensively studied Drosophila central nervous system, the …

Voice separation with an unknown number of multiple speakers

E Nachmani, Y Adi, L Wolf - International Conference on …, 2020 - proceedings.mlr.press
We present a new method for separating a mixed audio sequence, in which multiple voices
speak simultaneously. The new method employs gated neural networks that are trained to …