[PDF][PDF] Recent advances in end-to-end automatic speech recognition

J Li - APSIPA Transactions on Signal and Information …, 2022 - nowpublishers.com
Recently, the speech community is seeing a significant trend of moving from deep neural
network based hybrid modeling to end-to-end (E2E) modeling for automatic speech …

Adaptation algorithms for neural network-based speech recognition: An overview

P Bell, J Fainberg, O Klejch, J Li… - IEEE Open Journal …, 2020 - ieeexplore.ieee.org
We present a structured overview of adaptation algorithms for neural network-based speech
recognition, considering both hybrid hidden Markov model/neural network systems and end …

Deep transfer learning for automatic speech recognition: Towards better generalization

H Kheddar, Y Himeur, S Al-Maadeed, A Amira… - Knowledge-Based …, 2023 - Elsevier
Automatic speech recognition (ASR) has recently become an important challenge when
using deep learning (DL). It requires large-scale training datasets and high computational …

Overcoming catastrophic forgetting by incremental moment matching

SW Lee, JH Kim, J Jun, JW Ha… - Advances in neural …, 2017 - proceedings.neurips.cc
Catastrophic forgetting is a problem of neural networks that loses the information of the first
task after training the second task. Here, we propose a method, ie incremental moment …

Speaker-invariant training via adversarial learning

Z Meng, J Li, Z Chen, Y Zhao, V Mazalov… - … , Speech and Signal …, 2018 - ieeexplore.ieee.org
We propose a novel adversarial multi-task learning scheme, aiming at actively curtailing the
inter-talker feature variability while maximizing its senone discriminability so as to enhance …

Learning hidden unit contributions for unsupervised acoustic model adaptation

P Swietojanski, J Li, S Renals - IEEE/ACM Transactions on …, 2016 - ieeexplore.ieee.org
This work presents a broad study on the adaptation of neural network acoustic models by
means of learning hidden unit contributions (LHUC)—a method that linearly re-combines …

Adversarial teacher-student learning for unsupervised domain adaptation

Z Meng, J Li, Y Gong, BH Juang - 2018 IEEE International …, 2018 - ieeexplore.ieee.org
The teacher-student (T/S) learning has been shown effective in unsupervised domain
adaptation [1]. It is a form of transfer learning, not in terms of the transfer of recognition …

DNN speaker adaptation using parameterised sigmoid and ReLU hidden activation functions

C Zhang, PC Woodland - 2016 IEEE International Conference …, 2016 - ieeexplore.ieee.org
This paper investigates the use of parameterised sigmoid and rectified linear unit (ReLU)
hidden activation functions in deep neural network (DNN) speaker adaptation. The sigmoid …

[PDF][PDF] Rapid adaptation for deep neural networks through multi-task learning.

Z Huang, J Li, SM Siniscalchi, IF Chen, J Wu, CH Lee - Interspeech, 2015 - isca-archive.org
We propose a novel approach to addressing the adaptation effectiveness issue in parameter
adaptation for deep neural network (DNN) based acoustic models for automatic speech …

Develo** far-field speaker system via teacher-student learning

J Li, R Zhao, Z Chen, C Liu, X **ao… - … on Acoustics, Speech …, 2018 - ieeexplore.ieee.org
In this study, we develop the keyword spotting (KWS) and acoustic model (AM) components
in a far-field speaker system. Specifically, we use teacher-student (T/S) learning to adapt a …