Adapting GPT, GPT-2 and BERT language models for speech recognition

X Zheng, C Zhang, PC Woodland - 2021 IEEE Automatic …, 2021 - ieeexplore.ieee.org
Language models (LMs) pre-trained on massive amounts of text, in particular bidirectional
encoder representations from Transformers (BERT), generative pre-training (GPT), and GPT …

Innovative BERT-based reranking language models for speech recognition

SH Chiu, B Chen - 2021 IEEE Spoken Language Technology …, 2021 - ieeexplore.ieee.org
More recently, Bidirectional Encoder Representations from Transformers (BERT) was
proposed and has achieved impressive success on many natural language processing …

Bayesian neural network language modeling for speech recognition

B Xue, S Hu, J Xu, M Geng, X Liu… - IEEE/ACM Transactions …, 2022 - ieeexplore.ieee.org
State-of-the-art neural network language models (NNLMs) represented by long short term
memory recurrent neural networks (LSTM-RNNs) and Transformers are becoming highly …

Can contextual biasing remain effective with Whisper and GPT-2?

G Sun, X Zheng, C Zhang, PC Woodland - arxiv preprint arxiv:2306.01942, 2023 - arxiv.org
End-to-end automatic speech recognition (ASR) and large language models, such as
Whisper and GPT-2, have recently been scaled to use vast amounts of training data. Despite …

Effect and analysis of large-scale language model rescoring on competitive asr systems

T Udagawa, M Suzuki, G Kurata, N Itoh… - arxiv preprint arxiv …, 2022 - arxiv.org
Large-scale language models (LLMs) such as GPT-2, BERT and RoBERTa have been
successfully applied to ASR N-best rescoring. However, whether or how they can benefit …

Bayesian transformer language models for speech recognition

B Xue, J Yu, J Xu, S Liu, S Hu, Z Ye… - ICASSP 2021-2021 …, 2021 - ieeexplore.ieee.org
State-of-the-art neural language models (LMs) represented by Transformers are highly
complex. Their use of fixed, deterministic parameter estimates fail to account for model …

[PDF][PDF] Rapid Speaker Adaptation for Conformer Transducer: Attention and Bias Are All You Need.

Y Huang, G Ye, J Li, Y Gong - Interspeech, 2021 - isca-archive.org
Conformer transducer achieves new state-of-the-art end-to-end (E2E) system performance
and has become increasingly appealing for production. In this paper, we study how to …

Applying llms for rescoring n-best asr hypotheses of casual conversations: Effects of domain adaptation and context carry-over

A Ogawa, N Kamo, K Matsuura, T Ashihara… - arxiv preprint arxiv …, 2024 - arxiv.org
Large language models (LLMs) have been successfully applied for rescoring automatic
speech recognition (ASR) hypotheses. However, their ability to rescore ASR hypotheses of …

Using natural language processing techniques to improve manual test case descriptions

M Viggiato, D Paas, C Buzon, CP Bezemer - Proceedings of the 44th …, 2022 - dl.acm.org
Despite the recent advancements in test automation, testing often remains a manual, and
costly, activity in many industries. Manual test cases, often described only in natural …

A parallelizable lattice rescoring strategy with neural language models

K Li, D Povey, S Khudanpur - ICASSP 2021-2021 IEEE …, 2021 - ieeexplore.ieee.org
This paper proposes a parallel computation strategy and a posterior-based lattice expansion
algorithm for efficient lattice rescoring with neural language models (LMs) for automatic …