Adapting GPT, GPT-2 and BERT language models for speech recognition
Language models (LMs) pre-trained on massive amounts of text, in particular bidirectional
encoder representations from Transformers (BERT), generative pre-training (GPT), and GPT …
encoder representations from Transformers (BERT), generative pre-training (GPT), and GPT …
Innovative BERT-based reranking language models for speech recognition
More recently, Bidirectional Encoder Representations from Transformers (BERT) was
proposed and has achieved impressive success on many natural language processing …
proposed and has achieved impressive success on many natural language processing …
Bayesian neural network language modeling for speech recognition
State-of-the-art neural network language models (NNLMs) represented by long short term
memory recurrent neural networks (LSTM-RNNs) and Transformers are becoming highly …
memory recurrent neural networks (LSTM-RNNs) and Transformers are becoming highly …
Can contextual biasing remain effective with Whisper and GPT-2?
End-to-end automatic speech recognition (ASR) and large language models, such as
Whisper and GPT-2, have recently been scaled to use vast amounts of training data. Despite …
Whisper and GPT-2, have recently been scaled to use vast amounts of training data. Despite …
Effect and analysis of large-scale language model rescoring on competitive asr systems
Large-scale language models (LLMs) such as GPT-2, BERT and RoBERTa have been
successfully applied to ASR N-best rescoring. However, whether or how they can benefit …
successfully applied to ASR N-best rescoring. However, whether or how they can benefit …
Bayesian transformer language models for speech recognition
State-of-the-art neural language models (LMs) represented by Transformers are highly
complex. Their use of fixed, deterministic parameter estimates fail to account for model …
complex. Their use of fixed, deterministic parameter estimates fail to account for model …
[PDF][PDF] Rapid Speaker Adaptation for Conformer Transducer: Attention and Bias Are All You Need.
Conformer transducer achieves new state-of-the-art end-to-end (E2E) system performance
and has become increasingly appealing for production. In this paper, we study how to …
and has become increasingly appealing for production. In this paper, we study how to …
Applying llms for rescoring n-best asr hypotheses of casual conversations: Effects of domain adaptation and context carry-over
Large language models (LLMs) have been successfully applied for rescoring automatic
speech recognition (ASR) hypotheses. However, their ability to rescore ASR hypotheses of …
speech recognition (ASR) hypotheses. However, their ability to rescore ASR hypotheses of …
Using natural language processing techniques to improve manual test case descriptions
Despite the recent advancements in test automation, testing often remains a manual, and
costly, activity in many industries. Manual test cases, often described only in natural …
costly, activity in many industries. Manual test cases, often described only in natural …
A parallelizable lattice rescoring strategy with neural language models
This paper proposes a parallel computation strategy and a posterior-based lattice expansion
algorithm for efficient lattice rescoring with neural language models (LMs) for automatic …
algorithm for efficient lattice rescoring with neural language models (LMs) for automatic …