Adapting GPT, GPT-2 and BERT language models for speech recognition
Language models (LMs) pre-trained on massive amounts of text, in particular bidirectional
encoder representations from Transformers (BERT), generative pre-training (GPT), and GPT …
encoder representations from Transformers (BERT), generative pre-training (GPT), and GPT …
Can contextual biasing remain effective with Whisper and GPT-2?
End-to-end automatic speech recognition (ASR) and large language models, such as
Whisper and GPT-2, have recently been scaled to use vast amounts of training data. Despite …
Whisper and GPT-2, have recently been scaled to use vast amounts of training data. Despite …
Bayesian neural network language modeling for speech recognition
State-of-the-art neural network language models (NNLMs) represented by long short term
memory recurrent neural networks (LSTM-RNNs) and Transformers are becoming highly …
memory recurrent neural networks (LSTM-RNNs) and Transformers are becoming highly …
Innovative BERT-based reranking language models for speech recognition
More recently, Bidirectional Encoder Representations from Transformers (BERT) was
proposed and has achieved impressive success on many natural language processing …
proposed and has achieved impressive success on many natural language processing …
Recovering from privacy-preserving masking with large language models
Model adaptation is crucial to handle the discrepancy between proxy training data and
actual users' data received. To effectively perform adaptation, textual data of users is …
actual users' data received. To effectively perform adaptation, textual data of users is …
Effect and analysis of large-scale language model rescoring on competitive asr systems
Large-scale language models (LLMs) such as GPT-2, BERT and RoBERTa have been
successfully applied to ASR N-best rescoring. However, whether or how they can benefit …
successfully applied to ASR N-best rescoring. However, whether or how they can benefit …
[PDF][PDF] Rapid Speaker Adaptation for Conformer Transducer: Attention and Bias Are All You Need.
Conformer transducer achieves new state-of-the-art end-to-end (E2E) system performance
and has become increasingly appealing for production. In this paper, we study how to …
and has become increasingly appealing for production. In this paper, we study how to …
Bayesian transformer language models for speech recognition
State-of-the-art neural language models (LMs) represented by Transformers are highly
complex. Their use of fixed, deterministic parameter estimates fail to account for model …
complex. Their use of fixed, deterministic parameter estimates fail to account for model …
Applying llms for rescoring n-best asr hypotheses of casual conversations: Effects of domain adaptation and context carry-over
Large language models (LLMs) have been successfully applied for rescoring automatic
speech recognition (ASR) hypotheses. However, their ability to rescore ASR hypotheses of …
speech recognition (ASR) hypotheses. However, their ability to rescore ASR hypotheses of …
Using natural language processing techniques to improve manual test case descriptions
Despite the recent advancements in test automation, testing often remains a manual, and
costly, activity in many industries. Manual test cases, often described only in natural …
costly, activity in many industries. Manual test cases, often described only in natural …