Language modeling for code-mixing: The role of linguistic theory based synthetic data
Training language models for Code-mixed (CM) language is known to be a difficult problem
because of lack of data compounded by the increased confusability due to the presence of …
because of lack of data compounded by the increased confusability due to the presence of …
A review of the Mandarin-English code-switching corpus: SEAME
In this paper, we report the development of the South East Asia Mandarin-English (SEAME)
corpus, including 63 hours of transcribed spontaneous Mandarin-English code-switching …
corpus, including 63 hours of transcribed spontaneous Mandarin-English code-switching …
Lstm language models for lvcsr in first-pass decoding and lattice-rescoring
LSTM based language models are an important part of modern LVCSR systems as they
significantly improve performance over traditional backoff language models. Incorporating …
significantly improve performance over traditional backoff language models. Incorporating …
[PDF][PDF] Curriculum design for code-switching: Experiments with language identification and language modeling with deep neural networks
Curriculum learning strategies are known to improve the accuracy, robustness and
convergence rate for various language learning tasks using deep architectures (Bengio et …
convergence rate for various language learning tasks using deep architectures (Bengio et …
Semi-supervised adaptation of assistant based speech recognition models for different approach areas
M Kleinert, H Helmke, G Siol, H Ehr… - 2018 IEEE/AIAA 37th …, 2018 - ieeexplore.ieee.org
Air Navigation Service Providers (ANSPs) replace paper flight strips through different digital
solutions. The instructed commands from an air traffic controller (ATCos) are then available …
solutions. The instructed commands from an air traffic controller (ATCos) are then available …
Improvements to n-gram language model using text generated from neural language model
Although neural language models have emerged, n-gram language models are still used for
many speech recognition tasks. This paper proposes four methods to improve n-gram …
many speech recognition tasks. This paper proposes four methods to improve n-gram …
[PDF][PDF] Approximated and Domain-Adapted LSTM Language Models for First-Pass Decoding in Speech Recognition.
Abstract Traditionally, short-range Language Models (LMs) like the conventional n-gram
models have been used for language model adaptation. Recent work has improved …
models have been used for language model adaptation. Recent work has improved …
Improving N-gram language models with pre-trained deep transformer
Although n-gram language models (LMs) have been outperformed by the state-of-the-art
neural LMs, they are still widely used in speech recognition due to its high efficiency in …
neural LMs, they are still widely used in speech recognition due to its high efficiency in …
[PDF][PDF] Iterative Learning of Speech Recognition Models for Air Traffic Control.
Abstract Automatic Speech Recognition (ASR) has recently proved to be a useful tool to
reduce the workload of air traffic controllers leading to significant gains in operational …
reduce the workload of air traffic controllers leading to significant gains in operational …
On the N-gram Approximation of Pre-trained Language Models
Large pre-trained language models (PLMs) have shown remarkable performance across
various natural language understanding (NLU) tasks, particularly in low-resource settings …
various natural language understanding (NLU) tasks, particularly in low-resource settings …