Convolutional neural networks-an extensive arena of deep learning. A comprehensive study
N Singh, H Sabrol - Archives of Computational Methods in Engineering, 2021 - Springer
Deep learning is an evolving expanse of machine learning. Machine learning is observing
its neoteric span as deep learning is steadily becoming the pioneer in this field. With the …
its neoteric span as deep learning is steadily becoming the pioneer in this field. With the …
Very deep convolutional networks for end-to-end speech recognition
Sequence-to-sequence models have shown success in end-to-end speech recognition.
However these models have only used shallow acoustic encoder networks. In our work, we …
However these models have only used shallow acoustic encoder networks. In our work, we …
Reward augmented maximum likelihood for neural structured prediction
A key problem in structured output prediction is enabling direct optimization of the task
reward function that matters for test evaluation. This paper presents a simple and …
reward function that matters for test evaluation. This paper presents a simple and …
Imputer: Sequence modelling via imputation and dynamic programming
This paper presents the Imputer, a neural sequence model that generates output sequences
iteratively via imputations. The Imputer is an iterative generation model, requiring only a …
iteratively via imputations. The Imputer is an iterative generation model, requiring only a …
Non-monotonic sequential text generation
Standard sequential generation methods assume a pre-specified generation order, such as
text generation methods which generate words from left to right. In this work, we propose a …
text generation methods which generate words from left to right. In this work, we propose a …
Learning with algorithmic supervision via continuous relaxations
The integration of algorithmic components into neural architectures has gained increased
attention recently, as it allows training neural networks with new forms of supervision such …
attention recently, as it allows training neural networks with new forms of supervision such …
Noisy parallel approximate decoding for conditional recurrent language model
K Cho - arxiv preprint arxiv:1605.03835, 2016 - arxiv.org
Recent advances in conditional recurrent language modelling have mainly focused on
network architectures (eg, attention mechanism), learning algorithms (eg, scheduled …
network architectures (eg, attention mechanism), learning algorithms (eg, scheduled …
Optimal completion distillation for sequence learning
We present Optimal Completion Distillation (OCD), a training procedure for optimizing
sequence to sequence models based on edit distance. OCD is efficient, has no hyper …
sequence to sequence models based on edit distance. OCD is efficient, has no hyper …
Аналитический обзор интегральных систем распознавания речи
НМ Марковников… - Информатика и …, 2018 - proceedings.spiiras.nw.ru
Аннотация Приведен аналитический обзор разновидностей интегральных (end-to-end)
систем для распознавания речи, методов их построения, обучения и оптимизации …
систем для распознавания речи, методов их построения, обучения и оптимизации …
[PDF][PDF] On Online Attention-Based Speech Recognition and Joint Mandarin Character-Pinyin Training.
In this paper, we explore the use of attention-based models for online speech recognition
without the usage of language models or searching. Our model is based on an attention …
without the usage of language models or searching. Our model is based on an attention …