Turnitin
降AI改写
早检测系统
早降重系统
Turnitin-UK版
万方检测-期刊版
维普编辑部版
Grammarly检测
Paperpass检测
checkpass检测
PaperYY检测
[PDF][PDF] Recent advances in end-to-end automatic speech recognition
J Li - APSIPA Transactions on Signal and Information …, 2022 - nowpublishers.com
Recently, the speech community is seeing a significant trend of moving from deep neural
network based hybrid modeling to end-to-end (E2E) modeling for automatic speech …
network based hybrid modeling to end-to-end (E2E) modeling for automatic speech …
Self-supervised representation learning: Introduction, advances, and challenges
Self-supervised representation learning (SSRL) methods aim to provide powerful, deep
feature learning without the requirement of large annotated data sets, thus alleviating the …
feature learning without the requirement of large annotated data sets, thus alleviating the …
Scaling speech technology to 1,000+ languages
Expanding the language coverage of speech technology has the potential to improve
access to information for many more people. However, current speech technology is …
access to information for many more people. However, current speech technology is …
End-to-end speech recognition: A survey
In the last decade of automatic speech recognition (ASR) research, the introduction of deep
learning has brought considerable reductions in word error rate of more than 50% relative …
learning has brought considerable reductions in word error rate of more than 50% relative …
Modular deep learning
Transfer learning has recently become the dominant paradigm of machine learning. Pre-
trained models fine-tuned for downstream tasks achieve better performance with fewer …
trained models fine-tuned for downstream tasks achieve better performance with fewer …
Unsupervised cross-lingual representation learning for speech recognition
This paper presents XLSR which learns cross-lingual speech representations by pretraining
a single model from the raw waveform of speech in multiple languages. We build on …
a single model from the raw waveform of speech in multiple languages. We build on …
Exploring wav2vec 2.0 on speaker verification and language identification
Z Fan, M Li, S Zhou, B Xu - arxiv preprint arxiv:2012.06185, 2020 - arxiv.org
Wav2vec 2.0 is a recently proposed self-supervised framework for speech representation
learning. It follows a two-stage training process of pre-training and fine-tuning, and performs …
learning. It follows a two-stage training process of pre-training and fine-tuning, and performs …
mslam: Massively multilingual joint pre-training for speech and text
We present mSLAM, a multilingual Speech and LAnguage Model that learns cross-lingual
cross-modal representations of speech and text by pre-training jointly on large amounts of …
cross-modal representations of speech and text by pre-training jointly on large amounts of …
Improving continuous sign language recognition with cross-lingual signs
This work dedicates to continuous sign language recognition (CSLR), which is a weakly
supervised task dealing with the recognition of continuous signs from videos, without any …
supervised task dealing with the recognition of continuous signs from videos, without any …
Efficient adapter transfer of self-supervised speech models for automatic speech recognition
Self-supervised learning (SSL) is a powerful tool that allows learning of underlying
representations from unlabeled data. Transformer based models such as wav2vec 2.0 and …
representations from unlabeled data. Transformer based models such as wav2vec 2.0 and …