Turnitin
降AI改写
早检测系统
早降重系统
Turnitin-UK版
万方检测-期刊版
维普编辑部版
Grammarly检测
Paperpass检测
checkpass检测
PaperYY检测
A review of deep learning techniques for speech processing
The field of speech processing has undergone a transformative shift with the advent of deep
learning. The use of multiple processing layers has enabled the creation of models capable …
learning. The use of multiple processing layers has enabled the creation of models capable …
Self-supervised speech representation learning: A review
Although supervised deep learning has revolutionized speech and audio processing, it has
necessitated the building of specialist models for individual tasks and application scenarios …
necessitated the building of specialist models for individual tasks and application scenarios …
An unsupervised autoregressive model for speech representation learning
This paper proposes a novel unsupervised autoregressive neural model for learning generic
speech representations. In contrast to other speech representation learning methods that …
speech representations. In contrast to other speech representation learning methods that …
Generative pre-training for speech with autoregressive predictive coding
Learning meaningful and general representations from unannotated speech that are
applicable to a wide range of tasks remains challenging. In this paper we propose to use …
applicable to a wide range of tasks remains challenging. In this paper we propose to use …
Non-autoregressive predictive coding for learning speech representations from local dependencies
Self-supervised speech representations have been shown to be effective in a variety of
speech applications. However, existing representation learning methods generally rely on …
speech applications. However, existing representation learning methods generally rely on …
Unsupervised pre-training of bidirectional speech encoders via masked reconstruction
We propose an approach for pre-training speech representations via a masked
reconstruction loss. Our pre-trained encoder networks are bidirectional and can therefore be …
reconstruction loss. Our pre-trained encoder networks are bidirectional and can therefore be …
Learning hierarchical discrete linguistic units from visually-grounded speech
In this paper, we present a method for learning discrete linguistic units by incorporating
vector quantization layers into neural models of visually grounded speech. We show that our …
vector quantization layers into neural models of visually grounded speech. We show that our …
Improved speech representations with multi-target autoregressive predictive coding
Training objectives based on predictive coding have recently been shown to be very
effective at learning meaningful representations from unlabeled speech. One example is …
effective at learning meaningful representations from unlabeled speech. One example is …
Pre-training audio representations with self-supervision
M Tagliasacchi, B Gfeller… - IEEE Signal …, 2020 - ieeexplore.ieee.org
We explore self-supervision as a way to learn general purpose audio representations.
Specifically, we propose two self-supervised tasks: Audio2Vec, which aims at reconstructing …
Specifically, we propose two self-supervised tasks: Audio2Vec, which aims at reconstructing …
A brief overview of unsupervised neural speech representation learning
Unsupervised representation learning for speech processing has matured greatly in the last
few years. Work in computer vision and natural language processing has paved the way, but …
few years. Work in computer vision and natural language processing has paved the way, but …