Turnitin
降AI改写
早检测系统
早降重系统
Turnitin-UK版
万方检测-期刊版
维普编辑部版
Grammarly检测
Paperpass检测
checkpass检测
PaperYY检测
Tera: Self-supervised learning of transformer encoder representation for speech
We introduce a self-supervised speech pre-training method called TERA, which stands for
Transformer Encoder Representations from Alteration. Recent approaches often learn by …
Transformer Encoder Representations from Alteration. Recent approaches often learn by …
Slue: New benchmark tasks for spoken language understanding evaluation on natural speech
Progress in speech processing has been facilitated by shared datasets and benchmarks.
Historically these have focused on automatic speech recognition (ASR), speaker …
Historically these have focused on automatic speech recognition (ASR), speaker …
Expanding large pre-trained unimodal models with multimodal information injection for image-text multimodal classification
Fine-tuning pre-trained models for downstream tasks is mainstream in deep learning.
However, the pre-trained models are limited to be fine-tuned by data from a specific …
However, the pre-trained models are limited to be fine-tuned by data from a specific …
Applications, risk, challenges, and future prospects of ChatGPT in electronic records management
D Lin, R Zou - Journal of Artificial Intelligence Research, 2024 - sub.ifspress.hk
The widespread application and rapid development of ChatGPT are disrupting traditional
models across industries, bringing revolutionary changes to electronic records …
models across industries, bringing revolutionary changes to electronic records …
End-to-end neural transformer based spoken language understanding
Spoken language understanding (SLU) refers to the process of inferring the semantic
information from audio signals. While the neural transformers consistently deliver the best …
information from audio signals. While the neural transformers consistently deliver the best …
Semi-supervised spoken language understanding via self-supervised speech and language model pretraining
Much recent work on Spoken Language Understanding (SLU) is limited in at least one of
three ways: models were trained on oracle text input and neglected ASR errors, models …
three ways: models were trained on oracle text input and neglected ASR errors, models …
Speech-language pre-training for end-to-end spoken language understanding
End-to-end (E2E) spoken language understanding (SLU) can infer semantics directly from
speech signal without cascading an automatic speech recognizer (ASR) with a natural …
speech signal without cascading an automatic speech recognizer (ASR) with a natural …
St-bert: Cross-modal language model pre-training for end-to-end spoken language understanding
Language model pre-training has shown promising results in various downstream tasks. In
this context, we introduce a cross-modal pre-trained language model, called Speech-Text …
this context, we introduce a cross-modal pre-trained language model, called Speech-Text …
Understanding self-attention of self-supervised audio transformers
Self-supervised Audio Transformers (SAT) enable great success in many downstream
speech applications like ASR, but how they work has not been widely explored yet. In this …
speech applications like ASR, but how they work has not been widely explored yet. In this …
Exploring transfer learning for end-to-end spoken language understanding
Abstract Voice Assistants such as Alexa, Siri, and Google Assistant typically use a two-stage
Spoken Language Understanding pipeline; first, an Automatic Speech Recognition (ASR) …
Spoken Language Understanding pipeline; first, an Automatic Speech Recognition (ASR) …