[HTML][HTML] Stacked noise reduction auto encoder–OCEAN: a novel personalized recommendation model enhanced

B Wang, W Zheng, R Wang, S Lu, L Yin, L Wang, Z Yin… - Systems, 2024 - mdpi.com
With the continuous development of information technology and the rapid increase in new
users of social networking sites, recommendation technology is becoming more and more …

Speech-text pre-training for spoken dialog understanding with explicit cross-modal alignment

T Yu, H Gao, TE Lin, M Yang, Y Wu, W Ma… - Proceedings of the …, 2023 - aclanthology.org
Recently, speech-text pre-training methods have shown remarkable success in many
speech and natural language processing tasks. However, most previous pre-trained models …

Em-network: Oracle guided self-distillation for sequence learning

JW Yoon, S Ahn, H Lee, M Kim… - … on Machine Learning, 2023 - proceedings.mlr.press
We introduce EM-Network, a novel self-distillation approach that effectively leverages target
information for supervised sequence-to-sequence (seq2seq) learning. In contrast to …

Speech-text dialog pre-training for spoken dialog understanding with explicit cross-modal alignment

T Yu, H Gao, TE Lin, M Yang, Y Wu, W Ma… - arxiv preprint arxiv …, 2023 - arxiv.org
Recently, speech-text pre-training methods have shown remarkable success in many
speech and natural language processing tasks. However, most previous pre-trained models …

Teach me with a Whisper: Enhancing Large Language Models for Analyzing Spoken Transcripts using Speech Embeddings

F Hasan, Y Li, J Foulds, S Pan… - arxiv preprint arxiv …, 2023 - arxiv.org
Speech data has rich acoustic and paralinguistic information with important cues for
understanding a speaker's tone, emotion, and intent, yet traditional large language models …

CLASP: Cross-modal Alignment Using Pre-trained Unimodal Models

J Zhou, Z Zeng, H Gong, S Bhat - Findings of the Association for …, 2024 - aclanthology.org
Recent advancements in joint speech-text pre-training have significantly advanced the
processing of natural language. However, a key limitation is their reliance on parallel …

DoubleDistillation: Enhancing LLMs for Informal Text Analysis using Multistage Knowledge Distillation from Speech and Text

F Hasan, Y Li, JR Foulds, S Pan… - Proceedings of the 26th …, 2024 - dl.acm.org
Traditional large language models (LLMs) leverage extensive text corpora but lack access to
acoustic and para-linguistic cues present in speech. There is a growing interest in …

A Systematic Review of Adversarial Machine Learning and Deep Learning Applications

TA Abdalkareem, KA Zidan… - Al-Iraqia Journal for …, 2024 - ijser.aliraqia.edu.iq
The review delves into creating an understandable framework for machine learning in
robotics. It stresses the significance of machine learning in materials science and robotics …

Knowledge Distillation Methods for Sequence-to-Sequence Learning in Speech and Language Processing

윤지원 - 2024 - s-space.snu.ac.kr
Recently, sequence-to-sequence learning has shown remarkable performance in speech
and natural language processing. However, high-performing sequence models commonly …