Derail Yourself: Multi-turn LLM Jailbreak Attack through Self-discovered Clues
This study exposes the safety vulnerabilities of Large Language Models (LLMs) in multi-turn
interactions, where malicious users can obscure harmful intents across several queries. We …
interactions, where malicious users can obscure harmful intents across several queries. We …
[PDF][PDF] Investigating asr error correction with large language model and multilingual 1-best hypotheses
This paper investigates using pre-trained large language models (LLMs) to improve
multilingual automatic speech recognition (ASR) outputs. Current popular methods involve …
multilingual automatic speech recognition (ASR) outputs. Current popular methods involve …
Preference tuning with human feedback on language, speech, and vision tasks: A survey
Preference tuning is a crucial process for aligning deep generative models with human
preferences. This survey offers a thorough overview of recent advancements in preference …
preferences. This survey offers a thorough overview of recent advancements in preference …