Alignment of diffusion models: Fundamentals, challenges, and future

B Liu, S Shao, B Li, L Bai, Z Xu, H **ong, J Kwok… - arxiv preprint arxiv …, 2024 - arxiv.org
Diffusion models have emerged as the leading paradigm in generative modeling, excelling
in various applications. Despite their success, these models often misalign with human …

Navigating the overkill in large language models

C Shi, X Wang, Q Ge, S Gao, X Yang, T Gui… - arxiv preprint arxiv …, 2024 - arxiv.org
Large language models are meticulously aligned to be both helpful and harmless. However,
recent research points to a potential overkill which means models may refuse to answer …

How Reliable Is Human Feedback For Aligning Large Language Models?

MH Yeh, L Tao, J Wang, X Du, Y Li - arxiv preprint arxiv:2410.01957, 2024 - arxiv.org
Most alignment research today focuses on designing new learning algorithms using
datasets like Anthropic-HH, assuming human feedback data is inherently reliable. However …

MetaAlign: Align Large Language Models with Diverse Preferences during Inference Time

M Zhang, P Wang, C Tan, M Huang, D Zhang… - arxiv preprint arxiv …, 2024 - arxiv.org
Large Language Models (LLMs) acquire extensive knowledge and remarkable abilities from
extensive text corpora, making them powerful tools for various applications. To make LLMs …

On Uncertainty In Natural Language Processing

D Ulmer - arxiv preprint arxiv:2410.03446, 2024 - arxiv.org
The last decade in deep learning has brought on increasingly capable systems that are
deployed on a wide variety of applications. In natural language processing, the field has …

[PDF][PDF] The Alignment Formula: Large Language Models and Humans' Decisions in a False-Belief Task

M Zgreabăn - 2024 - studenttheses.uu.nl
The ability of humans to estimate other's beliefs, mental states, and attitudes is referred to as
Theory of Mind (ToM)[3–8]. ToM is widely researched due to its alleged implications in …