Turnitin
降AI改写
早检测系统
早降重系统
Turnitin-UK版
万方检测-期刊版
维普编辑部版
Grammarly检测
Paperpass检测
checkpass检测
PaperYY检测
PERSONA: A Reproducible Testbed for Pluralistic Alignment
The rapid advancement of language models (LMs) necessitates robust alignment with
diverse user values. However, current preference optimization approaches often fail to …
diverse user values. However, current preference optimization approaches often fail to …
Adaptagent: Adapting multimodal web agents with few-shot learning from human demonstrations
State-of-the-art multimodal web agents, powered by Multimodal Large Language Models
(MLLMs), can autonomously execute many web tasks by processing user instructions and …
(MLLMs), can autonomously execute many web tasks by processing user instructions and …
ChainBuddy: An AI Agent System for Generating LLM Pipelines
J Zhang, I Arawjo - arxiv preprint arxiv:2409.13588, 2024 - arxiv.org
As large language models (LLMs) advance, their potential applications have grown
significantly. However, it remains difficult to evaluate LLM behavior on user-specific tasks …
significantly. However, it remains difficult to evaluate LLM behavior on user-specific tasks …
SePPO: Semi-Policy Preference Optimization for Diffusion Alignment
Reinforcement learning from human feedback (RLHF) methods are emerging as a way to
fine-tune diffusion models (DMs) for visual generation. However, commonly used on-policy …
fine-tune diffusion models (DMs) for visual generation. However, commonly used on-policy …
Open-domain implicit format control for large language model generation
Controlling the format of outputs generated by large language models (LLMs) is a critical
functionality in various applications. Current methods typically employ constrained decoding …
functionality in various applications. Current methods typically employ constrained decoding …
Tuning-Free Personalized Alignment via Trial-Error-Explain In-Context Learning
Language models are aligned to the collective voice of many, resulting in generic outputs
that do not align with specific users' styles. In this work, we present Trial-Error-Explain In …
that do not align with specific users' styles. In this work, we present Trial-Error-Explain In …
Aligning LLMs with Domain Invariant Reward Models
D Wu, S Choudhury - arxiv preprint arxiv:2501.00911, 2025 - arxiv.org
Aligning large language models (LLMs) to human preferences is challenging in domains
where preference data is unavailable. We address the problem of learning reward models …
where preference data is unavailable. We address the problem of learning reward models …
No Preference Left Behind: Group Distributional Preference Optimization
Preferences within a group of people are not uniform but follow a distribution. While existing
alignment methods like Direct Preference Optimization (DPO) attempt to steer models to …
alignment methods like Direct Preference Optimization (DPO) attempt to steer models to …
Improving llm generation with inverse and forward alignment: Reward modeling, prompting, fine-tuning, and inference-time optimization
Large Language Models (LLMs) are often characterized as samplers or generators in the
literature, yet maximizing their capabilities in these roles is a complex challenge. Previous …
literature, yet maximizing their capabilities in these roles is a complex challenge. Previous …
Granting Non-AI Experts Creative Control Over AI Systems
MS Lam - Adjunct Proceedings of the 37th Annual ACM …, 2024 - dl.acm.org
Many harmful behaviors and problematic deployments of AI stem from the fact that AI experts
are not experts in the vast array of settings where AI is applied. Non-AI experts from these …
are not experts in the vast array of settings where AI is applied. Non-AI experts from these …