Turnitin
降AI改写
早检测系统
早降重系统
Turnitin-UK版
万方检测-期刊版
维普编辑部版
Grammarly检测
Paperpass检测
checkpass检测
PaperYY检测
The benefits, risks and bounds of personalizing the alignment of large language models to individuals
Large language models (LLMs) undergo 'alignment'so that they better reflect human values
or preferences, and are safer or more useful. However, alignment is intrinsically difficult …
or preferences, and are safer or more useful. However, alignment is intrinsically difficult …
From matching to generation: A survey on generative information retrieval
Information Retrieval (IR) systems are crucial tools for users to access information, widely
applied in scenarios like search engines, question answering, and recommendation …
applied in scenarios like search engines, question answering, and recommendation …
Prometheus 2: An open source language model specialized in evaluating other language models
Proprietary LMs such as GPT-4 are often employed to assess the quality of responses from
various LMs. However, concerns including transparency, controllability, and affordability …
various LMs. However, concerns including transparency, controllability, and affordability …
The prism alignment project: What participatory, representative and individualised human feedback reveals about the subjective and multicultural alignment of large …
Human feedback plays a central role in the alignment of Large Language Models (LLMs).
However, open questions remain about the methods (how), domains (where), people (who) …
However, open questions remain about the methods (how), domains (where), people (who) …
Arithmetic control of llms for diverse user preferences: Directional preference alignment with multi-objective rewards
Fine-grained control over large language models (LLMs) remains a significant challenge,
hindering their adaptability to diverse user needs. While Reinforcement Learning from …
hindering their adaptability to diverse user needs. While Reinforcement Learning from …
MaxMin-RLHF: Alignment with diverse human preferences
Reinforcement Learning from Human Feedback (RLHF) aligns language models to human
preferences by employing a singular reward model derived from preference data. However …
preferences by employing a singular reward model derived from preference data. However …
From persona to personalization: A survey on role-playing language agents
Recent advancements in large language models (LLMs) have significantly boosted the rise
of Role-Playing Language Agents (RPLAs), ie, specialized AI systems designed to simulate …
of Role-Playing Language Agents (RPLAs), ie, specialized AI systems designed to simulate …
Aligning to thousands of preferences via system message generalization
Although humans inherently have diverse values, current large language model (LLM)
alignment methods often assume that aligning LLMs with the general public's preferences is …
alignment methods often assume that aligning LLMs with the general public's preferences is …
Optimization methods for personalizing large language models through retrieval augmentation
This paper studies retrieval-augmented approaches for personalizing large language
models (LLMs), which potentially have a substantial impact on various applications and …
models (LLMs), which potentially have a substantial impact on various applications and …
Model merging in llms, mllms, and beyond: Methods, theories, applications and opportunities
Model merging is an efficient empowerment technique in the machine learning community
that does not require the collection of raw training data and does not require expensive …
that does not require the collection of raw training data and does not require expensive …