Parameter-efficient fine-tuning for large models: A comprehensive survey
Large models represent a groundbreaking advancement in multiple application fields,
enabling remarkable achievements across various tasks. However, their unprecedented …
enabling remarkable achievements across various tasks. However, their unprecedented …
Domain specialization as the key to make large language models disruptive: A comprehensive survey
Large language models (LLMs) have significantly advanced the field of natural language
processing (NLP), providing a highly useful, task-agnostic foundation for a wide range of …
processing (NLP), providing a highly useful, task-agnostic foundation for a wide range of …
Rrhf: Rank responses to align language models with human feedback without tears
Reinforcement Learning from Human Feedback (RLHF) facilitates the alignment of large
language models with human preferences, significantly enhancing the quality of interactions …
language models with human preferences, significantly enhancing the quality of interactions …
RRHF: Rank responses to align language models with human feedback
Abstract Reinforcement Learning from Human Feedback (RLHF) facilitates the alignment of
large language models with human preferences, significantly enhancing the quality of …
large language models with human preferences, significantly enhancing the quality of …
On the effectiveness of parameter-efficient fine-tuning
Fine-tuning pre-trained models has been ubiquitously proven to be effective in a wide range
of NLP tasks. However, fine-tuning the whole model is parameter inefficient as it always …
of NLP tasks. However, fine-tuning the whole model is parameter inefficient as it always …
[HTML][HTML] Chatdoctor: A medical chat model fine-tuned on a large language model meta-ai (llama) using medical domain knowledge
Objective The primary aim of this research was to address the limitations observed in the
medical knowledge of prevalent large language models (LLMs) such as ChatGPT, by …
medical knowledge of prevalent large language models (LLMs) such as ChatGPT, by …
Difffit: Unlocking transferability of large diffusion models via simple parameter-efficient fine-tuning
Diffusion models have proven to be highly effective in generating high-quality images.
However, adapting large pre-trained diffusion models to new domains remains an open …
However, adapting large pre-trained diffusion models to new domains remains an open …
Parameter-efficient fine-tuning methods for pretrained language models: A critical review and assessment
With the continuous growth in the number of parameters of transformer-based pretrained
language models (PLMs), particularly the emergence of large language models (LLMs) with …
language models (PLMs), particularly the emergence of large language models (LLMs) with …
How abilities in large language models are affected by supervised fine-tuning data composition
Large language models (LLMs) with enormous pre-training tokens and parameter amounts
emerge abilities, including math reasoning, code generation, and instruction following …
emerge abilities, including math reasoning, code generation, and instruction following …