Parameter-efficient fine-tuning for large models: A comprehensive survey

Z Han, C Gao, J Liu, J Zhang, SQ Zhang - arxiv preprint arxiv:2403.14608, 2024 - arxiv.org
Large models represent a groundbreaking advancement in multiple application fields,
enabling remarkable achievements across various tasks. However, their unprecedented …

Domain specialization as the key to make large language models disruptive: A comprehensive survey

C Ling, X Zhao, J Lu, C Deng, C Zheng, J Wang… - arxiv preprint arxiv …, 2023 - arxiv.org
Large language models (LLMs) have significantly advanced the field of natural language
processing (NLP), providing a highly useful, task-agnostic foundation for a wide range of …

Rrhf: Rank responses to align language models with human feedback without tears

Z Yuan, H Yuan, C Tan, W Wang, S Huang… - arxiv preprint arxiv …, 2023 - arxiv.org
Reinforcement Learning from Human Feedback (RLHF) facilitates the alignment of large
language models with human preferences, significantly enhancing the quality of interactions …

RRHF: Rank responses to align language models with human feedback

H Yuan, Z Yuan, C Tan, W Wang… - Advances in Neural …, 2024 - proceedings.neurips.cc
Abstract Reinforcement Learning from Human Feedback (RLHF) facilitates the alignment of
large language models with human preferences, significantly enhancing the quality of …

On the effectiveness of parameter-efficient fine-tuning

Z Fu, H Yang, AMC So, W Lam, L Bing… - Proceedings of the AAAI …, 2023 - ojs.aaai.org
Fine-tuning pre-trained models has been ubiquitously proven to be effective in a wide range
of NLP tasks. However, fine-tuning the whole model is parameter inefficient as it always …

[HTML][HTML] Chatdoctor: A medical chat model fine-tuned on a large language model meta-ai (llama) using medical domain knowledge

Y Li, Z Li, K Zhang, R Dan, S Jiang, Y Zhang - Cureus, 2023 - ncbi.nlm.nih.gov
Objective The primary aim of this research was to address the limitations observed in the
medical knowledge of prevalent large language models (LLMs) such as ChatGPT, by …

Difffit: Unlocking transferability of large diffusion models via simple parameter-efficient fine-tuning

E **e, L Yao, H Shi, Z Liu, D Zhou… - Proceedings of the …, 2023 - openaccess.thecvf.com
Diffusion models have proven to be highly effective in generating high-quality images.
However, adapting large pre-trained diffusion models to new domains remains an open …

Parameter-efficient fine-tuning methods for pretrained language models: A critical review and assessment

L Xu, H **e, SZJ Qin, X Tao, FL Wang - arxiv preprint arxiv:2312.12148, 2023 - arxiv.org
With the continuous growth in the number of parameters of transformer-based pretrained
language models (PLMs), particularly the emergence of large language models (LLMs) with …

How abilities in large language models are affected by supervised fine-tuning data composition

G Dong, H Yuan, K Lu, C Li, M Xue, D Liu… - arxiv preprint arxiv …, 2023 - arxiv.org
Large language models (LLMs) with enormous pre-training tokens and parameter amounts
emerge abilities, including math reasoning, code generation, and instruction following …