A comprehensive overview of large language models

H Naveed, AU Khan, S Qiu, M Saqib, S Anwar… - arxiv preprint arxiv …, 2023 - arxiv.org
Large Language Models (LLMs) have recently demonstrated remarkable capabilities in
natural language processing tasks and beyond. This success of LLMs has led to a large …

Neural prompt search

Y Zhang, K Zhou, Z Liu - IEEE Transactions on Pattern Analysis …, 2024 - ieeexplore.ieee.org
The size of vision models has grown exponentially over the last few years, especially after
the emergence of Vision Transformer. This has motivated the development of parameter …

On the effectiveness of parameter-efficient fine-tuning

Z Fu, H Yang, AMC So, W Lam, L Bing… - Proceedings of the AAAI …, 2023 - ojs.aaai.org
Fine-tuning pre-trained models has been ubiquitously proven to be effective in a wide range
of NLP tasks. However, fine-tuning the whole model is parameter inefficient as it always …

Deep model fusion: A survey

W Li, Y Peng, M Zhang, L Ding, H Hu… - arxiv preprint arxiv …, 2023 - arxiv.org
Deep model fusion/merging is an emerging technique that merges the parameters or
predictions of multiple deep learning models into a single one. It combines the abilities of …

A survey on stability of learning with limited labelled data and its sensitivity to the effects of randomness

B Pecher, I Srba, M Bielikova - ACM Computing Surveys, 2024 - dl.acm.org
Learning with limited labelled data, such as prompting, in-context learning, fine-tuning, meta-
learning, or few-shot learning, aims to effectively train a model using only a small amount of …

Exploring adapter-based transfer learning for recommender systems: Empirical studies and practical insights

J Fu, F Yuan, Y Song, Z Yuan, M Cheng… - Proceedings of the 17th …, 2024 - dl.acm.org
Adapters, a plug-in neural network module with some tunable parameters, have emerged as
a parameter-efficient transfer learning technique for adapting pre-trained models to …

Federated full-parameter tuning of billion-sized language models with communication cost under 18 kilobytes

Z Qin, D Chen, B Qian, B Ding, Y Li, S Deng - arxiv preprint arxiv …, 2023 - arxiv.org
Pre-trained large language models (LLMs) need fine-tuning to improve their responsiveness
to natural language instructions. Federated learning offers a way to fine-tune LLMs using the …

Exploring the capabilities of llms for code change related tasks

L Fan, J Liu, Z Liu, D Lo, X **a, S Li - ACM Transactions on Software …, 2024 - dl.acm.org
Developers deal with code-change-related tasks daily, eg, reviewing code. Pre-trained code
and code-change-oriented models have been adapted to help developers with such tasks …

Astraios: Parameter-efficient instruction tuning code large language models

TY Zhuo, A Zebaze, N Suppattarachai… - arxiv preprint arxiv …, 2024 - arxiv.org
The high cost of full-parameter fine-tuning (FFT) of Large Language Models (LLMs) has led
to a series of parameter-efficient fine-tuning (PEFT) methods. However, it remains unclear …

AutoPEFT: Automatic Configuration Search for Parameter-Efficient Fine-Tuning

H Zhou, X Wan, I Vulić, A Korhonen - Transactions of the Association …, 2024 - direct.mit.edu
Large pretrained language models are widely used in downstream NLP tasks via task-
specific fine-tuning, but such procedures can be costly. Recently, Parameter-Efficient Fine …