Parameter-efficient fine-tuning for large models: A comprehensive survey

Z Han, C Gao, J Liu, J Zhang, SQ Zhang - arxiv preprint arxiv:2403.14608, 2024 - arxiv.org
Large models represent a groundbreaking advancement in multiple application fields,
enabling remarkable achievements across various tasks. However, their unprecedented …

Llm-brain: Ai-driven fast generation of robot behaviour tree based on large language model

A Lykov, D Tsetserukou - 2024 2nd International Conference …, 2024 - ieeexplore.ieee.org
This paper introduces a pioneering methodology in autonomous robot control, denoted as
LLM-BRAIn, enabling the generation of adaptive behaviors in robots in response to operator …

Federated full-parameter tuning of billion-sized language models with communication cost under 18 kilobytes

Z Qin, D Chen, B Qian, B Ding, Y Li, S Deng - arxiv preprint arxiv …, 2023 - arxiv.org
Pre-trained large language models (LLMs) require fine-tuning to improve their
responsiveness to natural language instructions. Federated learning (FL) offers a way to …

Empirical guidelines for deploying llms onto resource-constrained edge devices

R Qin, D Liu, C Xu, Z Yan, Z Tan, Z Jia… - arxiv preprint arxiv …, 2024 - arxiv.org
The scaling laws have become the de facto guidelines for designing large language models
(LLMs), but they were studied under the assumption of unlimited computing resources for …

Low-rank adaptation of large language model rescoring for parameter-efficient speech recognition

Y Yu, CHH Yang, J Kolehmainen… - 2023 IEEE Automatic …, 2023 - ieeexplore.ieee.org
We propose a neural language modeling system based on low-rank adaptation (LoRA) for
speech recognition output rescoring. Although pretrained language models (LMs) like BERT …

Llm-mars: Large language model for behavior tree generation and nlp-enhanced dialogue in multi-agent robot systems

A Lykov, M Dronova, N Naglov, M Litvinov… - arxiv preprint arxiv …, 2023 - arxiv.org
This paper introduces LLM-MARS, first technology that utilizes a Large Language Model
based Artificial Intelligence for Multi-Agent Robot Systems. LLM-MARS enables dynamic …

Balancing cost and effectiveness of synthetic data generation strategies for llms

YC Chan, G Pu, A Shanker, P Suresh, P Jenks… - arxiv preprint arxiv …, 2024 - arxiv.org
As large language models (LLMs) are applied to more use cases, creating high quality, task-
specific datasets for fine-tuning becomes a bottleneck for model improvement. Using high …

Tuning a SAM-Based Model with Multi-Cognitive Visual Adapter to Remote Sensing Instance Segmentation

L Zheng, X Pu, F Xu - IEEE Journal of Selected Topics in …, 2024 - ieeexplore.ieee.org
The segment anything model (SAM), a foundational model designed for promptable
segmentation tasks, demonstrates exceptional generalization capabilities, making it highly …

Semantic are Beacons: A Semantic Perspective for Unveiling Parameter-Efficient Fine-Tuning in Knowledge Learning

R Wang, P Li - arxiv preprint arxiv:2405.18292, 2024 - arxiv.org
Parameter-Efficient Fine-Tuning (PEFT) methods enable efficient adaptation of Large
Language Models (LLMs) to various downstream applications. However, the effectiveness of …

Tri-Plane Mamba: Efficiently Adapting Segment Anything Model for 3D Medical Images

H Wang, Y Lin, X Ding, X Li - … on Medical Image Computing and Computer …, 2024 - Springer
General networks for 3D medical image segmentation have recently undergone extensive
exploration. Behind the exceptional performance of these networks lies a significant demand …