Large language models meet nlp: A survey

L Qin, Q Chen, X Feng, Y Wu, Y Zhang, Y Li… - arxiv preprint arxiv …, 2024 - arxiv.org
While large language models (LLMs) like ChatGPT have shown impressive capabilities in
Natural Language Processing (NLP) tasks, a systematic investigation of their potential in this …

RST-LoRA: A Discourse-Aware Low-Rank Adaptation for Long Document Abstractive Summarization

D Pu, V Demberg - arxiv preprint arxiv:2405.00657, 2024 - arxiv.org
For long document summarization, discourse structure is important to discern the key
content of the text and the differences in importance level between sentences. Unfortunately …

A safety realignment framework via subspace-oriented model fusion for large language models

X Yi, S Zheng, L Wang, X Wang, L He - arxiv preprint arxiv:2405.09055, 2024 - arxiv.org
The current safeguard mechanisms for large language models (LLMs) are indeed
susceptible to jailbreak attacks, making them inherently fragile. Even the process of fine …

Simul-LLM: A framework for exploring high-quality simultaneous translation with large language models

V Agostinelli, M Wild, M Raffel, KAA Fuad… - arxiv preprint arxiv …, 2023 - arxiv.org
Large language models (LLMs) with billions of parameters and pretrained on massive
amounts of data are now capable of near or better than state-of-the-art performance in a …

PEMA: An Offsite-Tunable Plug-in External Memory Adaptation for Language Models

HJ Kim, YJ Kim, JY Bak - Proceedings of the 2024 Conference of …, 2024 - aclanthology.org
Pre-trained language models (PLMs) show impressive performance in various downstream
NLP tasks. However, pre-training large language models demands substantial memory and …

Unlocking Parameter-Efficient Fine-Tuning for Low-Resource Language Translation

T Su, X Peng, S Thillainathan, D Guzmán… - arxiv preprint arxiv …, 2024 - arxiv.org
Parameter-efficient fine-tuning (PEFT) methods are increasingly vital in adapting large-scale
pre-trained language models for diverse tasks, offering a balance between adaptability and …

[PDF][PDF] Parameter-Efficient Adapter Based on Pre-trained Models for Speech Translation

N Chen, Y Wang, F Bao - Proc. Interspeech 2024, 2024 - isca-archive.org
Multi-task learning (MTL) approach leverages pre-trained models in speech and machine
translation and has significantly advanced speech-to-text translation tasks. However, it …

[CYTOWANIE][C] A Catalan-German machine translation system based on the M2M-100 multilingual model

P Garriga Riba - 2022