Large language models meet nlp: A survey
While large language models (LLMs) like ChatGPT have shown impressive capabilities in
Natural Language Processing (NLP) tasks, a systematic investigation of their potential in this …
Natural Language Processing (NLP) tasks, a systematic investigation of their potential in this …
RST-LoRA: A Discourse-Aware Low-Rank Adaptation for Long Document Abstractive Summarization
D Pu, V Demberg - arxiv preprint arxiv:2405.00657, 2024 - arxiv.org
For long document summarization, discourse structure is important to discern the key
content of the text and the differences in importance level between sentences. Unfortunately …
content of the text and the differences in importance level between sentences. Unfortunately …
A safety realignment framework via subspace-oriented model fusion for large language models
The current safeguard mechanisms for large language models (LLMs) are indeed
susceptible to jailbreak attacks, making them inherently fragile. Even the process of fine …
susceptible to jailbreak attacks, making them inherently fragile. Even the process of fine …
Simul-LLM: A framework for exploring high-quality simultaneous translation with large language models
Large language models (LLMs) with billions of parameters and pretrained on massive
amounts of data are now capable of near or better than state-of-the-art performance in a …
amounts of data are now capable of near or better than state-of-the-art performance in a …
PEMA: An Offsite-Tunable Plug-in External Memory Adaptation for Language Models
Pre-trained language models (PLMs) show impressive performance in various downstream
NLP tasks. However, pre-training large language models demands substantial memory and …
NLP tasks. However, pre-training large language models demands substantial memory and …
Unlocking Parameter-Efficient Fine-Tuning for Low-Resource Language Translation
Parameter-efficient fine-tuning (PEFT) methods are increasingly vital in adapting large-scale
pre-trained language models for diverse tasks, offering a balance between adaptability and …
pre-trained language models for diverse tasks, offering a balance between adaptability and …
[PDF][PDF] Parameter-Efficient Adapter Based on Pre-trained Models for Speech Translation
N Chen, Y Wang, F Bao - Proc. Interspeech 2024, 2024 - isca-archive.org
Multi-task learning (MTL) approach leverages pre-trained models in speech and machine
translation and has significantly advanced speech-to-text translation tasks. However, it …
translation and has significantly advanced speech-to-text translation tasks. However, it …
[CYTOWANIE][C] A Catalan-German machine translation system based on the M2M-100 multilingual model
P Garriga Riba - 2022