Few-shot parameter-efficient fine-tuning is better and cheaper than in-context learning

H Liu, D Tam, M Muqeeth, J Mohta… - Advances in …, 2022 - proceedings.neurips.cc
Few-shot in-context learning (ICL) enables pre-trained language models to perform a
previously-unseen task without any gradient-based training by feeding a small number of …

ChatGPT vs human-authored text: Insights into controllable text summarization and sentence style transfer

D Pu, V Demberg - arxiv preprint arxiv:2306.07799, 2023 - arxiv.org
Large-scale language models, like ChatGPT, have garnered significant media attention and
stunned the public with their remarkable capacity for generating coherent text from short …

Infoprompt: Information-theoretic soft prompt tuning for natural language understanding

J Wu, T Yu, R Wang, Z Song, R Zhang… - Advances in …, 2024 - proceedings.neurips.cc
Soft prompt tuning achieves superior performances across a wide range of few-shot tasks.
However, the performances of prompt tuning can be highly sensitive to the initialization of …

UniSumm and SummZoo: Unified model and diverse benchmark for few-shot summarization

Y Chen, Y Liu, R Xu, Z Yang, C Zhu, M Zeng… - arxiv preprint arxiv …, 2022 - arxiv.org
The high annotation costs and diverse demands of various summarization tasks motivate the
development of few-shot summarization. However, despite the emergence of many …

Query refinement prompts for closed-book long-form question answering

RK Amplayo, K Webster, M Collins, D Das… - arxiv preprint arxiv …, 2022 - arxiv.org
Large language models (LLMs) have been shown to perform well in answering questions
and in producing long-form texts, both in few-shot closed-book settings. While the former can …

[HTML][HTML] Align-then-abstract representation learning for low-resource summarization

G Moro, L Ragazzi - Neurocomputing, 2023 - Elsevier
Generative transformer-based models have achieved state-of-the-art performance in text
summarization. Nevertheless, they still struggle in real-world scenarios with long documents …

Towards summary candidates fusion

M Ravaut, S Joty, NF Chen - arxiv preprint arxiv:2210.08779, 2022 - arxiv.org
Sequence-to-sequence deep neural models fine-tuned for abstractive summarization can
achieve great performance on datasets with enough human annotations. Yet, it has been …

Learning instructions with unlabeled data for zero-shot cross-task generalization

Y Gu, P Ke, X Zhu, M Huang - arxiv preprint arxiv:2210.09175, 2022 - arxiv.org
Training language models to learn from human instructions for zero-shot cross-task
generalization has attracted much attention in NLP communities. Recently, instruction tuning …

Recent trends in unsupervised summarization

M Khosravani, A Trabelsi - arxiv preprint arxiv:2305.11231, 2023 - arxiv.org
Unsupervised summarization is a powerful technique that enables training summarizing
models without requiring labeled datasets. This survey covers different recent techniques …

SPC: Soft Prompt Construction for Cross Domain Generalization

W Zhao, A Gupta, T Chung, J Huang - Proceedings of the 8th …, 2023 - aclanthology.org
Recent advances in prompt tuning have proven effective as a new language modeling
paradigm for various natural language understanding tasks. However, it is challenging to …