Few-shot parameter-efficient fine-tuning is better and cheaper than in-context learning
Few-shot in-context learning (ICL) enables pre-trained language models to perform a
previously-unseen task without any gradient-based training by feeding a small number of …
previously-unseen task without any gradient-based training by feeding a small number of …
ChatGPT vs human-authored text: Insights into controllable text summarization and sentence style transfer
D Pu, V Demberg - arxiv preprint arxiv:2306.07799, 2023 - arxiv.org
Large-scale language models, like ChatGPT, have garnered significant media attention and
stunned the public with their remarkable capacity for generating coherent text from short …
stunned the public with their remarkable capacity for generating coherent text from short …
Infoprompt: Information-theoretic soft prompt tuning for natural language understanding
Soft prompt tuning achieves superior performances across a wide range of few-shot tasks.
However, the performances of prompt tuning can be highly sensitive to the initialization of …
However, the performances of prompt tuning can be highly sensitive to the initialization of …
UniSumm and SummZoo: Unified model and diverse benchmark for few-shot summarization
The high annotation costs and diverse demands of various summarization tasks motivate the
development of few-shot summarization. However, despite the emergence of many …
development of few-shot summarization. However, despite the emergence of many …
Query refinement prompts for closed-book long-form question answering
Large language models (LLMs) have been shown to perform well in answering questions
and in producing long-form texts, both in few-shot closed-book settings. While the former can …
and in producing long-form texts, both in few-shot closed-book settings. While the former can …
[HTML][HTML] Align-then-abstract representation learning for low-resource summarization
Generative transformer-based models have achieved state-of-the-art performance in text
summarization. Nevertheless, they still struggle in real-world scenarios with long documents …
summarization. Nevertheless, they still struggle in real-world scenarios with long documents …
Towards summary candidates fusion
Sequence-to-sequence deep neural models fine-tuned for abstractive summarization can
achieve great performance on datasets with enough human annotations. Yet, it has been …
achieve great performance on datasets with enough human annotations. Yet, it has been …
Learning instructions with unlabeled data for zero-shot cross-task generalization
Training language models to learn from human instructions for zero-shot cross-task
generalization has attracted much attention in NLP communities. Recently, instruction tuning …
generalization has attracted much attention in NLP communities. Recently, instruction tuning …
Recent trends in unsupervised summarization
Unsupervised summarization is a powerful technique that enables training summarizing
models without requiring labeled datasets. This survey covers different recent techniques …
models without requiring labeled datasets. This survey covers different recent techniques …
SPC: Soft Prompt Construction for Cross Domain Generalization
Recent advances in prompt tuning have proven effective as a new language modeling
paradigm for various natural language understanding tasks. However, it is challenging to …
paradigm for various natural language understanding tasks. However, it is challenging to …