Deep learning, graph-based text representation and classification: a survey, perspectives and challenges

P Pham, LTT Nguyen, W Pedrycz, B Vo - Artificial Intelligence Review, 2023 - Springer
Recently, with the rapid developments of the Internet and social networks, there have been
tremendous increase in the amount of complex-structured text resources. These information …

The flan collection: Designing data and methods for effective instruction tuning

S Longpre, L Hou, T Vu, A Webson… - International …, 2023 - proceedings.mlr.press
We study the design decision of publicly available instruction tuning methods, by
reproducing and breaking down the development of Flan 2022 (Chung et al., 2022) …

Spot: Better frozen model adaptation through soft prompt transfer

T Vu, B Lester, N Constant, R Al-Rfou, D Cer - arxiv preprint arxiv …, 2021 - arxiv.org
There has been growing interest in parameter-efficient methods to apply pre-trained
language models to downstream tasks. Building on the Prompt Tuning approach of Lester et …

Merging models with fisher-weighted averaging

MS Matena, CA Raffel - Advances in Neural Information …, 2022 - proceedings.neurips.cc
Averaging the parameters of models that have the same architecture and initialization can
provide a means of combining their respective capabilities. In this paper, we take the …

Ext5: Towards extreme multi-task scaling for transfer learning

V Aribandi, Y Tay, T Schuster, J Rao, HS Zheng… - arxiv preprint arxiv …, 2021 - arxiv.org
Despite the recent success of multi-task learning and transfer learning for natural language
processing (NLP), few works have systematically studied the effect of scaling up the number …

Multitask prompted training enables zero-shot task generalization

V Sanh, A Webson, C Raffel, SH Bach… - arxiv preprint arxiv …, 2021 - arxiv.org
Large language models have recently been shown to attain reasonable zero-shot
generalization on a diverse set of tasks (Brown et al., 2020). It has been hypothesized that …

On transferability of prompt tuning for natural language processing

Y Su, X Wang, Y Qin, CM Chan, Y Lin, H Wang… - arxiv preprint arxiv …, 2021 - arxiv.org
Prompt tuning (PT) is a promising parameter-efficient method to utilize extremely large pre-
trained language models (PLMs), which can achieve comparable performance to full …

Opt-iml: Scaling language model instruction meta learning through the lens of generalization

S Iyer, XV Lin, R Pasunuru, T Mihaylov, D Simig… - arxiv preprint arxiv …, 2022 - arxiv.org
Recent work has shown that fine-tuning large pre-trained language models on a collection
of tasks described via instructions, aka instruction-tuning, improves their zero and few-shot …

Multitask prompt tuning enables parameter-efficient transfer learning

Z Wang, R Panda, L Karlinsky, R Feris, H Sun… - arxiv preprint arxiv …, 2023 - arxiv.org
Prompt tuning, in which a base pretrained model is adapted to each task via conditioning on
learned prompt vectors, has emerged as a promising approach for efficiently adapting large …

Crossfit: A few-shot learning challenge for cross-task generalization in nlp

Q Ye, BY Lin, X Ren - arxiv preprint arxiv:2104.08835, 2021 - arxiv.org
Humans can learn a new language task efficiently with only few examples, by leveraging
their knowledge obtained when learning prior tasks. In this paper, we explore whether and …