Deep learning, graph-based text representation and classification: a survey, perspectives and challenges
Recently, with the rapid developments of the Internet and social networks, there have been
tremendous increase in the amount of complex-structured text resources. These information …
tremendous increase in the amount of complex-structured text resources. These information …
The flan collection: Designing data and methods for effective instruction tuning
We study the design decision of publicly available instruction tuning methods, by
reproducing and breaking down the development of Flan 2022 (Chung et al., 2022) …
reproducing and breaking down the development of Flan 2022 (Chung et al., 2022) …
Spot: Better frozen model adaptation through soft prompt transfer
There has been growing interest in parameter-efficient methods to apply pre-trained
language models to downstream tasks. Building on the Prompt Tuning approach of Lester et …
language models to downstream tasks. Building on the Prompt Tuning approach of Lester et …
Merging models with fisher-weighted averaging
MS Matena, CA Raffel - Advances in Neural Information …, 2022 - proceedings.neurips.cc
Averaging the parameters of models that have the same architecture and initialization can
provide a means of combining their respective capabilities. In this paper, we take the …
provide a means of combining their respective capabilities. In this paper, we take the …
Ext5: Towards extreme multi-task scaling for transfer learning
Despite the recent success of multi-task learning and transfer learning for natural language
processing (NLP), few works have systematically studied the effect of scaling up the number …
processing (NLP), few works have systematically studied the effect of scaling up the number …
Multitask prompted training enables zero-shot task generalization
Large language models have recently been shown to attain reasonable zero-shot
generalization on a diverse set of tasks (Brown et al., 2020). It has been hypothesized that …
generalization on a diverse set of tasks (Brown et al., 2020). It has been hypothesized that …
On transferability of prompt tuning for natural language processing
Prompt tuning (PT) is a promising parameter-efficient method to utilize extremely large pre-
trained language models (PLMs), which can achieve comparable performance to full …
trained language models (PLMs), which can achieve comparable performance to full …
Opt-iml: Scaling language model instruction meta learning through the lens of generalization
Recent work has shown that fine-tuning large pre-trained language models on a collection
of tasks described via instructions, aka instruction-tuning, improves their zero and few-shot …
of tasks described via instructions, aka instruction-tuning, improves their zero and few-shot …
Multitask prompt tuning enables parameter-efficient transfer learning
Prompt tuning, in which a base pretrained model is adapted to each task via conditioning on
learned prompt vectors, has emerged as a promising approach for efficiently adapting large …
learned prompt vectors, has emerged as a promising approach for efficiently adapting large …
Crossfit: A few-shot learning challenge for cross-task generalization in nlp
Humans can learn a new language task efficiently with only few examples, by leveraging
their knowledge obtained when learning prior tasks. In this paper, we explore whether and …
their knowledge obtained when learning prior tasks. In this paper, we explore whether and …