A comprehensive survey of few-shot learning: Evolution, applications, challenges, and opportunities
Few-shot learning (FSL) has emerged as an effective learning method and shows great
potential. Despite the recent creative works in tackling FSL tasks, learning valid information …
potential. Despite the recent creative works in tackling FSL tasks, learning valid information …
Ext5: Towards extreme multi-task scaling for transfer learning
Despite the recent success of multi-task learning and transfer learning for natural language
processing (NLP), few works have systematically studied the effect of scaling up the number …
processing (NLP), few works have systematically studied the effect of scaling up the number …
Nonparametric masked language modeling
Existing language models (LMs) predict tokens with a softmax over a finite vocabulary,
which can make it difficult to predict rare tokens or phrases. We introduce NPM, the first …
which can make it difficult to predict rare tokens or phrases. We introduce NPM, the first …
Fusing finetuned models for better pretraining
Pretrained models are the standard starting point for training. This approach consistently
outperforms the use of a random initialization. However, pretraining is a costly endeavour …
outperforms the use of a random initialization. However, pretraining is a costly endeavour …
OmniTab: Pretraining with natural and synthetic data for few-shot table-based question answering
The information in tables can be an important complement to text, making table-based
question answering (QA) systems of great value. The intrinsic complexity of handling tables …
question answering (QA) systems of great value. The intrinsic complexity of handling tables …
Learning to retrieve passages without supervision
Dense retrievers for open-domain question answering (ODQA) have been shown to achieve
impressive performance by training on large datasets of question-passage pairs. In this work …
impressive performance by training on large datasets of question-passage pairs. In this work …
WECHSEL: Effective initialization of subword embeddings for cross-lingual transfer of monolingual language models
Large pretrained language models (LMs) have become the central building block of many
NLP applications. Training these models requires ever more computational resources and …
NLP applications. Training these models requires ever more computational resources and …
SkillSpan: Hard and soft skill extraction from English job postings
Skill Extraction (SE) is an important and widely-studied task useful to gain insights into labor
market dynamics. However, there is a lacuna of datasets and annotation guidelines; …
market dynamics. However, there is a lacuna of datasets and annotation guidelines; …
Turning tables: Generating examples from semi-structured tables for endowing language models with reasoning skills
Models pre-trained with a language modeling objective possess ample world knowledge
and language skills, but are known to struggle in tasks that require reasoning. In this work …
and language skills, but are known to struggle in tasks that require reasoning. In this work …
ProQA: Structural prompt-based pre-training for unified question answering
Question Answering (QA) is a longstanding challenge in natural language processing.
Existing QA works mostly focus on specific question types, knowledge domains, or …
Existing QA works mostly focus on specific question types, knowledge domains, or …