A comprehensive survey of few-shot learning: Evolution, applications, challenges, and opportunities

Y Song, T Wang, P Cai, SK Mondal… - ACM Computing Surveys, 2023 - dl.acm.org
Few-shot learning (FSL) has emerged as an effective learning method and shows great
potential. Despite the recent creative works in tackling FSL tasks, learning valid information …

Ext5: Towards extreme multi-task scaling for transfer learning

V Aribandi, Y Tay, T Schuster, J Rao, HS Zheng… - arxiv preprint arxiv …, 2021 - arxiv.org
Despite the recent success of multi-task learning and transfer learning for natural language
processing (NLP), few works have systematically studied the effect of scaling up the number …

Nonparametric masked language modeling

S Min, W Shi, M Lewis, X Chen, W Yih… - arxiv preprint arxiv …, 2022 - arxiv.org
Existing language models (LMs) predict tokens with a softmax over a finite vocabulary,
which can make it difficult to predict rare tokens or phrases. We introduce NPM, the first …

Fusing finetuned models for better pretraining

L Choshen, E Venezian, N Slonim, Y Katz - arxiv preprint arxiv …, 2022 - arxiv.org
Pretrained models are the standard starting point for training. This approach consistently
outperforms the use of a random initialization. However, pretraining is a costly endeavour …

OmniTab: Pretraining with natural and synthetic data for few-shot table-based question answering

Z Jiang, Y Mao, P He, G Neubig, W Chen - arxiv preprint arxiv …, 2022 - arxiv.org
The information in tables can be an important complement to text, making table-based
question answering (QA) systems of great value. The intrinsic complexity of handling tables …

Learning to retrieve passages without supervision

O Ram, G Shachaf, O Levy, J Berant… - arxiv preprint arxiv …, 2021 - arxiv.org
Dense retrievers for open-domain question answering (ODQA) have been shown to achieve
impressive performance by training on large datasets of question-passage pairs. In this work …

WECHSEL: Effective initialization of subword embeddings for cross-lingual transfer of monolingual language models

B Minixhofer, F Paischer, N Rekabsaz - arxiv preprint arxiv:2112.06598, 2021 - arxiv.org
Large pretrained language models (LMs) have become the central building block of many
NLP applications. Training these models requires ever more computational resources and …

SkillSpan: Hard and soft skill extraction from English job postings

M Zhang, KN Jensen, SD Sonniks, B Plank - arxiv preprint arxiv …, 2022 - arxiv.org
Skill Extraction (SE) is an important and widely-studied task useful to gain insights into labor
market dynamics. However, there is a lacuna of datasets and annotation guidelines; …

Turning tables: Generating examples from semi-structured tables for endowing language models with reasoning skills

O Yoran, A Talmor, J Berant - arxiv preprint arxiv:2107.07261, 2021 - arxiv.org
Models pre-trained with a language modeling objective possess ample world knowledge
and language skills, but are known to struggle in tasks that require reasoning. In this work …

ProQA: Structural prompt-based pre-training for unified question answering

W Zhong, Y Gao, N Ding, Y Qin, Z Liu, M Zhou… - arxiv preprint arxiv …, 2022 - arxiv.org
Question Answering (QA) is a longstanding challenge in natural language processing.
Existing QA works mostly focus on specific question types, knowledge domains, or …