Text data augmentation for deep learning

C Shorten, TM Khoshgoftaar, B Furht - Journal of big Data, 2021 - Springer
Abstract Natural Language Processing (NLP) is one of the most captivating applications of
Deep Learning. In this survey, we consider how the Data Augmentation training strategy can …

A metaverse: Taxonomy, components, applications, and open challenges

SM Park, YG Kim - IEEE access, 2022 - ieeexplore.ieee.org
Unlike previous studies on the Metaverse based on Second Life, the current Metaverse is
based on the social value of Generation Z that online and offline selves are not different …

The learnability of in-context learning

N Wies, Y Levine, A Shashua - Advances in Neural …, 2024 - proceedings.neurips.cc
In-context learning is a surprising and important phenomenon that emerged when modern
language models were scaled to billions of learned parameters. Without modifying a large …

A taxonomy and review of generalization research in NLP

D Hupkes, M Giulianelli, V Dankers, M Artetxe… - Nature Machine …, 2023 - nature.com
The ability to generalize well is one of the primary desiderata for models of natural language
processing (NLP), but what 'good generalization'entails and how it should be evaluated is …

State-of-the-art generalisation research in NLP: a taxonomy and review

D Hupkes, M Giulianelli, V Dankers, M Artetxe… - arxiv preprint arxiv …, 2022 - arxiv.org
The ability to generalise well is one of the primary desiderata of natural language
processing (NLP). Yet, what'good generalisation'entails and how it should be evaluated is …

[HTML][HTML] Filtered corpus training (fict) shows that language models can generalize from indirect evidence

A Patil, J Jumelet, YY Chiu, A Lapastora… - Transactions of the …, 2024 - direct.mit.edu
This paper introduces Fi ltered C orpus T raining, a method that trains language models
(LMs) on corpora with certain linguistic constructions filtered out from the training data, and …

Language acquisition: do children and language models follow similar learning stages?

L Evanson, Y Lakretz, JR King - arxiv preprint arxiv:2306.03586, 2023 - arxiv.org
During language acquisition, children follow a typical sequence of learning stages, whereby
they first learn to categorize phonemes before they develop their lexicon and eventually …

Quiet-star: Language models can teach themselves to think before speaking

E Zelikman, G Harik, Y Shao, V Jayasiri… - arxiv preprint arxiv …, 2024 - arxiv.org
When writing and talking, people sometimes pause to think. Although reasoning-focused
works have often framed reasoning as a method of answering questions or completing …

Language models use monotonicity to assess NPI licensing

J Jumelet, M Denić, J Szymanik, D Hupkes… - arxiv preprint arxiv …, 2021 - arxiv.org
We investigate the semantic knowledge of language models (LMs), focusing on (1) whether
these LMs create categories of linguistic environments based on their semantic monotonicity …

Interpretability of Language Models via Task Spaces

L Weber, J Jumelet, E Bruni, D Hupkes - arxiv preprint arxiv:2406.06441, 2024 - arxiv.org
The usual way to interpret language models (LMs) is to test their performance on different
benchmarks and subsequently infer their internal processes. In this paper, we present an …