Recent advances in natural language inference: A survey of benchmarks, resources, and approaches

S Storks, Q Gao, JY Chai - arxiv preprint arxiv:1904.01172, 2019 - arxiv.org
In the NLP community, recent years have seen a surge of research activities that address
machines' ability to perform deep language understanding which goes beyond what is …

Mathematical language models: A survey

W Liu, H Hu, J Zhou, Y Ding, J Li, J Zeng, M He… - arxiv preprint arxiv …, 2023 - arxiv.org
In recent years, there has been remarkable progress in leveraging Language Models (LMs),
encompassing Pre-trained Language Models (PLMs) and Large-scale Language Models …

Piqa: Reasoning about physical commonsense in natural language

Y Bisk, R Zellers, J Gao, Y Choi - … of the AAAI conference on artificial …, 2020 - ojs.aaai.org
To apply eyeshadow without a brush, should I use a cotton swab or a toothpick? Questions
requiring this kind of physical commonsense pose a challenge to today's natural language …

Swag: A large-scale adversarial dataset for grounded commonsense inference

R Zellers, Y Bisk, R Schwartz, Y Choi - arxiv preprint arxiv:1808.05326, 2018 - arxiv.org
Given a partial description like" she opened the hood of the car," humans can reason about
the situation and anticipate what might come next (" then, she examined the engine"). In this …

Do NLP models know numbers? probing numeracy in embeddings

E Wallace, Y Wang, S Li, S Singh… - arxiv preprint arxiv …, 2019 - arxiv.org
The ability to understand and work with numbers (numeracy) is critical for many complex
reasoning tasks. Currently, most NLP models treat numbers in text in the same way as other …

Generative data augmentation for commonsense reasoning

Y Yang, C Malaviya, J Fernandez… - arxiv preprint arxiv …, 2020 - arxiv.org
Recent advances in commonsense reasoning depend on large-scale human-annotated
training data to achieve peak performance. However, manual curation of training examples …

DREAM: A challenge data set and models for dialogue-based reading comprehension

K Sun, D Yu, J Chen, D Yu, Y Choi… - Transactions of the …, 2019 - direct.mit.edu
We present DREAM, the first dialogue-based multiple-choice reading comprehension data
set. Collected from English as a Foreign Language examinations designed by human …

oLMpics-on what language model pre-training captures

A Talmor, Y Elazar, Y Goldberg… - Transactions of the …, 2020 - direct.mit.edu
Recent success of pre-trained language models (LMs) has spurred widespread interest in
the language capabilities that they possess. However, efforts to understand whether LM …

Representing numbers in NLP: a survey and a vision

A Thawani, J Pujara, PA Szekely, F Ilievski - arxiv preprint arxiv …, 2021 - arxiv.org
NLP systems rarely give special consideration to numbers found in text. This starkly
contrasts with the consensus in neuroscience that, in the brain, numbers are represented …

Birds have four legs?! numersense: Probing numerical commonsense knowledge of pre-trained language models

BY Lin, S Lee, R Khanna, X Ren - arxiv preprint arxiv:2005.00683, 2020 - arxiv.org
Recent works show that pre-trained language models (PTLMs), such as BERT, possess
certain commonsense and factual knowledge. They suggest that it is promising to use …