A survey on machine reading comprehension systems

R Baradaran, R Ghiasi, H Amirkhani - Natural Language Engineering, 2022 - cambridge.org
Machine Reading Comprehension (MRC) is a challenging task and hot topic in Natural
Language Processing. The goal of this field is to develop systems for answering the …

Prompting gpt-3 to be reliable

C Si, Z Gan, Z Yang, S Wang, J Wang… - arxiv preprint arxiv …, 2022 - arxiv.org
Large language models (LLMs) show impressive abilities via few-shot prompting.
Commercialized APIs such as OpenAI GPT-3 further increase their use in real-world …

Revisiting out-of-distribution robustness in nlp: Benchmarks, analysis, and llms evaluations

L Yuan, Y Chen, G Cui, H Gao, F Zou… - Advances in …, 2023 - proceedings.neurips.cc
This paper reexamines the research on out-of-distribution (OOD) robustness in the field of
NLP. We find that the distribution shift settings in previous studies commonly lack adequate …

Ppt: Pre-trained prompt tuning for few-shot learning

Y Gu, X Han, Z Liu, M Huang - arxiv preprint arxiv:2109.04332, 2021 - arxiv.org
Prompts for pre-trained language models (PLMs) have shown remarkable performance by
bridging the gap between pre-training tasks and various downstream tasks. Among these …

Spot: Better frozen model adaptation through soft prompt transfer

T Vu, B Lester, N Constant, R Al-Rfou, D Cer - arxiv preprint arxiv …, 2021 - arxiv.org
There has been growing interest in parameter-efficient methods to apply pre-trained
language models to downstream tasks. Building on the Prompt Tuning approach of Lester et …

Unifiedqa: Crossing format boundaries with a single qa system

D Khashabi, S Min, T Khot, A Sabharwal… - arxiv preprint arxiv …, 2020 - arxiv.org
Question answering (QA) tasks have been posed using a variety of formats, such as
extractive span selection, multiple choice, etc. This has led to format-specialized models …

Intermediate-task transfer learning with pretrained models for natural language understanding: When and why does it work?

Y Pruksachatkun, J Phang, H Liu, PM Htut… - arxiv preprint arxiv …, 2020 - arxiv.org
While pretrained models such as BERT have shown large gains across natural language
understanding tasks, their performance can be improved by further training the model on a …

MRQA 2019 shared task: Evaluating generalization in reading comprehension

A Fisch, A Talmor, R Jia, M Seo, E Choi… - arxiv preprint arxiv …, 2019 - arxiv.org
We present the results of the Machine Reading for Question Answering (MRQA) 2019
shared task on evaluating the generalization capabilities of reading comprehension …

Attempt: Parameter-efficient multi-task tuning via attentional mixtures of soft prompts

A Asai, M Salehi, ME Peters, H Hajishirzi - arxiv preprint arxiv:2205.11961, 2022 - arxiv.org
This work introduces a new multi-task, parameter-efficient language model (LM) tuning
method that learns to transfer knowledge across different tasks via a mixture of soft prompts …

Limitations of transformers on clinical text classification

S Gao, M Alawad, MT Young, J Gounley… - IEEE journal of …, 2021 - ieeexplore.ieee.org
Bidirectional Encoder Representations from Transformers (BERT) and BERT-based
approaches are the current state-of-the-art in many natural language processing (NLP) …