Recent advances in natural language processing via large pre-trained language models: A survey

B Min, H Ross, E Sulem, APB Veyseh… - ACM Computing …, 2023 - dl.acm.org
Large, pre-trained language models (PLMs) such as BERT and GPT have drastically
changed the Natural Language Processing (NLP) field. For numerous NLP tasks …

Paradigm shift in natural language processing

TX Sun, XY Liu, XP Qiu, XJ Huang - Machine Intelligence Research, 2022 - Springer
In the era of deep learning, modeling for most natural language processing (NLP) tasks has
converged into several mainstream paradigms. For example, we usually adopt the …

Explaining machine learning models with interactive natural language conversations using TalkToModel

D Slack, S Krishna, H Lakkaraju, S Singh - Nature Machine Intelligence, 2023 - nature.com
Practitioners increasingly use machine learning (ML) models, yet models have become
more complex and harder to understand. To understand complex models, researchers have …

Event extraction as machine reading comprehension

J Liu, Y Chen, K Liu, W Bi, X Liu - Proceedings of the 2020 …, 2020 - aclanthology.org
Event extraction (EE) is a crucial information extraction task that aims to extract event
information in texts. Previous methods for EE typically model it as a classification task, which …

Multi-task pre-training for plug-and-play task-oriented dialogue system

Y Su, L Shu, E Mansimov, A Gupta, D Cai… - arxiv preprint arxiv …, 2021 - arxiv.org
Pre-trained language models have been recently shown to benefit task-oriented dialogue
(TOD) systems. Despite their success, existing methods often formulate this task as a …

MultiWOZ 2.1: A consolidated multi-domain dialogue dataset with state corrections and state tracking baselines

M Eric, R Goel, S Paul, A Kumar, A Sethi, P Ku… - arxiv preprint arxiv …, 2019 - arxiv.org
MultiWOZ 2.0 (Budzianowski et al., 2018) is a recently released multi-domain dialogue
dataset spanning 7 distinct domains and containing over 10,000 dialogues. Though …

TOD-BERT: Pre-trained natural language understanding for task-oriented dialogue

CS Wu, S Hoi, R Socher, C **ong - arxiv preprint arxiv:2004.06871, 2020 - arxiv.org
The underlying difference of linguistic patterns between general text and task-oriented
dialogue makes existing pre-trained language models less useful in practice. In this work …

Mintl: Minimalist transfer learning for task-oriented dialogue systems

Z Lin, A Madotto, GI Winata, P Fung - arxiv preprint arxiv:2009.12005, 2020 - arxiv.org
In this paper, we propose Minimalist Transfer Learning (MinTL) to simplify the system design
process of task-oriented dialogue systems and alleviate the over-dependency on annotated …

Trippy: A triple copy strategy for value independent neural dialog state tracking

M Heck, C van Niekerk, N Lubis, C Geishauser… - arxiv preprint arxiv …, 2020 - arxiv.org
Task-oriented dialog systems rely on dialog state tracking (DST) to monitor the user's goal
during the course of an interaction. Multi-domain and open-vocabulary settings complicate …

Recent advances and challenges in task-oriented dialog systems

Z Zhang, R Takanobu, Q Zhu, ML Huang… - Science China …, 2020 - Springer
Due to the significance and value in human-computer interaction and natural language
processing, task-oriented dialog systems are attracting more and more attention in both …