Recent advances in deep learning based dialogue systems: A systematic survey
Dialogue systems are a popular natural language processing (NLP) task as it is promising in
real-life applications. It is also a complicated task since many NLP tasks deserving study are …
real-life applications. It is also a complicated task since many NLP tasks deserving study are …
Lost in the middle: How language models use long contexts
While recent language models have the ability to take long contexts as input, relatively little
is known about how well they use longer context. We analyze the performance of language …
is known about how well they use longer context. We analyze the performance of language …
Adversarial NLI: A new benchmark for natural language understanding
We introduce a new large-scale NLI benchmark dataset, collected via an iterative,
adversarial human-and-model-in-the-loop procedure. We show that training models on this …
adversarial human-and-model-in-the-loop procedure. We show that training models on this …
[CITATION][C] Reasoning with transformer-based models: Deep learning, but shallow reasoning
Recent years have seen impressive performance of transformer-based models on different
natural language processing tasks. However, it is not clear to what degree the transformers …
natural language processing tasks. However, it is not clear to what degree the transformers …
Dynabench: Rethinking benchmarking in NLP
We introduce Dynabench, an open-source platform for dynamic dataset creation and model
benchmarking. Dynabench runs in a web browser and supports human-and-model-in-the …
benchmarking. Dynabench runs in a web browser and supports human-and-model-in-the …
Out of order: How important is the sequential order of words in a sentence in natural language understanding tasks?
Do state-of-the-art natural language understanding models care about word order-one of the
most important characteristics of a sequence? Not always! We found 75% to 90% of the …
most important characteristics of a sequence? Not always! We found 75% to 90% of the …
A model-agnostic data manipulation method for persona-based dialogue generation
Towards building intelligent dialogue agents, there has been a growing interest in
introducing explicit personas in generation models. However, with limited persona-based …
introducing explicit personas in generation models. However, with limited persona-based …
What context features can transformer language models use?
Transformer-based language models benefit from conditioning on contexts of hundreds to
thousands of previous tokens. What aspects of these contexts contribute to accurate model …
thousands of previous tokens. What aspects of these contexts contribute to accurate model …
Conversations are not flat: Modeling the dynamic information flow across dialogue utterances
Nowadays, open-domain dialogue models can generate acceptable responses according to
the historical context based on the large-scale pre-trained language models. However, they …
the historical context based on the large-scale pre-trained language models. However, they …
Tool wear estimation using a CNN-transformer model with semi-supervised learning
In the machining industry, tool wear has a great influence on machining efficiency, product
quality, and production costs. To achieve accurate tool wear estimation, a novel CNN …
quality, and production costs. To achieve accurate tool wear estimation, a novel CNN …