Pre-train, prompt, and predict: A systematic survey of prompting methods in natural language processing

P Liu, W Yuan, J Fu, Z Jiang, H Hayashi… - ACM computing …, 2023 - dl.acm.org
This article surveys and organizes research works in a new paradigm in natural language
processing, which we dub “prompt-based learning.” Unlike traditional supervised learning …

A review on the attention mechanism of deep learning

Z Niu, G Zhong, H Yu - Neurocomputing, 2021 - Elsevier
Attention has arguably become one of the most important concepts in the deep learning
field. It is inspired by the biological systems of humans that tend to focus on the distinctive …

Holistic evaluation of language models

P Liang, R Bommasani, T Lee, D Tsipras… - arxiv preprint arxiv …, 2022 - arxiv.org
Language models (LMs) are becoming the foundation for almost all major language
technologies, but their capabilities, limitations, and risks are not well understood. We present …

Multiscale feature extraction and fusion of image and text in VQA

S Lu, Y Ding, M Liu, Z Yin, L Yin, W Zheng - International Journal of …, 2023 - Springer
Abstract The Visual Question Answering (VQA) system is the process of finding useful
information from images related to the question to answer the question correctly. It can be …

Learning to prompt for continual learning

Z Wang, Z Zhang, CY Lee, H Zhang… - Proceedings of the …, 2022 - openaccess.thecvf.com
The mainstream paradigm behind continual learning has been to adapt the model
parameters to non-stationary data distributions, where catastrophic forgetting is the central …

Finetuned language models are zero-shot learners

J Wei, M Bosma, VY Zhao, K Guu, AW Yu… - arxiv preprint arxiv …, 2021 - arxiv.org
This paper explores a simple method for improving the zero-shot learning abilities of
language models. We show that instruction tuning--finetuning language models on a …

Fantastically ordered prompts and where to find them: Overcoming few-shot prompt order sensitivity

Y Lu, M Bartolo, A Moore, S Riedel… - arxiv preprint arxiv …, 2021 - arxiv.org
When primed with only a handful of training samples, very large, pretrained language
models such as GPT-3 have shown competitive results when compared to fully-supervised …

Lift yourself up: Retrieval-augmented text generation with self-memory

X Cheng, D Luo, X Chen, L Liu… - Advances in Neural …, 2023 - proceedings.neurips.cc
With direct access to human-written reference as memory, retrieval-augmented generation
has achieved much progress in a wide range of text generation tasks. Since better memory …

Memot: Multi-object tracking with memory

J Cai, M Xu, W Li, Y **ong, W **a… - Proceedings of the …, 2022 - openaccess.thecvf.com
We propose an online tracking algorithm that performs the object detection and data
association under a common framework, capable of linking objects after a long time span …

Stmtrack: Template-free visual tracking with space-time memory networks

Z Fu, Q Liu, Z Fu, Y Wang - … of the IEEE/CVF conference on …, 2021 - openaccess.thecvf.com
Boosting performance of the offline trained siamese trackers is getting harder nowadays
since the fixed information of the template cropped from the first frame has been almost …