An overview of deep semi-supervised learning

Y Ouali, C Hudelot, M Tami - arxiv preprint arxiv:2006.05278, 2020 - arxiv.org
Deep neural networks demonstrated their ability to provide remarkable performances on a
wide range of supervised learning tasks (eg, image classification) when trained on extensive …

A survey of multilingual neural machine translation

R Dabre, C Chu, A Kunchukuttan - ACM Computing Surveys (CSUR), 2020 - dl.acm.org
We present a survey on multilingual neural machine translation (MNMT), which has gained
a lot of traction in recent years. MNMT has been useful in improving translation quality as a …

Toolformer: Language models can teach themselves to use tools

T Schick, J Dwivedi-Yu, R Dessì… - Advances in …, 2023 - proceedings.neurips.cc
Abstract Language models (LMs) exhibit remarkable abilities to solve new tasks from just a
few examples or textual instructions, especially at scale. They also, paradoxically, struggle …

Self-instruct: Aligning language models with self-generated instructions

Y Wang, Y Kordi, S Mishra, A Liu, NA Smith… - arxiv preprint arxiv …, 2022 - arxiv.org
Large" instruction-tuned" language models (ie, finetuned to respond to instructions) have
demonstrated a remarkable ability to generalize zero-shot to new tasks. Nevertheless, they …

Large language models can self-improve

J Huang, SS Gu, L Hou, Y Wu, X Wang, H Yu… - arxiv preprint arxiv …, 2022 - arxiv.org
Large Language Models (LLMs) have achieved excellent performances in various tasks.
However, fine-tuning an LLM requires extensive supervision. Human, on the other hand …

Reinforced self-training (rest) for language modeling

C Gulcehre, TL Paine, S Srinivasan… - arxiv preprint arxiv …, 2023 - arxiv.org
Reinforcement learning from human feedback (RLHF) can improve the quality of large
language model's (LLM) outputs by aligning them with human preferences. We propose a …

Want to reduce labeling cost? GPT-3 can help

S Wang, Y Liu, Y Xu, C Zhu, M Zeng - arxiv preprint arxiv:2108.13487, 2021 - arxiv.org
Data annotation is a time-consuming and labor-intensive process for many NLP tasks.
Although there exist various methods to produce pseudo data labels, they are often task …

Rethinking pre-training and self-training

B Zoph, G Ghiasi, TY Lin, Y Cui, H Liu… - Advances in neural …, 2020 - proceedings.neurips.cc
Pre-training is a dominant paradigm in computer vision. For example, supervised ImageNet
pre-training is commonly used to initialize the backbones of object detection and …

Meta pseudo labels

H Pham, Z Dai, Q **e, QV Le - Proceedings of the IEEE/CVF …, 2021 - openaccess.thecvf.com
Abstract We present Meta Pseudo Labels, a semi-supervised learning method that achieves
a new state-of-the-art top-1 accuracy of 90.2% on ImageNet, which is 1.6% better than the …

Usb: A unified semi-supervised learning benchmark for classification

Y Wang, H Chen, Y Fan, W Sun… - Advances in …, 2022 - proceedings.neurips.cc
Semi-supervised learning (SSL) improves model generalization by leveraging massive
unlabeled data to augment limited labeled samples. However, currently, popular SSL …