Learning from noisy labels with deep neural networks: A survey

H Song, M Kim, D Park, Y Shin… - IEEE transactions on …, 2022 - ieeexplore.ieee.org
Deep learning has achieved remarkable success in numerous domains with help from large
amounts of big data. However, the quality of data labels is a concern because of the lack of …

A survey on deep semi-supervised learning

X Yang, Z Song, I King, Z Xu - IEEE Transactions on …, 2022 - ieeexplore.ieee.org
Deep semi-supervised learning is a fast-growing field with a range of practical applications.
This paper provides a comprehensive survey on both fundamentals and recent advances in …

Generalized cross entropy loss for training deep neural networks with noisy labels

Z Zhang, M Sabuncu - Advances in neural information …, 2018 - proceedings.neurips.cc
Deep neural networks (DNNs) have achieved tremendous success in a variety of
applications across many disciplines. Yet, their superior performance comes with the …

Co-teaching: Robust training of deep neural networks with extremely noisy labels

B Han, Q Yao, X Yu, G Niu, M Xu… - Advances in neural …, 2018 - proceedings.neurips.cc
Deep learning with noisy labels is practically challenging, as the capacity of deep models is
so high that they can totally memorize these noisy labels sooner or later during training …

A survey on curriculum learning

X Wang, Y Chen, W Zhu - IEEE transactions on pattern analysis …, 2021 - ieeexplore.ieee.org
Curriculum learning (CL) is a training strategy that trains a machine learning model from
easier data to harder data, which imitates the meaningful learning order in human curricula …

Temporal ensembling for semi-supervised learning

S Laine, T Aila - arxiv preprint arxiv:1610.02242, 2016 - arxiv.org
In this paper, we present a simple and efficient method for training deep neural networks in a
semi-supervised setting where only a small portion of training data is labeled. We introduce …

Dividemix: Learning with noisy labels as semi-supervised learning

J Li, R Socher, SCH Hoi - arxiv preprint arxiv:2002.07394, 2020 - arxiv.org
Deep neural networks are known to be annotation-hungry. Numerous efforts have been
devoted to reducing the annotation cost when learning with deep networks. Two prominent …

Learning to reweight examples for robust deep learning

M Ren, W Zeng, B Yang… - … conference on machine …, 2018 - proceedings.mlr.press
Deep neural networks have been shown to be very powerful modeling tools for many
supervised learning tasks involving complex input patterns. However, they can also easily …

Mentornet: Learning data-driven curriculum for very deep neural networks on corrupted labels

L Jiang, Z Zhou, T Leung, LJ Li… - … conference on machine …, 2018 - proceedings.mlr.press
Recent deep networks are capable of memorizing the entire data even when the labels are
completely random. To overcome the overfitting on corrupted labels, we propose a novel …

Early-learning regularization prevents memorization of noisy labels

S Liu, J Niles-Weed, N Razavian… - Advances in neural …, 2020 - proceedings.neurips.cc
We propose a novel framework to perform classification via deep learning in the presence of
noisy annotations. When trained on noisy labels, deep neural networks have been observed …