Why should we add early exits to neural networks?

S Scardapane, M Scarpiniti, E Baccarelli… - Cognitive Computation, 2020 - Springer
Deep neural networks are generally designed as a stack of differentiable layers, in which a
prediction is obtained only after running the full stack. Recently, some contributions have …

On efficient training of large-scale deep learning models: A literature review

L Shen, Y Sun, Z Yu, L Ding, X Tian, D Tao - arxiv preprint arxiv …, 2023 - arxiv.org
The field of deep learning has witnessed significant progress, particularly in computer vision
(CV), natural language processing (NLP), and speech. The use of large-scale models …

Occluded person re-identification

J Zhuo, Z Chen, J Lai, G Wang - 2018 IEEE international …, 2018 - ieeexplore.ieee.org
Person re-identification (re-id) suffers from a serious occlusion problem when applied to
crowded public places. In this paper, we propose to retrieve a full-body person image by …

Greedy layerwise learning can scale to imagenet

E Belilovsky, M Eickenberg… - … conference on machine …, 2019 - proceedings.mlr.press
Shallow supervised 1-hidden layer neural networks have a number of favorable properties
that make them easier to interpret, analyze, and optimize than their deep counterparts, but …

Efficienttrain++: Generalized curriculum learning for efficient visual backbone training

Y Wang, Y Yue, R Lu, Y Han, S Song… - IEEE Transactions on …, 2024 - ieeexplore.ieee.org
The superior performance of modern computer vision backbones (eg, vision Transformers
learned on ImageNet-1 K/22 K) usually comes with a costly training procedure. This study …

Efficienttrain: Exploring generalized curriculum learning for training visual backbones

Y Wang, Y Yue, R Lu, T Liu, Z Zhong… - Proceedings of the …, 2023 - openaccess.thecvf.com
The superior performance of modern deep networks usually comes with a costly training
procedure. This paper presents a new curriculum learning approach for the efficient training …

Automated progressive learning for efficient training of vision transformers

C Li, B Zhuang, G Wang, X Liang… - Proceedings of the …, 2022 - openaccess.thecvf.com
Recent advances in vision Transformers (ViTs) have come with a voracious appetite for
computing power, high-lighting the urgent need to develop efficient training methods for …

[PDF][PDF] Semi-Supervised Robust Deep Neural Networks for Multi-Label Classification.

H Cevikalp, B Benligiray, ÖN Gerek… - CVPR …, 2019 - openaccess.thecvf.com
In this paper, we propose a robust method for semisupervised training of deep neural
networks for multi-label image classification. To this end, we use ramp loss, which is more …

Early-exit deep neural network-a comprehensive survey

H Rahmath P, V Srivastava, K Chaurasia… - ACM Computing …, 2024 - dl.acm.org
Deep neural networks (DNNs) typically have a single exit point that makes predictions by
running the entire stack of neural layers. Since not all inputs require the same amount of …

Semi-supervised segmentation of salt bodies in seismic images using an ensemble of convolutional neural networks

Y Babakhin, A Sanakoyeu, H Kitamura - Pattern Recognition: 41st DAGM …, 2019 - Springer
Seismic image analysis plays a crucial role in a wide range of industrial applications and
has been receiving significant attention. One of the essential challenges of seismic imaging …