A survey on deep neural network pruning: Taxonomy, comparison, analysis, and recommendations

H Cheng, M Zhang, JQ Shi - IEEE Transactions on Pattern …, 2024 - ieeexplore.ieee.org
Modern deep neural networks, particularly recent large language models, come with
massive model sizes that require significant computational and storage resources. To …

Deep metric learning: A survey

M Kaya, HŞ Bilge - Symmetry, 2019 - mdpi.com
Metric learning aims to measure the similarity among samples while using an optimal
distance metric for learning tasks. Metric learning methods, which generally use a linear …

Scaling vision transformers to 22 billion parameters

M Dehghani, J Djolonga, B Mustafa… - International …, 2023 - proceedings.mlr.press
The scaling of Transformers has driven breakthrough capabilities for language models. At
present, the largest large language models (LLMs) contain upwards of 100B parameters …

Visual prompt tuning

M Jia, L Tang, BC Chen, C Cardie, S Belongie… - … on Computer Vision, 2022 - Springer
The current modus operandi in adapting pre-trained models involves updating all the
backbone parameters, ie., full fine-tuning. This paper introduces Visual Prompt Tuning (VPT) …

Scaling & shifting your features: A new baseline for efficient model tuning

D Lian, D Zhou, J Feng, X Wang - Advances in Neural …, 2022 - proceedings.neurips.cc
Existing fine-tuning methods either tune all parameters of the pre-trained model (full fine-
tuning), which is not efficient, or only tune the last linear layer (linear probing), which suffers …

Scaling vision with sparse mixture of experts

C Riquelme, J Puigcerver, B Mustafa… - Advances in …, 2021 - proceedings.neurips.cc
Abstract Sparsely-gated Mixture of Experts networks (MoEs) have demonstrated excellent
scalability in Natural Language Processing. In Computer Vision, however, almost all …

Neural prompt search

Y Zhang, K Zhou, Z Liu - arxiv preprint arxiv:2206.04673, 2022 - arxiv.org
The size of vision models has grown exponentially over the last few years, especially after
the emergence of Vision Transformer. This has motivated the development of parameter …

Convolutional bypasses are better vision transformer adapters

S Jie, ZH Deng, S Chen, Z ** - ECAI 2024, 2024 - ebooks.iospress.nl
The pretrain-then-finetune paradigm has been widely adopted in computer vision. But as the
size of Vision Transformer (ViT) grows exponentially, the full finetuning becomes prohibitive …

A fine-grained analysis on distribution shift

O Wiles, S Gowal, F Stimberg, S Alvise-Rebuffi… - arxiv preprint arxiv …, 2021 - arxiv.org
Robustness to distribution shifts is critical for deploying machine learning models in the real
world. Despite this necessity, there has been little work in defining the underlying …

Within the lack of chest COVID-19 X-ray dataset: a novel detection model based on GAN and deep transfer learning

M Loey, F Smarandache, NE M. Khalifa - Symmetry, 2020 - mdpi.com
The coronavirus (COVID-19) pandemic is putting healthcare systems across the world under
unprecedented and increasing pressure according to the World Health Organization (WHO) …