A survey on deep neural network pruning: Taxonomy, comparison, analysis, and recommendations
Modern deep neural networks, particularly recent large language models, come with
massive model sizes that require significant computational and storage resources. To …
massive model sizes that require significant computational and storage resources. To …
Deep metric learning: A survey
M Kaya, HŞ Bilge - Symmetry, 2019 - mdpi.com
Metric learning aims to measure the similarity among samples while using an optimal
distance metric for learning tasks. Metric learning methods, which generally use a linear …
distance metric for learning tasks. Metric learning methods, which generally use a linear …
Scaling vision transformers to 22 billion parameters
The scaling of Transformers has driven breakthrough capabilities for language models. At
present, the largest large language models (LLMs) contain upwards of 100B parameters …
present, the largest large language models (LLMs) contain upwards of 100B parameters …
Visual prompt tuning
The current modus operandi in adapting pre-trained models involves updating all the
backbone parameters, ie., full fine-tuning. This paper introduces Visual Prompt Tuning (VPT) …
backbone parameters, ie., full fine-tuning. This paper introduces Visual Prompt Tuning (VPT) …
Scaling & shifting your features: A new baseline for efficient model tuning
Existing fine-tuning methods either tune all parameters of the pre-trained model (full fine-
tuning), which is not efficient, or only tune the last linear layer (linear probing), which suffers …
tuning), which is not efficient, or only tune the last linear layer (linear probing), which suffers …
Scaling vision with sparse mixture of experts
Abstract Sparsely-gated Mixture of Experts networks (MoEs) have demonstrated excellent
scalability in Natural Language Processing. In Computer Vision, however, almost all …
scalability in Natural Language Processing. In Computer Vision, however, almost all …
Neural prompt search
The size of vision models has grown exponentially over the last few years, especially after
the emergence of Vision Transformer. This has motivated the development of parameter …
the emergence of Vision Transformer. This has motivated the development of parameter …
Convolutional bypasses are better vision transformer adapters
The pretrain-then-finetune paradigm has been widely adopted in computer vision. But as the
size of Vision Transformer (ViT) grows exponentially, the full finetuning becomes prohibitive …
size of Vision Transformer (ViT) grows exponentially, the full finetuning becomes prohibitive …
A fine-grained analysis on distribution shift
Robustness to distribution shifts is critical for deploying machine learning models in the real
world. Despite this necessity, there has been little work in defining the underlying …
world. Despite this necessity, there has been little work in defining the underlying …
Within the lack of chest COVID-19 X-ray dataset: a novel detection model based on GAN and deep transfer learning
The coronavirus (COVID-19) pandemic is putting healthcare systems across the world under
unprecedented and increasing pressure according to the World Health Organization (WHO) …
unprecedented and increasing pressure according to the World Health Organization (WHO) …