A comprehensive survey on pretrained foundation models: A history from bert to chatgpt
Abstract Pretrained Foundation Models (PFMs) are regarded as the foundation for various
downstream tasks across different data modalities. A PFM (eg, BERT, ChatGPT, GPT-4) is …
downstream tasks across different data modalities. A PFM (eg, BERT, ChatGPT, GPT-4) is …
Rethinking semantic segmentation: A prototype view
Prevalent semantic segmentation solutions, despite their different network designs (FCN
based or attention based) and mask decoding strategies (parametric softmax based or pixel …
based or attention based) and mask decoding strategies (parametric softmax based or pixel …
Imagenet-21k pretraining for the masses
ImageNet-1K serves as the primary dataset for pretraining deep learning models for
computer vision tasks. ImageNet-21K dataset, which is bigger and more diverse, is used …
computer vision tasks. ImageNet-21K dataset, which is bigger and more diverse, is used …
With a little help from my friends: Nearest-neighbor contrastive learning of visual representations
Self-supervised learning algorithms based on instance discrimination train encoders to be
invariant to pre-defined transformations of the same instance. While most methods treat …
invariant to pre-defined transformations of the same instance. While most methods treat …
Vicreg: Variance-invariance-covariance regularization for self-supervised learning
Recent self-supervised methods for image representation learning are based on maximizing
the agreement between embedding vectors from different views of the same image. A trivial …
the agreement between embedding vectors from different views of the same image. A trivial …
Hard negative mixing for contrastive learning
Contrastive learning has become a key component of self-supervised learning approaches
for computer vision. By learning to embed two augmented versions of the same image close …
for computer vision. By learning to embed two augmented versions of the same image close …
Unsupervised learning of visual features by contrasting cluster assignments
Unsupervised image representations have significantly reduced the gap with supervised
pretraining, notably with the recent achievements of contrastive learning methods. These …
pretraining, notably with the recent achievements of contrastive learning methods. These …
Self-supervised learning: Generative or contrastive
Deep supervised learning has achieved great success in the last decade. However, its
defects of heavy dependence on manual labels and vulnerability to attacks have driven …
defects of heavy dependence on manual labels and vulnerability to attacks have driven …
Unsupervised semantic segmentation by contrasting object mask proposals
Being able to learn dense semantic representations of images without supervision is an
important problem in computer vision. However, despite its significance, this problem …
important problem in computer vision. However, despite its significance, this problem …
Self-supervised pretraining of visual features in the wild
Recently, self-supervised learning methods like MoCo, SimCLR, BYOL and SwAV have
reduced the gap with supervised methods. These results have been achieved in a control …
reduced the gap with supervised methods. These results have been achieved in a control …