Omg-seg: Is one model good enough for all segmentation?

X Li, H Yuan, W Li, H Ding, S Wu… - Proceedings of the …, 2024 - openaccess.thecvf.com
In this work we address various segmentation tasks each traditionally tackled by distinct or
partially unified models. We propose OMG-Seg One Model that is Good enough to efficiently …

Mimic before reconstruct: Enhancing masked autoencoders with feature mimicking

P Gao, Z Lin, R Zhang, R Fang, H Li, H Li… - International Journal of …, 2024 - Springer
Masked Autoencoders (MAE) have been popular paradigms for large-scale vision
representation pre-training. However, MAE solely reconstructs the low-level RGB signals …

SCE-MAE: Selective Correspondence Enhancement with Masked Autoencoder for Self-Supervised Landmark Estimation

K Yin, V Rao, R Jiang, X Liu… - Proceedings of the …, 2024 - openaccess.thecvf.com
Self-supervised landmark estimation is a challenging task that demands the formation of
locally distinct feature representations to identify sparse facial landmarks in the absence of …

Event camera data dense pre-training

Y Yang, L Pan, L Liu - European Conference on Computer Vision, 2024 - Springer
This paper introduces a self-supervised learning framework designed for pre-training neural
networks tailored to dense prediction tasks using event camera data. Our approach utilizes …

Recent advances of local mechanisms in computer vision: a survey and outlook of recent work

Q Wang, Y Yin - arxiv preprint arxiv:2306.01929, 2023 - arxiv.org
Inspired by the fact that human brains can emphasize discriminative parts of the input and
suppress irrelevant ones, substantial local mechanisms have been designed to boost the …

Pre-training with random orthogonal projection image modeling

M Haghighat, P Moghadam, S Mohamed… - arxiv preprint arxiv …, 2023 - arxiv.org
Masked Image Modeling (MIM) is a powerful self-supervised strategy for visual pre-training
without the use of labels. MIM applies random crops to input images, processes them with …

Contrastive learning with consistent representations

Z Wang, Y Wang, Z Chen, H Hu, P Li - arxiv preprint arxiv:2302.01541, 2023 - arxiv.org
Contrastive learning demonstrates great promise for representation learning. Data
augmentations play a critical role in contrastive learning by providing informative views of …

Frequency-Guided Masking for Enhanced Vision Self-Supervised Learning

AK Monsefi, M Zhou, NK Monsefi, SN Lim… - arxiv preprint arxiv …, 2024 - arxiv.org
We present a novel frequency-based Self-Supervised Learning (SSL) approach that
significantly enhances its efficacy for pre-training. Prior work in this direction masks out pre …

PGP: Prior-Guided Pretraining for Small-sample Esophageal Cancer Segmentation

Q Shi, W Duan, W Chen, H Yang, H Lu… - 2024 IEEE …, 2024 - ieeexplore.ieee.org
Transformer-based models have demonstrated substantial potential in medical image
segmentation tasks due to their exceptional ability to capture long-range dependencies. To …

Self-Supervised Learning with Siamese Structure

Z Gao - 2024 - qmro.qmul.ac.uk
Recent progress in self-supervised representation learning has shown that self-supervised
pre-training can leverage unlabeled data to learn generalizable representations that benefit …