Sam-clip: Merging vision foundation models towards semantic and spatial understanding

H Wang, PKA Vasu, F Faghri… - Proceedings of the …, 2024 - openaccess.thecvf.com
The landscape of publicly available vision foundation models (VFMs) such as CLIP and
SAM is expanding rapidly. VFMs are endowed with distinct capabilities stemming from their …

Replay in minds and machines

L Wittkuhn, S Chien, S Hall-McMaster… - … & Biobehavioral Reviews, 2021 - Elsevier
Experience-related brain activity patterns reactivate during sleep, wakeful rest, and brief
pauses from active behavior. In parallel, machine learning research has found that …

Rainbow memory: Continual learning with a memory of diverse samples

J Bang, H Kim, YJ Yoo, JW Ha… - Proceedings of the IEEE …, 2021 - openaccess.thecvf.com
Continual learning is a realistic learning scenario for AI models. Prevalent scenario of
continual learning, however, assumes disjoint sets of classes as tasks and is less realistic …

Always be dreaming: A new approach for data-free class-incremental learning

J Smith, YC Hsu, J Balloch, Y Shen… - Proceedings of the …, 2021 - openaccess.thecvf.com
Modern computer vision applications suffer from catastrophic forgetting when incrementally
learning new concepts over time. The most successful approaches to alleviate this forgetting …

Open-vclip: Transforming clip to an open-vocabulary video model via interpolated weight optimization

Z Weng, X Yang, A Li, Z Wu… - … Conference on Machine …, 2023 - proceedings.mlr.press
Abstract Contrastive Language-Image Pretraining (CLIP) has demonstrated impressive zero-
shot learning abilities for image understanding, yet limited effort has been made to …

Architecture matters in continual learning

SI Mirzadeh, A Chaudhry, D Yin, T Nguyen… - ar** large foundation models up to date on latest data is inherently expensive. To avoid
the prohibitive costs of constantly retraining, it is imperative to continually train these models …

Building an open-vocabulary video CLIP model with better architectures, optimization and data

Z Wu, Z Weng, W Peng, X Yang, A Li… - … on Pattern Analysis …, 2024 - ieeexplore.ieee.org
Despite significant results achieved by Contrastive Language-Image Pretraining (CLIP) in
zero-shot image recognition, limited effort has been made exploring its potential for zero …