A primer on motion capture with deep learning: principles, pitfalls, and perspectives
Extracting behavioral measurements non-invasively from video is stymied by the fact that it is
a hard computational problem. Recent advances in deep learning have tremendously …
a hard computational problem. Recent advances in deep learning have tremendously …
Dense contrastive learning for self-supervised visual pre-training
To date, most existing self-supervised learning methods are designed and optimized for
image classification. These pre-trained models can be sub-optimal for dense prediction …
image classification. These pre-trained models can be sub-optimal for dense prediction …
Rethinking pre-training and self-training
Pre-training is a dominant paradigm in computer vision. For example, supervised ImageNet
pre-training is commonly used to initialize the backbones of object detection and …
pre-training is commonly used to initialize the backbones of object detection and …
Multi-task self-training for learning general representations
Despite the fast progress in training specialized models for various tasks, learning a single
general model that works well for many tasks is still challenging for computer vision. Here …
general model that works well for many tasks is still challenging for computer vision. Here …
On warm-starting neural network training
In many real-world deployments of machine learning systems, data arrive piecemeal. These
learning scenarios may be passive, where data arrive incrementally due to structural …
learning scenarios may be passive, where data arrive incrementally due to structural …
Pretraining boosts out-of-domain robustness for pose estimation
Neural networks are highly effective tools for pose estimation. However, as in other
computer vision tasks, robustness to out-of-domain data remains a challenge, especially for …
computer vision tasks, robustness to out-of-domain data remains a challenge, especially for …
A Survey of incremental deep learning for defect detection in manufacturing
Deep learning based visual cognition has greatly improved the accuracy of defect detection,
reducing processing times and increasing product throughput across a variety of …
reducing processing times and increasing product throughput across a variety of …
Gaia: A transfer learning system of object detection that fits your needs
Transfer learning with pre-training on large-scale datasets has played an increasingly
significant role in computer vision and natural language processing recently. However, as …
significant role in computer vision and natural language processing recently. However, as …
Neural data server: A large-scale search engine for transfer learning data
Transfer learning has proven to be a successful technique to train deep learning models in
the domains where little training data is available. The dominant approach is to pretrain a …
the domains where little training data is available. The dominant approach is to pretrain a …
Exploring visual interpretability for contrastive language-image pre-training
Contrastive Language-Image Pre-training (CLIP) learns rich representations via readily
available supervision of natural language. It improves the performance of downstream vision …
available supervision of natural language. It improves the performance of downstream vision …