Assemblyhands: Towards egocentric activity understanding via 3d hand pose estimation

T Ohkawa, K He, F Sener, T Hodan… - Proceedings of the …, 2023 - openaccess.thecvf.com
We present AssemblyHands, a large-scale benchmark dataset with accurate 3D hand pose
annotations, to facilitate the study of egocentric activities with challenging hand-object …

Dare-gram: Unsupervised domain adaptation regression by aligning inverse gram matrices

I Nejjar, Q Wang, O Fink - … of the IEEE/CVF conference on …, 2023 - openaccess.thecvf.com
Abstract Unsupervised Domain Adaptation Regression (DAR) aims to bridge the domain
gap between a labeled source dataset and an unlabelled target dataset for regression …

Weakly supervised temporal sentence grounding with uncertainty-guided self-training

Y Huang, L Yang, Y Sato - … of the IEEE/CVF conference on …, 2023 - openaccess.thecvf.com
The task of weakly supervised temporal sentence grounding aims at finding the
corresponding temporal moments of a language description in the video, given video …

Challenges and solutions for vision-based hand gesture interpretation: A review

K Gao, H Zhang, X Liu, X Wang, L **e, B Ji… - Computer Vision and …, 2024 - Elsevier
Hand gesture is one of the most efficient and natural interfaces in current human–computer
interaction (HCI) systems. Despite the great progress achieved in hand gesture-based HCI …

Single-to-dual-view adaptation for egocentric 3d hand pose estimation

R Liu, T Ohkawa, M Zhang… - Proceedings of the IEEE …, 2024 - openaccess.thecvf.com
The pursuit of accurate 3D hand pose estimation stands as a keystone for understanding
human activity in the realm of egocentric vision. The majority of existing estimation methods …

Benchmarks and challenges in pose estimation for egocentric hand interactions with objects

Z Fan, T Ohkawa, L Yang, N Lin, Z Zhou, S Zhou… - … on Computer Vision, 2024 - Springer
We interact with the world with our hands and see it through our own (egocentric)
perspective. A holistic 3D understanding of such interactions from egocentric views is …

Efficient annotation and learning for 3d hand pose estimation: A survey

T Ohkawa, R Furuta, Y Sato - International Journal of Computer Vision, 2023 - Springer
In this survey, we present a systematic review of 3D hand pose estimation from the
perspective of efficient annotation and learning. 3D hand pose estimation has been an …

Clip-hand3D: Exploiting 3D hand pose estimation via context-aware prompting

S Guo, Q Cai, L Qi, J Dong - Proceedings of the 31st ACM International …, 2023 - dl.acm.org
Contrastive Language-Image Pre-training (CLIP) starts to emerge in many computer vision
tasks and has achieved promising performance. However, it remains underexplored …

Maps: A noise-robust progressive learning approach for source-free domain adaptive keypoint detection

Y Ding, J Liang, B Jiang, A Zheng… - IEEE Transactions on …, 2023 - ieeexplore.ieee.org
Existing cross-domain keypoint detection methods always require accessing the source data
during adaptation, which may violate the data privacy law and pose serious security …

Pre-Training for 3D Hand Pose Estimation with Contrastive Learning on Large-Scale Hand Images in the Wild

N Lin, T Ohkawa, M Zhang, Y Huang, R Furuta… - arxiv preprint arxiv …, 2024 - arxiv.org
We present a contrastive learning framework based on in-the-wild hand images tailored for
pre-training 3D hand pose estimators, dubbed HandCLR. Pre-training on large-scale …