Egoexolearn: A dataset for bridging asynchronous ego-and exo-centric view of procedural activities in real world

Y Huang, G Chen, J Xu, M Zhang… - Proceedings of the …, 2024 - openaccess.thecvf.com
Being able to map the activities of others into one's own point of view is one fundamental
human skill even from a very early age. Taking a step toward understanding this human …

A survey of visual affordance recognition based on deep learning

D Chen, D Kong, J Li, S Wang… - IEEE Transactions on Big …, 2023 - ieeexplore.ieee.org
Visual affordance recognition is an important research topic in robotics, human-computer
interaction, and other computer vision tasks. In recent years, deep learning-based …

Egochoir: Capturing 3d human-object interaction regions from egocentric views

Y Yang, W Zhai, C Wang, C Yu… - Advances in Neural …, 2025 - proceedings.neurips.cc
Understanding egocentric human-object interaction (HOI) is a fundamental aspect of human-
centric perception, facilitating applications like AR/VR and embodied AI. For the egocentric …

Empower dexterous robotic hand for human-centric smart manufacturing: A perception and skill learning perspective

B Gao, J Fan, P Zheng - Robotics and Computer-Integrated Manufacturing, 2025 - Elsevier
Recent rapid developments of dexterous robotic hands have greatly enhanced the
manipulative capabilities of robots, enabling them to perform industrial tasks in human-like …

Multi-label affordance map** from egocentric vision

L Mur-Labadia, JJ Guerrero… - Proceedings of the …, 2023 - openaccess.thecvf.com
Accurate affordance detection and segmentation with pixel precision is an important piece in
many complex systems based on interactions, such as robots and assitive devices. We …

Diagnosing human-object interaction detectors

F Zhu, Y **e, W **e, H Jiang - International Journal of Computer Vision, 2025 - Springer
We have witnessed significant progress in human-object interaction (HOI) detection.
However, relying solely on mAP (mean Average Precision) scores as a summary metric …

FineBio: a fine-grained video dataset of biological experiments with hierarchical annotation

T Yagi, M Ohashi, Y Huang, R Furuta, S Adachi… - arxiv preprint arxiv …, 2024 - arxiv.org
In the development of science, accurate and reproducible documentation of the
experimental process is crucial. Automatic recognition of the actions in experiments from …

Text-driven Affordance Learning from Egocentric Vision

T Yoshida, S Kurita, T Nishimura, S Mori - arxiv preprint arxiv:2404.02523, 2024 - arxiv.org
Visual affordance learning is a key component for robots to understand how to interact with
objects. Conventional approaches in this field rely on pre-defined objects and actions, falling …

Exo2egodvc: Dense video captioning of egocentric procedural activities using web instructional videos

T Ohkawa, T Yagi, T Nishimura, R Furuta… - arxiv preprint arxiv …, 2023 - arxiv.org
We propose a novel benchmark for cross-view knowledge transfer of dense video
captioning, adapting models from web instructional videos with exocentric views to an …

[HTML][HTML] Egocentric zone-aware action recognition across environments

SA Peirone, G Goletto, M Planamente, A Bottino… - Pattern Recognition …, 2025 - Elsevier
Human activities exhibit a strong correlation between actions and the places where these
are performed, such as washing something at a sink. More specifically, in daily living …