Egoexolearn: A dataset for bridging asynchronous ego-and exo-centric view of procedural activities in real world

Y Huang, G Chen, J Xu, M Zhang… - Proceedings of the …, 2024 - openaccess.thecvf.com
Being able to map the activities of others into one's own point of view is one fundamental
human skill even from a very early age. Taking a step toward understanding this human …

Vinci: A Real-time Embodied Smart Assistant based on Egocentric Vision-Language Model

Y Huang, J Xu, B Pei, Y He, G Chen, L Yang… - arxiv preprint arxiv …, 2024 - arxiv.org
We introduce Vinci, a real-time embodied smart assistant built upon an egocentric vision-
language model. Designed for deployment on portable devices such as smartphones and …

COM Kitchens: An Unedited Overhead-view Video Dataset as a Vision-Language Benchmark

K Maeda, T Hirasawa, A Hashimoto… - … on Computer Vision, 2024 - Springer
Procedural video understanding is gaining attention in the vision and language community.
Deep learning-based video analysis requires extensive data. Consequently, existing works …

Exocentric To Egocentric Transfer For Action Recognition: A Short Survey

A Thatipelli, SY Lo, AK Roy-Chowdhury - arxiv preprint arxiv:2410.20621, 2024 - arxiv.org
Egocentric vision captures the scene from the point of view of the camera wearer while
exocentric vision captures the overall scene context. Jointly modeling ego and exo views is …

BioVL-QR: Egocentric Biochemical Video-and-Language Dataset Using Micro QR Codes

T Nishimura, K Yamamoto, Y Haneji, K Kajimura… - arxiv preprint arxiv …, 2024 - arxiv.org
This paper introduces a biochemical vision-and-language dataset, which consists of 24
egocentric experiment videos, corresponding protocols, and video-and-language …

[KNIHA][B] Exocentric to Egocentric Transfer for Action Recognition

A Thatipelli - 2024 - search.proquest.com
Egocentric vision captures the scene from the point of the view of the camera wearer while
exocentric vision captures the overall scene context. Jointly modelling ego and exo views is …