Assemblyhands: Towards egocentric activity understanding via 3d hand pose estimation

T Ohkawa, K He, F Sener, T Hodan… - Proceedings of the …, 2023 - openaccess.thecvf.com
We present AssemblyHands, a large-scale benchmark dataset with accurate 3D hand pose
annotations, to facilitate the study of egocentric activities with challenging hand-object …

Aria digital twin: A new benchmark dataset for egocentric 3d machine perception

X Pan, N Charron, Y Yang, S Peters… - Proceedings of the …, 2023 - openaccess.thecvf.com
Abstract We introduce the Aria Digital Twin (ADT)-an egocentric dataset captured using Aria
glasses with extensive object, environment, and human level ground truth. This ADT release …

Put myself in your shoes: Lifting the egocentric perspective from exocentric videos

M Luo, Z Xue, A Dimakis, K Grauman - European Conference on Computer …, 2024 - Springer
We investigate exocentric-to-egocentric cross-view translation, which aims to generate a first-
person (egocentric) view of an actor based on a video recording that captures the actor from …

Ego-Humans: An Ego-Centric 3D Multi-Human Benchmark

R Khirodkar, A Bansal, L Ma… - Proceedings of the …, 2023 - openaccess.thecvf.com
We present EgoHumans, a new multi-view multi-human video benchmark to advance the
state-of-the-art of egocentric human 3D pose estimation and tracking. Existing egocentric …

In-hand 3d object scanning from an rgb sequence

S Hampali, T Hodan, L Tran, L Ma… - Proceedings of the …, 2023 - openaccess.thecvf.com
We propose a method for in-hand 3D scanning of an unknown object with a monocular
camera. Our method relies on a neural implicit surface representation that captures both the …

Imu2clip: Multimodal contrastive learning for imu motion sensors from egocentric videos and text

S Moon, A Madotto, Z Lin, A Dirafzoon, A Saraf… - arxiv preprint arxiv …, 2022 - arxiv.org
We present IMU2CLIP, a novel pre-training approach to align Inertial Measurement Unit
(IMU) motion sensor recordings with video and text, by projecting them into the joint …

RealityReplay: Detecting and Replaying Temporal Changes In Situ Using Mixed Reality

H Cho, ML Komar, D Lindlbauer - Proceedings of the ACM on Interactive …, 2023 - dl.acm.org
Humans easily miss events in their surroundings due to limited short-term memory and field
of view. This happens, for example, while watching an instructor's machine repair …

OmniActions: Predicting Digital Actions in Response to Real-World Multimodal Sensory Inputs with LLMs

JN Li, Y Xu, T Grossman, S Santosa, M Li - Proceedings of the CHI …, 2024 - dl.acm.org
The progression to “Pervasive Augmented Reality” envisions easy access to multimodal
information continuously. However, in many everyday scenarios, users are occupied …

SIMMC-VR: A Task-oriented Multimodal Dialog Dataset with Situated and Immersive VR Streams

TL Wu, S Kottur, A Madotto, M Azab… - Proceedings of the …, 2023 - aclanthology.org
Building an AI assistant that can seamlessly converse and instruct humans, in a user-centric
situated scenario, requires several essential abilities:(1) spatial and temporal understanding …

Egoblur: Responsible innovation in aria

N Raina, G Somasundaram, K Zheng, S Miglani… - arxiv preprint arxiv …, 2023 - arxiv.org
Project Aria pushes the frontiers of Egocentric AI with large-scale real-world data collection
using purposely designed glasses with privacy first approach. To protect the privacy of …