Assemblyhands: Towards egocentric activity understanding via 3d hand pose estimation
We present AssemblyHands, a large-scale benchmark dataset with accurate 3D hand pose
annotations, to facilitate the study of egocentric activities with challenging hand-object …
annotations, to facilitate the study of egocentric activities with challenging hand-object …
Aria digital twin: A new benchmark dataset for egocentric 3d machine perception
Abstract We introduce the Aria Digital Twin (ADT)-an egocentric dataset captured using Aria
glasses with extensive object, environment, and human level ground truth. This ADT release …
glasses with extensive object, environment, and human level ground truth. This ADT release …
Put myself in your shoes: Lifting the egocentric perspective from exocentric videos
We investigate exocentric-to-egocentric cross-view translation, which aims to generate a first-
person (egocentric) view of an actor based on a video recording that captures the actor from …
person (egocentric) view of an actor based on a video recording that captures the actor from …
Ego-Humans: An Ego-Centric 3D Multi-Human Benchmark
We present EgoHumans, a new multi-view multi-human video benchmark to advance the
state-of-the-art of egocentric human 3D pose estimation and tracking. Existing egocentric …
state-of-the-art of egocentric human 3D pose estimation and tracking. Existing egocentric …
In-hand 3d object scanning from an rgb sequence
We propose a method for in-hand 3D scanning of an unknown object with a monocular
camera. Our method relies on a neural implicit surface representation that captures both the …
camera. Our method relies on a neural implicit surface representation that captures both the …
Imu2clip: Multimodal contrastive learning for imu motion sensors from egocentric videos and text
We present IMU2CLIP, a novel pre-training approach to align Inertial Measurement Unit
(IMU) motion sensor recordings with video and text, by projecting them into the joint …
(IMU) motion sensor recordings with video and text, by projecting them into the joint …
RealityReplay: Detecting and Replaying Temporal Changes In Situ Using Mixed Reality
Humans easily miss events in their surroundings due to limited short-term memory and field
of view. This happens, for example, while watching an instructor's machine repair …
of view. This happens, for example, while watching an instructor's machine repair …
OmniActions: Predicting Digital Actions in Response to Real-World Multimodal Sensory Inputs with LLMs
The progression to “Pervasive Augmented Reality” envisions easy access to multimodal
information continuously. However, in many everyday scenarios, users are occupied …
information continuously. However, in many everyday scenarios, users are occupied …
SIMMC-VR: A Task-oriented Multimodal Dialog Dataset with Situated and Immersive VR Streams
Building an AI assistant that can seamlessly converse and instruct humans, in a user-centric
situated scenario, requires several essential abilities:(1) spatial and temporal understanding …
situated scenario, requires several essential abilities:(1) spatial and temporal understanding …
Egoblur: Responsible innovation in aria
Project Aria pushes the frontiers of Egocentric AI with large-scale real-world data collection
using purposely designed glasses with privacy first approach. To protect the privacy of …
using purposely designed glasses with privacy first approach. To protect the privacy of …