Deep learning-based human pose estimation: A survey

C Zheng, W Wu, C Chen, T Yang, S Zhu, J Shen… - ACM Computing …, 2023 - dl.acm.org
Human pose estimation aims to locate the human body parts and build human body
representation (eg, body skeleton) from input data such as images and videos. It has drawn …

Recovering 3d human mesh from monocular images: A survey

Y Tian, H Zhang, Y Liu, L Wang - IEEE transactions on pattern …, 2023 - ieeexplore.ieee.org
Estimating human pose and shape from monocular images is a long-standing problem in
computer vision. Since the release of statistical body models, 3D human mesh recovery has …

Non-rigid neural radiance fields: Reconstruction and novel view synthesis of a dynamic scene from monocular video

E Tretschk, A Tewari, V Golyanik… - Proceedings of the …, 2021 - openaccess.thecvf.com
Abstract We present Non-Rigid Neural Radiance Fields (NR-NeRF), a reconstruction and
novel view synthesis approach for general non-rigid dynamic scenes. Our approach takes …

Function4d: Real-time human volumetric capture from very sparse consumer rgbd sensors

T Yu, Z Zheng, K Guo, P Liu… - Proceedings of the …, 2021 - openaccess.thecvf.com
Human volumetric capture is a long-standing topic in computer vision and computer
graphics. Although high-quality results can be achieved using sophisticated off-line systems …

Physical inertial poser (pip): Physics-aware real-time human motion tracking from sparse inertial sensors

X Yi, Y Zhou, M Habermann… - Proceedings of the …, 2022 - openaccess.thecvf.com
Motion capture from sparse inertial sensors has shown great potential compared to image-
based approaches since occlusions do not lead to a reduced tracking quality and the …

Structured local radiance fields for human avatar modeling

Z Zheng, H Huang, T Yu, H Zhang… - Proceedings of the …, 2022 - openaccess.thecvf.com
It is extremely challenging to create an animatable clothed human avatar from RGB videos,
especially for loose clothes due to the difficulties in motion modeling. To address this …

Vid2avatar: 3d avatar reconstruction from videos in the wild via self-supervised scene decomposition

C Guo, T Jiang, X Chen, J Song… - Proceedings of the …, 2023 - openaccess.thecvf.com
Abstract We present Vid2Avatar, a method to learn human avatars from monocular in-the-
wild videos. Reconstructing humans that move naturally from monocular in-the-wild videos …

Behave: Dataset and method for tracking human object interactions

BL Bhatnagar, X **e, IA Petrov… - Proceedings of the …, 2022 - openaccess.thecvf.com
Modelling interactions between humans and objects in natural environments is central to
many applications including gaming, virtual and mixed reality, as well as human behavior …

SCANimate: Weakly supervised learning of skinned clothed avatar networks

S Saito, J Yang, Q Ma, MJ Black - Proceedings of the IEEE …, 2021 - openaccess.thecvf.com
We present SCANimate, an end-to-end trainable framework that takes raw 3D scans of a
clothed human and turns them into an animatable avatar. These avatars are driven by pose …

4k4d: Real-time 4d view synthesis at 4k resolution

Z Xu, S Peng, H Lin, G He, J Sun… - Proceedings of the …, 2024 - openaccess.thecvf.com
This paper targets high-fidelity and real-time view synthesis of dynamic 3D scenes at 4K
resolution. Recent methods on dynamic view synthesis have shown impressive rendering …