Motionlcm: Real-time controllable motion generation via latent consistency model

W Dai, LH Chen, J Wang, J Liu, B Dai… - European Conference on …, 2024 - Springer
This work introduces MotionLCM, extending controllable motion generation to a real-time
level. Existing methods for spatial-temporal control in text-conditioned motion generation …

Emdm: Efficient motion diffusion model for fast and high-quality motion generation

W Zhou, Z Dou, Z Cao, Z Liao, J Wang, W Wang… - … on Computer Vision, 2024 - Springer
Abstract We introduce Efficient Motion Diffusion Model (EMDM) for fast and high-quality
human motion generation. Current state-of-the-art generative diffusion models have …

Tlcontrol: Trajectory and language control for human motion synthesis

W Wan, Z Dou, T Komura, W Wang… - … on Computer Vision, 2024 - Springer
Controllable human motion synthesis is essential for applications in AR/VR, gaming and
embodied AI. Existing methods often focus solely on either language or full trajectory control …

Disentangled clothed avatar generation from text descriptions

J Wang, Y Liu, Z Dou, Z Yu, Y Liang, C Lin… - … on Computer Vision, 2024 - Springer
In this paper, we introduce a novel text-to-avatar generation method that separately
generates the human body and the clothes and allows high-quality animation on the …

Plan, Posture and Go: Towards Open-Vocabulary Text-to-Motion Generation

J Liu, W Dai, C Wang, Y Cheng, Y Tang… - European Conference on …, 2024 - Springer
Conventional text-to-motion generation methods are usually trained on limited text-motion
pairs, making them hard to generalize to open-vocabulary scenarios. Some works use the …

DICE: End-to-end Deformation Capture of Hand-Face Interactions from a Single Image

Q Wu, Z Dou, S Xu, S Shimada, C Wang, Z Yu… - arxiv preprint arxiv …, 2024 - arxiv.org
Reconstructing 3D hand-face interactions with deformations from a single image is a
challenging yet crucial task with broad applications in AR, VR, and gaming. The challenges …

InterDreamer: Zero-Shot Text to 3D Dynamic Human-Object Interaction

S Xu, Z Wang, YX Wang, LY Gui - arxiv preprint arxiv:2403.19652, 2024 - arxiv.org
Text-conditioned human motion generation has experienced significant advancements with
diffusion models trained on extensive motion capture data and corresponding textual …

COLLAGE: Collaborative Human-Agent Interaction Generation using Hierarchical Latent Diffusion and Language Models

D Daiya, D Conover, A Bera - arxiv preprint arxiv:2409.20502, 2024 - arxiv.org
We propose a novel framework COLLAGE for generating collaborative agent-object-agent
interactions by leveraging large language models (LLMs) and hierarchical motion-specific …

[ΒΙΒΛΙΟ][B] Computer Vision-ECCV 2024: 18th European Conference, Milan, Italy, September 29-October 4, 2024, Proceedings, Part XXIV.

A Leonardis - 2024 - books.google.com
The multi-volume set of LNCS books with volume numbers 15059 up to 15147 constitutes
the refereed proceedings of the 18th European Conference on Computer Vision, ECCV …

Harmonizing Stochasticity and Determinism: Scene-responsive Diverse Human Motion Prediction

Z Lou, Q Cui, T Wang, Z Song, L Zhang… - The Thirty-eighth Annual … - openreview.net
Diverse human motion prediction (HMP) is a fundamental application in computer vision that
has recently attracted considerable interest. Prior methods primarily focus on the stochastic …