Turnitin
降AI改写
早检测系统
早降重系统
Turnitin-UK版
万方检测-期刊版
维普编辑部版
Grammarly检测
Paperpass检测
checkpass检测
PaperYY检测
Vasa-1: Lifelike audio-driven talking faces generated in real time
We introduce VASA, a framework for generating lifelike talking faces with appealing visual
affective skills (VAS) given a single static image and a speech audio clip. Our premiere …
affective skills (VAS) given a single static image and a speech audio clip. Our premiere …
Facechain-imagineid: Freely crafting high-fidelity diverse talking faces from disentangled audio
In this paper we abstract the process of people hearing speech extracting meaningful cues
and creating various dynamically audio-consistent talking faces termed Listening and …
and creating various dynamically audio-consistent talking faces termed Listening and …
Deepfake generation and detection: A benchmark and survey
Deepfake is a technology dedicated to creating highly realistic facial images and videos
under specific conditions, which has significant application potential in fields such as …
under specific conditions, which has significant application potential in fields such as …
Portrait4d: Learning one-shot 4d head avatar synthesis using synthetic data
Existing one-shot 4D head synthesis methods usually learn from monocular videos with the
aid of 3DMM reconstruction yet the latter is evenly challenging which restricts them from …
aid of 3DMM reconstruction yet the latter is evenly challenging which restricts them from …
Vlogger: Multimodal diffusion for embodied avatar synthesis
We propose VLOGGER, a method for audio-driven human video generation from a single
input image of a person, which builds on the success of recent generative diffusion models …
input image of a person, which builds on the success of recent generative diffusion models …
Hallo: Hierarchical audio-driven visual synthesis for portrait image animation
The field of portrait image animation, driven by speech audio input, has experienced
significant advancements in the generation of realistic and dynamic portraits. This research …
significant advancements in the generation of realistic and dynamic portraits. This research …
Multilingual video dubbing—a technology review and current challenges
D Bigioi, P Corcoran - Frontiers in signal processing, 2023 - frontiersin.org
The proliferation of multi-lingual content on today's streaming services has created a need
for automated multi-lingual dubbing tools. In this article, current state-of-the-art approaches …
for automated multi-lingual dubbing tools. In this article, current state-of-the-art approaches …
A Comprehensive Survey on Human Video Generation: Challenges, Methods, and Insights
Human video generation is a dynamic and rapidly evolving task that aims to synthesize 2D
human body video sequences with generative models given control conditions such as text …
human body video sequences with generative models given control conditions such as text …
Learning to generate conditional tri-plane for 3d-aware expression controllable portrait animation
In this paper, we present Export3D, a one-shot 3D-aware portrait animation method that is
able to control the facial expression and camera view of a given portrait image. To achieve …
able to control the facial expression and camera view of a given portrait image. To achieve …
VOODOO 3D: volumetric portrait disentanglement for one-shot 3d head reenactment
We present a 3D-aware one-shot head reenactment method based on a fully volumetric
neural disentanglement framework for source appearance and driver expressions. Our …
neural disentanglement framework for source appearance and driver expressions. Our …