Turnitin
降AI改写
早检测系统
早降重系统
Turnitin-UK版
万方检测-期刊版
维普编辑部版
Grammarly检测
Paperpass检测
checkpass检测
PaperYY检测
Moda: Map**-once audio-driven portrait animation with dual attentions
Audio-driven portrait animation aims to synthesize portrait videos that are conditioned by
given audio. Animating high-fidelity and multimodal video portraits has a variety of …
given audio. Animating high-fidelity and multimodal video portraits has a variety of …
Revisiting generalizability in deepfake detection: Improving metrics and stabilizing transfer
Abstract" Generalizability" is seen as the hallmark quality of a good deepfake detection
model. However, standard out-of-domain evaluation datasets are very similar in form to the …
model. However, standard out-of-domain evaluation datasets are very similar in form to the …
Dagan++: Depth-aware generative adversarial network for talking head video generation
Predominant techniques on talking head generation largely depend on 2D information,
including facial appearances and motions from input face images. Nevertheless, dense 3D …
including facial appearances and motions from input face images. Nevertheless, dense 3D …
Multimodal-driven talking face generation via a unified diffusion-based generator
Multimodal-driven talking face generation refers to animating a portrait with the given pose,
expression, and gaze transferred from the driving image and video, or estimated from the …
expression, and gaze transferred from the driving image and video, or estimated from the …
Make your actor talk: Generalizable and high-fidelity lip sync with motion and appearance disentanglement
We aim to edit the lip movements in talking video according to the given speech while
preserving the personal identity and visual details. The task can be decomposed into two …
preserving the personal identity and visual details. The task can be decomposed into two …
Dreamhead: Learning spatial-temporal correspondence via hierarchical diffusion for audio-driven talking head synthesis
Audio-driven talking head synthesis strives to generate lifelike video portraits from provided
audio. The diffusion model, recognized for its superior quality and robust generalization, has …
audio. The diffusion model, recognized for its superior quality and robust generalization, has …
EAT-Face: Emotion-Controllable Audio-Driven Talking Face Generation via Diffusion Model
Audio-driven talking face generation is a promising task with a lot of attention. Despite
abundant efforts are devoted to video quality and lip synchronization, most existing works do …
abundant efforts are devoted to video quality and lip synchronization, most existing works do …