Turnitin
降AI改写
早检测系统
早降重系统
Turnitin-UK版
万方检测-期刊版
维普编辑部版
Grammarly检测
Paperpass检测
checkpass检测
PaperYY检测
Make-a-video: Text-to-video generation without text-video data
We propose Make-A-Video--an approach for directly translating the tremendous recent
progress in Text-to-Image (T2I) generation to Text-to-Video (T2V). Our intuition is simple …
progress in Text-to-Image (T2I) generation to Text-to-Video (T2V). Our intuition is simple …
Tryondiffusion: A tale of two unets
Given two images depicting a person and a garment worn by another person, our goal is to
generate a visualization of how the garment might look on the input person. A key challenge …
generate a visualization of how the garment might look on the input person. A key challenge …
Extracting motion and appearance via inter-frame attention for efficient video frame interpolation
Effectively extracting inter-frame motion and appearance information is important for video
frame interpolation (VFI). Previous works either extract both types of information in a mixed …
frame interpolation (VFI). Previous works either extract both types of information in a mixed …
Consistent view synthesis with pose-guided diffusion models
Novel view synthesis from a single image has been a cornerstone problem for many Virtual
Reality applications that provide immersive experiences. However, most existing techniques …
Reality applications that provide immersive experiences. However, most existing techniques …
Amt: All-pairs multi-field transforms for efficient frame interpolation
Abstract We present All-Pairs Multi-Field Transforms (AMT), a new network architecture for
video frame interpolation. It is based on two essential designs. First, we build bidirectional …
video frame interpolation. It is based on two essential designs. First, we build bidirectional …
Video interpolation with diffusion models
We present VIDIM a generative model for video interpolation which creates short videos
given a start and end frame. In order to achieve high fidelity and generate motions unseen in …
given a start and end frame. In order to achieve high fidelity and generate motions unseen in …
A vision chip with complementary pathways for open-world sensing
Image sensors face substantial challenges when dealing with dynamic, diverse and
unpredictable scenes in open-world applications. However, the development of image …
unpredictable scenes in open-world applications. However, the development of image …
Tooncrafter: Generative cartoon interpolation
We introduce ToonCrafter, a novel approach that transcends traditional correspondence-
based cartoon video interpolation, paving the way for generative interpolation. Traditional …
based cartoon video interpolation, paving the way for generative interpolation. Traditional …
Towards scalable neural representation for diverse videos
Implicit neural representations (INR) have gained increasing attention in representing 3D
scenes and images, and have been recently applied to encode videos (eg, NeRV, E-NeRV) …
scenes and images, and have been recently applied to encode videos (eg, NeRV, E-NeRV) …
Videogen: A reference-guided latent diffusion approach for high definition text-to-video generation
In this paper, we present VideoGen, a text-to-video generation approach, which can
generate a high-definition video with high frame fidelity and strong temporal consistency …
generate a high-definition video with high frame fidelity and strong temporal consistency …