Turnitin
降AI改写
早检测系统
早降重系统
Turnitin-UK版
万方检测-期刊版
维普编辑部版
Grammarly检测
Paperpass检测
checkpass检测
PaperYY检测
Multi-modal machine learning in engineering design: A review and future directions
In the rapidly advancing field of multi-modal machine learning (MMML), the convergence of
multiple data modalities has the potential to reshape various applications. This paper …
multiple data modalities has the potential to reshape various applications. This paper …
A survey on multimodal bidirectional machine learning translation of image and natural language processing
W Nam, B Jang - Expert Systems with Applications, 2024 - Elsevier
Advances in multimodal machine learning help artificial intelligence to resemble human
intellect more closely, which perceives the world from multiple modalities. We surveyed state …
intellect more closely, which perceives the world from multiple modalities. We surveyed state …
Inversion-based style transfer with diffusion models
The artistic style within a painting is the means of expression, which includes not only the
painting material, colors, and brushstrokes, but also the high-level attributes, including …
painting material, colors, and brushstrokes, but also the high-level attributes, including …
Long-clip: Unlocking the long-text capability of clip
Abstract Contrastive Language-Image Pre-training (CLIP) has been the cornerstone for zero-
shot classification, text-image retrieval, and text-image generation by aligning image and …
shot classification, text-image retrieval, and text-image generation by aligning image and …
Iterative prompt learning for unsupervised backlit image enhancement
We propose a novel unsupervised backlit image enhancement method, abbreviated as CLIP-
LIT, by exploring the potential of Contrastive Language-Image Pre-Training (CLIP) for pixel …
LIT, by exploring the potential of Contrastive Language-Image Pre-Training (CLIP) for pixel …
High-resolution image synthesis with latent diffusion models
By decomposing the image formation process into a sequential application of denoising
autoencoders, diffusion models (DMs) achieve state-of-the-art synthesis results on image …
autoencoders, diffusion models (DMs) achieve state-of-the-art synthesis results on image …
Text2live: Text-driven layered image and video editing
We present a method for zero-shot, text-driven editing of natural images and videos. Given
an image or a video and a text prompt, our goal is to edit the appearance of existing objects …
an image or a video and a text prompt, our goal is to edit the appearance of existing objects …
Zero-shot text-guided object generation with dream fields
We combine neural rendering with multi-modal image and text representations to synthesize
diverse 3D objects solely from natural language descriptions. Our method, Dream Fields …
diverse 3D objects solely from natural language descriptions. Our method, Dream Fields …
Avatarclip: Zero-shot text-driven generation and animation of 3d avatars
3D avatar creation plays a crucial role in the digital age. However, the whole production
process is prohibitively time-consuming and labor-intensive. To democratize this technology …
process is prohibitively time-consuming and labor-intensive. To democratize this technology …
Motionclip: Exposing human motion generation to clip space
We introduce MotionCLIP, a 3D human motion auto-encoder featuring a latent embedding
that is disentangled, well behaved, and supports highly semantic textual descriptions …
that is disentangled, well behaved, and supports highly semantic textual descriptions …