Turnitin
降AI改写
早检测系统
早降重系统
Turnitin-UK版
万方检测-期刊版
维普编辑部版
Grammarly检测
Paperpass检测
checkpass检测
PaperYY检测
Hallucination of multimodal large language models: A survey
This survey presents a comprehensive analysis of the phenomenon of hallucination in
multimodal large language models (MLLMs), also known as Large Vision-Language Models …
multimodal large language models (MLLMs), also known as Large Vision-Language Models …
Rlaif-v: Aligning mllms through open-source ai feedback for super gpt-4v trustworthiness
Learning from feedback reduces the hallucination of multimodal large language models
(MLLMs) by aligning them with human preferences. While traditional methods rely on labor …
(MLLMs) by aligning them with human preferences. While traditional methods rely on labor …
Llafs: When large language models meet few-shot segmentation
This paper proposes LLaFS the first attempt to leverage large language models (LLMs) in
few-shot segmentation. In contrast to the conventional few-shot segmentation methods that …
few-shot segmentation. In contrast to the conventional few-shot segmentation methods that …
Controlmllm: Training-free visual prompt learning for multimodal large language models
In this work, we propose a training-free method to inject visual prompts into Multimodal
Large Language Models (MLLMs) through learnable latent variable optimization. We …
Large Language Models (MLLMs) through learnable latent variable optimization. We …
Model tailor: Mitigating catastrophic forgetting in multi-modal large language models
Catastrophic forgetting emerges as a critical challenge when fine-tuning multi-modal large
language models (MLLMs), where improving performance on unseen tasks often leads to a …
language models (MLLMs), where improving performance on unseen tasks often leads to a …
Agla: Mitigating object hallucinations in large vision-language models with assembly of global and local attention
Despite their great success across various multimodal tasks, Large Vision-Language
Models (LVLMs) are facing a prevalent problem with object hallucinations, where the …
Models (LVLMs) are facing a prevalent problem with object hallucinations, where the …
Sam2-adapter: Evaluating & adapting segment anything 2 in downstream tasks: Camouflage, shadow, medical image segmentation, and more
The advent of large models, also known as foundation models, has significantly transformed
the AI research landscape, with models like Segment Anything (SAM) achieving notable …
the AI research landscape, with models like Segment Anything (SAM) achieving notable …
Alleviating hallucination in large vision-language models with active retrieval augmentation
Despite the remarkable ability of large vision-language models (LVLMs) in image
comprehension, these models frequently generate plausible yet factually incorrect …
comprehension, these models frequently generate plausible yet factually incorrect …
Discrete latent perspective learning for segmentation and detection
In this paper, we address the challenge of Perspective-Invariant Learning in machine
learning and computer vision, which involves enabling a network to understand images from …
learning and computer vision, which involves enabling a network to understand images from …
Eventhallusion: Diagnosing event hallucinations in video llms
Recently, Multimodal Large Language Models (MLLMs) have made significant progress in
the video comprehension field. Despite remarkable content reasoning and instruction …
the video comprehension field. Despite remarkable content reasoning and instruction …