Turnitin
降AI改写
早检测系统
早降重系统
Turnitin-UK版
万方检测-期刊版
维普编辑部版
Grammarly检测
Paperpass检测
checkpass检测
PaperYY检测
From show to tell: A survey on deep learning-based image captioning
Connecting Vision and Language plays an essential role in Generative Intelligence. For this
reason, large research efforts have been devoted to image captioning, ie describing images …
reason, large research efforts have been devoted to image captioning, ie describing images …
Visuals to text: A comprehensive review on automatic image captioning
Y Ming, N Hu, C Fan, F Feng… - IEEE/CAA Journal of …, 2022 - researchportal.port.ac.uk
Image captioning refers to automatic generation of descriptive texts according to the visual
content of images. It is a technique integrating multiple disciplines including the computer …
content of images. It is a technique integrating multiple disciplines including the computer …
Positive-augmented contrastive learning for image and video captioning evaluation
The CLIP model has been recently proven to be very effective for a variety of cross-modal
tasks, including the evaluation of captions generated from vision-and-language …
tasks, including the evaluation of captions generated from vision-and-language …
Injecting semantic concepts into end-to-end image captioning
Tremendous progress has been made in recent years in develo** better image captioning
models, yet most of them rely on a separate object detector to extract regional features …
models, yet most of them rely on a separate object detector to extract regional features …
BRIDGE: Bridging Gaps in Image Captioning Evaluation with Stronger Visual Cues
Effectively aligning with human judgment when evaluating machine-generated image
captions represents a complex yet intriguing challenge. Existing evaluation metrics like …
captions represents a complex yet intriguing challenge. Existing evaluation metrics like …
Emscore: Evaluating video captioning via coarse-grained and fine-grained embedding matching
Current metrics for video captioning are mostly based on the text-level comparison between
reference and candidate captions. However, they have some insuperable drawbacks, eg …
reference and candidate captions. However, they have some insuperable drawbacks, eg …
Improving image captioning descriptiveness by ranking and llm-based fusion
State-of-The-Art (SoTA) image captioning models often rely on the Microsoft COCO (MS-
COCO) dataset for training. This dataset contains annotations provided by human …
COCO) dataset for training. This dataset contains annotations provided by human …
InfoMetIC: an informative metric for reference-free image caption evaluation
Automatic image captioning evaluation is critical for benchmarking and promoting advances
in image captioning research. Existing metrics only provide a single score to measure …
in image captioning research. Existing metrics only provide a single score to measure …
Positive-Augmented Contrastive Learning for Vision-and-Language Evaluation and Training
Despite significant advancements in caption generation, existing evaluation metrics often fail
to capture the full quality or fine-grained details of captions. This is mainly due to their …
to capture the full quality or fine-grained details of captions. This is mainly due to their …
Deep learning approaches for image captioning: Opportunities, challenges and future potential
Generative intelligence relies heavily on the integration of vision and language. Much of the
research has focused on image captioning, which involves describing images with …
research has focused on image captioning, which involves describing images with …