Pre-trained language models in biomedical domain: A systematic survey
Pre-trained language models (PLMs) have been the de facto paradigm for most natural
language processing tasks. This also benefits the biomedical domain: researchers from …
language processing tasks. This also benefits the biomedical domain: researchers from …
Clip in medical imaging: A comprehensive survey
Contrastive Language-Image Pre-training (CLIP), a simple yet effective pre-training
paradigm, successfully introduces text supervision to vision models. It has shown promising …
paradigm, successfully introduces text supervision to vision models. It has shown promising …
Making the most of text semantics to improve biomedical vision–language processing
Multi-modal data abounds in biomedicine, such as radiology images and reports.
Interpreting this data at scale is essential for improving clinical care and accelerating clinical …
Interpreting this data at scale is essential for improving clinical care and accelerating clinical …
Contrastive learning of medical visual representations from paired images and text
Learning visual representations of medical images (eg, X-rays) is core to medical image
understanding but its progress has been held back by the scarcity of human annotations …
understanding but its progress has been held back by the scarcity of human annotations …
Knowledge-enhanced visual-language pre-training on chest radiology images
While multi-modal foundation models pre-trained on large-scale data have been successful
in natural language understanding and vision recognition, their use in medical domains is …
in natural language understanding and vision recognition, their use in medical domains is …
[PDF][PDF] Large-scale domain-specific pretraining for biomedical vision-language processing
Contrastive pretraining on parallel image-text data has attained great success in vision-
language processing (VLP), as exemplified by CLIP and related methods. However, prior …
language processing (VLP), as exemplified by CLIP and related methods. However, prior …
Multi-granularity cross-modal alignment for generalized medical visual representation learning
Learning medical visual representations directly from paired radiology reports has become
an emerging topic in representation learning. However, existing medical image-text joint …
an emerging topic in representation learning. However, existing medical image-text joint …
Medklip: Medical knowledge enhanced language-image pre-training for x-ray diagnosis
In this paper, we consider enhancing medical visual-language pre-training (VLP) with
domain-specific knowledge, by exploiting the paired image-text reports from the radiological …
domain-specific knowledge, by exploiting the paired image-text reports from the radiological …
Prior: Prototype representation joint learning from medical images and reports
Contrastive learning based vision-language joint pre-training has emerged as a successful
representation learning strategy. In this paper, we present a prototype representation …
representation learning strategy. In this paper, we present a prototype representation …
Lvit: language meets vision transformer in medical image segmentation
Deep learning has been widely used in medical image segmentation and other aspects.
However, the performance of existing medical image segmentation models has been limited …
However, the performance of existing medical image segmentation models has been limited …