HyperAIHyperAI

Command Palette

Search for a command to run...

3 months ago

PMC-CLIP: Contrastive Language-Image Pre-training using Biomedical Documents

Weixiong Lin Ziheng Zhao Xiaoman Zhang Chaoyi Wu Ya Zhang Yanfeng Wang Weidi Xie

PMC-CLIP: Contrastive Language-Image Pre-training using Biomedical Documents

Abstract

Foundation models trained on large-scale dataset gain a recent surge in CV and NLP. In contrast, development in biomedical domain lags far behind due to data scarcity. To address this issue, we build and release PMC-OA, a biomedical dataset with 1.6M image-caption pairs collected from PubMedCentral's OpenAccess subset, which is 8 times larger than before. PMC-OA covers diverse modalities or diseases, with majority of the image-caption samples aligned at finer-grained level, i.e., subfigure and subcaption. While pretraining a CLIP-style model on PMC-OA, our model named PMC-CLIP achieves state-of-the-art results on various downstream tasks, including image-text retrieval on ROCO, MedMNIST image classification, Medical VQA, i.e. +8.1% R@10 on image-text retrieval, +3.9% accuracy on image classification.

Code Repositories

WeixiongLin/PMC-CLIP
Official
pytorch
mbzuai-oryx/unimed-clip
pytorch
Mentioned in GitHub

Benchmarks

BenchmarkMethodologyMetrics
visual-question-answering-vqa-on-pmc-vqaPMC-CLIP
Accuracy: 24.7

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp
PMC-CLIP: Contrastive Language-Image Pre-training using Biomedical Documents | Papers | HyperAI