CONCH  by mahmoodlab

Vision-language model for computational pathology

Created 1 year ago
409 stars

Top 71.3% on SourcePulse

GitHubView on GitHub
1 Expert Loves This Project
Project Summary

CONCH is a vision-language foundation model for computational pathology, designed to advance AI in histopathology by integrating image and text data. It offers researchers and practitioners a powerful tool for a wide range of downstream tasks, including classification, segmentation, and retrieval, with the potential to reduce the need for extensive task-specific fine-tuning.

How It Works

CONCH employs contrastive learning from captions, pretraining on a large dataset of 1.17 million histopathology image-caption pairs. This approach allows it to learn rich representations from both visual and textual information, distinguishing it from models trained solely on images. This dual-modality learning enables better performance on non-H&E stained images and tasks requiring cross-modal understanding.

Quick Start & Requirements

  • Install via pip install -e . after cloning the repository and activating a Python 3.10 conda environment.
  • Requires PyTorch.
  • Model weights must be downloaded separately from Hugging Face after requesting access.
  • Detailed examples for loading and using the model are available in linked notebooks.

Highlighted Details

  • State-of-the-art performance on 14 diverse computational pathology benchmarks.
  • Trained on a unique dataset of 1.17M image-caption pairs, avoiding common public pathology datasets to prevent benchmark contamination.
  • Capable of zero-shot classification, image-text retrieval, and serving as a general-purpose encoder for histopathology images and text.
  • Benchmarked against multiple leading vision and vision-language models on various slide and ROI tasks.

Maintenance & Community

The project is associated with the Mahmood Lab. Recent updates include comparisons with other models like Virchow and Prov-GigaPath, and the release of related models like TITAN. The README lists numerous research applications and publications utilizing CONCH.

Licensing & Compatibility

Released under CC-BY-NC-ND 4.0 license. This strictly prohibits commercial use, sale, or monetization. Use is restricted to non-commercial, academic research purposes with proper attribution. Downloading requires Hugging Face registration and agreement to terms, including not distributing or reproducing the model.

Limitations & Caveats

The publicly released weights exclude the multimodal decoder due to potential PHI leakage concerns, though this does not affect the vision encoder's performance on key tasks. Commercial use is prohibited without prior approval.

Health Check
Last Commit

5 months ago

Responsiveness

Inactive

Pull Requests (30d)
0
Issues (30d)
0
Star History
11 stars in the last 30 days

Explore Similar Projects

Starred by Stas Bekman Stas Bekman(Author of "Machine Learning Engineering Open Book"; Research Engineer at Snowflake), Douwe Kiela Douwe Kiela(Cofounder of Contextual AI), and
1 more.

lens by ContextualAI

0.3%
353
Vision-language research paper using LLMs
Created 2 years ago
Updated 1 month ago
Starred by Jiayi Pan Jiayi Pan(Author of SWE-Gym; MTS at xAI), Shizhe Diao Shizhe Diao(Author of LMFlow; Research Scientist at NVIDIA), and
1 more.

METER by zdou0830

0%
373
Multimodal framework for vision-and-language transformer research
Created 3 years ago
Updated 2 years ago
Starred by Chip Huyen Chip Huyen(Author of "AI Engineering", "Designing Machine Learning Systems"), Simon Willison Simon Willison(Coauthor of Django), and
10 more.

LAVIS by salesforce

0.2%
11k
Library for language-vision AI research
Created 3 years ago
Updated 10 months ago
Feedback? Help us improve.