Discover and explore top open-source AI tools and projects—updated daily.
zhaozihengUniversal 3D medical image segmentation via text prompts
Top 95.1% on SourcePulse
Large-vocabulary segmentation for medical images is addressed by SAT, a knowledge-enhanced universal segmentation model. It targets researchers and practitioners needing versatile segmentation across multiple modalities and anatomical structures. The primary benefit is a single, efficient model capable of segmenting hundreds of classes, replacing the need for numerous specialist models.
How It Works
SAT implements a knowledge-enhanced universal segmentation approach. It is built upon an extensive collection of 72 public 3D medical segmentation datasets, enabling segmentation of 497 classes across MR, CT, and PET modalities. Segmentation is guided by text prompts using anatomical terminology, offering a unified and efficient solution compared to training individual specialist models.
Quick Start & Requirements
pip install -e dynamic-network-architectures-main (from the model directory). Key requirements include torch>=1.10.0, numpy==1.21.5, monai==1.1.0, transformers==4.21.3, nibabel==4.0.2, einops==0.6.1, and positional_encodings==6.0.1. Install mamba_ssm for the U-Mamba variant.jsonl format.zzh99/SAT).Highlighted Details
Maintenance & Community
The project is associated with NPJ Digital Medicine and is a baseline for a prominent CVPR 2025 challenge, indicating active development and recognition. No specific community channels (Discord, Slack) are listed.
Licensing & Compatibility
The README does not explicitly state the software license. Compatibility for commercial use or closed-source linking is undetermined without a specified license.
Limitations & Caveats
High GPU memory requirements are necessary for both SAT-Pro (~34-62GB) and SAT-Nano (~24-36GB) during inference. Training demands substantial resources, recommending 8+ A100-80G GPUs for SAT-Nano and 16+ for SAT-Pro.
1 week ago
Inactive
OpenGVLab
milesial