ComfyUI node for image segmentation using text prompts
Top 38.1% on sourcepulse
This ComfyUI custom node enables image segmentation using natural language prompts, leveraging GroundingDINO and Segment Anything (SAM) models. It targets users of the ComfyUI workflow manager who need precise object selection and masking capabilities within their image generation pipelines. The primary benefit is the ability to semantically identify and isolate any object in an image via text descriptions.
How It Works
The node integrates GroundingDINO for object detection based on text prompts and SAM for generating masks for detected objects. This two-stage approach allows for flexible and accurate segmentation, where GroundingDINO identifies potential objects matching the input text, and SAM refines these detections into precise masks. This combination offers a powerful way to interactively segment images using semantic understanding.
Quick Start & Requirements
pip3 install -r requirements.txt
ComfyUI/models/bert-base-uncased
, ComfyUI/models/grounding-dino
, and ComfyUI/models/sams
.requirements.txt
.Highlighted Details
sd-webui-segment-anything
.Maintenance & Community
continue-revolution
.Licensing & Compatibility
Limitations & Caveats
The README does not specify the license, which may impact commercial use. It also does not detail performance benchmarks or specific hardware requirements beyond standard Python dependencies.
1 year ago
1 week