Tool for training supervised models using foundation models, no labeling needed
Top 19.9% on sourcepulse
Autodistill enables users to train custom computer vision models without manual data labeling by leveraging large foundation models. It targets developers and researchers seeking to rapidly deploy efficient, specialized models for edge or cloud inference, bypassing the traditional bottleneck of data annotation.
How It Works
Autodistill employs a distillation pipeline: a large, capable "Base Model" (e.g., Grounding SAM, LLaVA) processes unlabeled images using an "Ontology" to generate auto-labeled datasets. These datasets then train a smaller, faster "Target Model" (e.g., YOLOv8, DETR), resulting in a deployable "Distilled Model." This approach democratizes model training by reducing reliance on human annotators and expensive labeling services.
Quick Start & Requirements
pip install autodistill autodistill-grounded-sam autodistill-yolov8
autodistill images --base="grounding_dino" --target="yolov8" --ontology '{"prompt": "label"}' --output="./dataset"
Highlighted Details
Maintenance & Community
Licensing & Compatibility
autodistill
package is licensed under Apache 2.0.Limitations & Caveats
2 months ago
1 week