Discover and explore top open-source AI tools and projects—updated daily.
autodistillTool for training supervised models using foundation models, no labeling needed
Top 18.9% on SourcePulse
Autodistill enables users to train custom computer vision models without manual data labeling by leveraging large foundation models. It targets developers and researchers seeking to rapidly deploy efficient, specialized models for edge or cloud inference, bypassing the traditional bottleneck of data annotation.
How It Works
Autodistill employs a distillation pipeline: a large, capable "Base Model" (e.g., Grounding SAM, LLaVA) processes unlabeled images using an "Ontology" to generate auto-labeled datasets. These datasets then train a smaller, faster "Target Model" (e.g., YOLOv8, DETR), resulting in a deployable "Distilled Model." This approach democratizes model training by reducing reliance on human annotators and expensive labeling services.
Quick Start & Requirements
pip install autodistill autodistill-grounded-sam autodistill-yolov8autodistill images --base="grounding_dino" --target="yolov8" --ontology '{"prompt": "label"}' --output="./dataset"Highlighted Details
Maintenance & Community
Licensing & Compatibility
autodistill package is licensed under Apache 2.0.Limitations & Caveats
5 months ago
1 week
xtreme1-io
nidhaloff
apple
oumi-ai
GeorgeSeif