Real-time segmentation model for NVIDIA Jetson
Top 45.9% on sourcepulse
NanoSAM is a highly optimized variant of the Segment Anything (SAM) model, designed for real-time image segmentation on NVIDIA Jetson platforms. It targets developers and researchers working with edge AI applications, offering significantly reduced latency and resource requirements compared to larger models.
How It Works
NanoSAM achieves its performance by distilling a smaller MobileSAM model (specifically, the ResNet18 variant of the image encoder) using unlabeled images. This knowledge distillation process transfers capabilities from a larger teacher model to a more compact student model. The resulting NanoSAM model is then optimized for NVIDIA TensorRT, enabling efficient execution on edge devices.
Quick Start & Requirements
python3 setup.py develop --user
.torch2trt
, NVIDIA TensorRT (optional but recommended for engine building), transformers
(for OWL-ViT example), trt_pose
(for pose example).trtexec
.Highlighted Details
Maintenance & Community
The project is maintained by NVIDIA AI IoT. Links to relevant NVIDIA Jetson resources are provided.
Licensing & Compatibility
The repository does not explicitly state a license. However, it is built upon SAM and MobileSAM, which have permissive licenses. Compatibility for commercial use is not specified.
Limitations & Caveats
The MobileSAM image encoder requires FP32 precision in TensorRT due to erroneous results with FP16. The tracking example is experimental and may not be robust.
1 year ago
1 day