Discover and explore top open-source AI tools and projects—updated daily.
BradNeubergDeep learning pipeline for orbital satellite imagery analysis
Top 88.4% on SourcePulse
This project provides a deep learning pipeline for detecting clouds in orbital satellite imagery, primarily targeting users and startups like Planet Labs who need to pre-process vast amounts of data. It enables automated cloud detection and localization, facilitating downstream tasks such as change detection and deforestation monitoring, and is adaptable for other satellite imagery analysis tasks.
How It Works
The system comprises three core components: an annotation tool for bootstrapping training data by drawing bounding boxes on satellite images, a training pipeline that fine-tunes an AlexNet model using Caffe on GPU instances (e.g., AWS EC2) with annotated data, and a bounding box inference system that applies the trained model to new imagery to identify and delineate clouds. This modular approach allows for customization for various satellite detection challenges beyond cloud identification.
Quick Start & Requirements
brew install gdal, virtualenv, virtualenvwrapper, and pip install -r requirements.txt within a dedicated virtual environment (annotate-django). Data import involves downloading Planet Labs imagery via download_planetlabs.py and populating a database with populate_db.pip install -r requirements.txt in the root directory. The PYTHONPATH must include Caffe's Python bindings and ./src. Data preparation involves creating LevelDB files using prepare_data.py, followed by training with train.py.CAFFE_HOME and SELECTIVE_SEARCH environment variables must be set.Highlighted Details
Maintenance & Community
Contributors include Brad Neuberg, Johann Hauswald, and Max Nova. Parts of the project originated from Dropbox's Hack Week. It is released as version 1.0, with special thanks to Dropbox and Planet Labs. No specific community channels (like Discord/Slack) or roadmap are detailed in the README.
Licensing & Compatibility
The project is available under the Apache 2.0 license. While the code is permissively licensed, the use of Planet Labs data is subject to their ownership, and raw imagery is not publicly available.
Limitations & Caveats
The setup process is complex, requiring specific environment configurations, including Caffe, Python 2.7 for a critical dependency (Selective Search fork), and potentially AWS infrastructure for efficient training. Raw training data is not included due to copyright restrictions, necessitating users to acquire and prepare their own data.
9 years ago
Inactive
Shengjia Zhao(Chief Scientist at Meta Superintelligence Lab),
google
grahamjenson
google-research
triton-inference-server
tensorflow