Image search tool using natural language and computer vision
Top 58.3% on sourcepulse
Memery provides natural language and computer vision-based search for local image collections, targeting users overwhelmed by large, unorganized image folders. It enables searching with descriptive text or example images, offering a significant improvement over manual thumbnail browsing for locating specific visual content.
How It Works
Memery leverages OpenAI's CLIP (Contrastive Language-Image Pretraining) model. CLIP learns a shared latent space for images and their textual descriptions by training a vision transformer and a language transformer. This allows Memery to effectively map natural language queries to relevant images within a local dataset, facilitating semantic search.
Quick Start & Requirements
pip install git+https://github.com/deepfates/memery.git
or pip install memery
(after cloning and poetry install
). Also requires pip install git+https://github.com/openai/CLIP.git
.pip install torch==1.7.1+cpu torchvision==0.8.2+cpu torchaudio==0.7.2 -f https://download.pytorch.org/whl/torch_stable.html
).memery serve [directory]
memery recall PATH/TO/IMAGE/FOLDER -t 'text_query'
Memery
from memery.core
.Highlighted Details
Maintenance & Community
The project was initially developed by Max Anton Brewer (@deepfates), with significant contributions from @wkrettek. The primary contact for support is via GitHub Issues or direct message on Twitter.
Licensing & Compatibility
The repository does not explicitly state a license in the README. This requires clarification for commercial use or integration into closed-source projects.
Limitations & Caveats
The project is noted as not yet optimized for performance, with search times scaling linearly with dataset size (O(n)). The GUI's directory selection is basic, relying on path strings rather than a file picker. The default installation prioritizes GPU usage, requiring manual steps for CPU-only operation.
9 months ago
1 week