Image search demo using natural language queries
Top 98.4% on sourcepulse
This project provides a natural language-based image search engine using CLIP and NCNN, targeting mobile (Android) and desktop (x86) platforms. It enables users to find images within a gallery by describing them in text, offering a seamless search experience akin to built-in phone gallery features.
How It Works
The system leverages CLIP's encode_image
to extract features from gallery images, creating a feature vector database. For a given text query, CLIP's encode_text
generates a text feature vector. Similarity is then calculated between these vectors, allowing for text-to-image matching. The project currently displays the single highest probability match for simplicity but can be extended to return multiple relevant images.
Quick Start & Requirements
.bin
files for NCNN are available for download from the project's releases page and must be placed in the assert
folder of the respective demo projects.Highlighted Details
Maintenance & Community
The project notes that work is infrequent due to other commitments, with a focus on gaining stars. No community links or further contributor information are provided in the README.
Licensing & Compatibility
The README does not explicitly state a license. The project utilizes NCNN and CLIP, which have their own licenses. Compatibility for commercial use or closed-source linking is not specified.
Limitations & Caveats
The project is described as having slow updates and is primarily seeking stars. It currently only displays the top match, not a ranked list of results. The README does not specify the exact CLIP model version or licensing details.
2 years ago
Inactive