Discover and explore top open-source AI tools and projects—updated daily.
AI-FanGeAI navigation and assistance system for the visually impaired
Top 35.7% on SourcePulse
Summary
AI-FanGe/OpenAIglasses_for_Navigation is an open-source framework for AI-driven assistive navigation and interaction, designed for visually impaired users. It integrates computer vision and NLP to provide real-time guidance for navigation, object recognition, and environmental awareness, aiming to enhance independence and safety.
How It Works
The system uses a FastAPI backend to process real-time video/audio streams. It leverages deep learning models like YOLOv8 for segmentation (blind paths, lanes) and YOLO-E for open-vocabulary object search, alongside MediaPipe for hand tracking. Optical flow (Lucas-Kanade) stabilizes input, while Aliyun DashScope provides ASR and multimodal chat (Qwen-Omni-Turbo) for voice interaction. Feedback is multimodal: visual annotations, voice, and hand guidance.
Quick Start & Requirements
pip install -r requirements.txt. GPU acceleration requires CUDA 11.8+..env), and running python app_main.py.Highlighted Details
Maintenance & Community
The project is explicitly for "exchange and learning only" and not production-ready. The README provides no details on active maintenance, contributors, or community channels (e.g., Discord, Slack).
Licensing & Compatibility
Released under the permissive MIT License, allowing broad usage, including commercial applications and integration into closed-source projects, with standard attribution requirements.
Limitations & Caveats
2 months ago
Inactive
open-mmlab
dnhkng
OthersideAI