Discover and explore top open-source AI tools and projects—updated daily.
DenisovAVLocal AI inference SDK for Flutter apps
Top 95.1% on SourcePulse
This Flutter plugin enables on-device execution of various large language models, including Gemma, TinyLlama, Phi, and others, directly within iOS, Android, and Web applications. It targets Flutter developers seeking to integrate advanced AI capabilities like text generation, multimodal input, and function calling into their applications without relying on external servers, thereby enhancing user privacy and enabling offline functionality.
How It Works
The plugin leverages MediaPipe and LiterTLM formats for efficient on-device inference. It manages the download, storage, and initialization of diverse LLM models, supporting both CPU and GPU backends. Key features include multimodal (text + image) input for specific models like Gemma 3 Nano, function calling capabilities for integrating with external services, and a "thinking mode" for DeepSeek models to expose their reasoning process. It also supports LoRA weights for efficient model fine-tuning and provides robust download management with retry logic.
Quick Start & Requirements
flutter_gemma: latest_version to pubspec.yaml and run flutter pub get.platform :ios, '16.0', file sharing enabled, local network usage description, memory entitlements (com.apple.developer.kernel.extended-virtual-addressing, etc.), and static pod linkage (use_frameworks! :linkage => :static).Highlighted Details
Maintenance & Community
The provided README does not contain specific details regarding maintainers, community channels (like Discord/Slack), or project roadmaps.
Licensing & Compatibility
The license for this plugin is not explicitly stated in the provided README. Compatibility for commercial use or linking with closed-source applications would depend on the underlying model licenses and the plugin's own license, which requires clarification.
Limitations & Caveats
Web platform support is restricted to GPU backend models. Larger models (e.g., 7B parameters) may be too resource-intensive for typical on-device inference. Multimodal models require significant memory (8GB+ RAM recommended) and specific iOS configurations (iOS 16.0+, memory entitlements). Function calling and thinking mode are only available for specific, supported models. The plugin's license is not specified.
1 week ago
Inactive
sgomez
pytorch