Android app for local SLM/LLM inference
Top 70.2% on sourcepulse
SmolChat-Android enables on-device inference of GGUF-formatted Large Language Models (LLMs) directly on Android devices. It targets users who want to run AI models locally for privacy, offline use, or experimentation, providing a straightforward interface for interacting with these models.
How It Works
The application leverages the ggerganov/llama.cpp
C++ library, which is compiled for Android using the NDK. A JNI binding (smollm.cpp
) facilitates communication between the Kotlin-based Android application and the C++ inference engine. This approach allows for efficient execution of LLMs on mobile hardware by utilizing llama.cpp
's optimized inference capabilities.
Quick Start & Requirements
https://github.com/shubham0204/SmolChat-Android
.git clone --depth=1 --recurse-submodules https://github.com/shubham0204/SmolChat-Android
) and build with Android Studio.Highlighted Details
llama.cpp
.Maintenance & Community
The project is maintained by shubham0204. Further community engagement details (Discord, Slack, etc.) are not specified in the README.
Licensing & Compatibility
The project appears to be licensed under the MIT License, allowing for commercial use and integration with closed-source applications.
Limitations & Caveats
The README mentions potential future integration with Vulkan for GPU acceleration, implying current inference may be CPU-bound. Some planned features like automatic chat naming and background services are not yet implemented.
1 month ago
1 day