Framework for autonomous Android/computer use via LLMs
Top 64.7% on sourcepulse
This framework enables autonomous control of Android and macOS devices using Large Language Models (LLMs), allowing users to automate complex tasks through natural language prompts. It targets developers and power users seeking to integrate LLM capabilities into device automation workflows.
How It Works
The system leverages a dual-LLM approach: a "planner" LLM to break down tasks into actionable steps and a "finder" LLM to identify UI elements on the target device. It supports various LLMs, including OpenAI's GPT models, Google's Gemini, and local models via Ollama. The framework interacts with Android devices via ADB and macOS via internal scripting.
Quick Start & Requirements
pip install git+https://github.com/BandarLabs/clickclickclick.git
adb
installed, USB debugging enabled on Android, Python >= 3.11.OPENAI_API_KEY
, GEMINI_API_KEY
) in environment variables or config/models.yaml
.click3 run "task"
), Gradio web UI (click3 gradio
), or API (Uvicorn server).Highlighted Details
Maintenance & Community
The project is actively developed by BandarLabs. Community contributions are encouraged via GitHub issues and pull requests.
Licensing & Compatibility
Limitations & Caveats
The project is described as "highly experimental" and may evolve significantly. Current macOS support is noted as not fully functional.
2 days ago
1 day