Discover and explore top open-source AI tools and projects—updated daily.
withcataiCLI tool for local AI assistant using GGUF models
Top 64.0% on SourcePulse
CatAI provides a local AI assistant experience, enabling users to run GGUF models on their own computers with a chat UI and a simple Node.js API. It targets developers and power users seeking to leverage large language models offline, offering features like real-time streaming and fast model downloads.
How It Works
CatAI utilizes node-llama-cpp, a Node.js binding for llama.cpp, to run GGUF models efficiently. This approach allows for cross-platform compatibility (Windows, Linux, macOS) and leverages the performance optimizations of llama.cpp for local inference. The project also includes a CLI for model management and a web API for programmatic interaction.
Quick Start & Requirements
npm install -g cataicatai install meta-llama-3-8b-q4_k_mcatai updarwin-x64, linux-x64, win32-x64-msvc.~/catai.Highlighted Details
Maintenance & Community
Licensing & Compatibility
llama.cpp license, which is typically MIT.Limitations & Caveats
The project relies on node-llama-cpp which is in beta. Specific platform support depends on the underlying llama.cpp build for node-llama-cpp.
1 year ago
Inactive
abhishekkrthakur
smol-ai
xorbitsai
openinterpreter