CLI tool for local AI assistant using GGUF models
Top 65.0% on sourcepulse
CatAI provides a local AI assistant experience, enabling users to run GGUF models on their own computers with a chat UI and a simple Node.js API. It targets developers and power users seeking to leverage large language models offline, offering features like real-time streaming and fast model downloads.
How It Works
CatAI utilizes node-llama-cpp
, a Node.js binding for llama.cpp
, to run GGUF models efficiently. This approach allows for cross-platform compatibility (Windows, Linux, macOS) and leverages the performance optimizations of llama.cpp
for local inference. The project also includes a CLI for model management and a web API for programmatic interaction.
Quick Start & Requirements
npm install -g catai
catai install meta-llama-3-8b-q4_k_m
catai up
darwin-x64
, linux-x64
, win32-x64-msvc
.~/catai
.Highlighted Details
Maintenance & Community
Licensing & Compatibility
llama.cpp
license, which is typically MIT.Limitations & Caveats
The project relies on node-llama-cpp
which is in beta. Specific platform support depends on the underlying llama.cpp
build for node-llama-cpp
.
1 year ago
1 day