Discover and explore top open-source AI tools and projects—updated daily.
Local-first AI framework for RAG and agentic applications
Top 60.6% on SourcePulse
LlamaFarm is an open-source framework designed for building retrieval-augmented generation (RAG) and agentic AI applications locally. It targets developers seeking a flexible, local-first environment to deploy and manage AI models, databases, and complex pipelines. LlamaFarm simplifies development through a unified CLI and a configuration-driven approach, allowing users to swap components like LLM providers or vector stores without extensive code rewrites, thereby accelerating AI application development.
How It Works
LlamaFarm provides a composable architecture for AI applications, prioritizing a local-first developer experience. It uses opinionated defaults such as Ollama for local model serving and Chroma for vector storage, but these are fully extendable. Users can integrate alternative runtimes like vLLM or OpenAI-compatible hosts, and various databases or parsers by modifying YAML configuration files. The framework enforces schema-based configuration, ensuring consistency and version control, and offers a friendly CLI (lf
) for project management, dataset handling, and interactive chat sessions.
Quick Start & Requirements
curl -fsSL https://raw.githubusercontent.com/llama-farm/llamafarm/main/install.sh | bash
winget install LlamaFarm.CLI
lf init my-project
(generates llamafarm.yaml
).lf start
.Highlighted Details
llamafarm.yaml
.Maintenance & Community
LlamaFarm fosters community engagement through Discord for real-time chat, GitHub Issues for bug reports and feature requests, and Discussions for proposals and roadmap discussions. A contributing guide outlines expectations for code style, testing, and documentation updates.
Licensing & Compatibility
The project is licensed under the Apache 2.0 License, which permits commercial use and integration into closed-source applications.
Limitations & Caveats
While designed for local-first development, the README emphasizes Ollama as the primary default local runtime, with other options being extendable. Users are advised to adjust Ollama's context window for optimal RAG performance with long documents. Detailed production deployment strategies beyond API compatibility are not explicitly covered in the README.
8 hours ago
Inactive