Discover and explore top open-source AI tools and projects—updated daily.
veylissLocal-first RAG system for document Q&A and LLM integration
Top 91.3% on SourcePulse
A local-first, self-hostable Retrieval Augmented Generation (RAG) system designed to integrate local documents with large language models for conversational search. It targets individuals and small teams seeking a private, customizable AI knowledge base, offering support for various document types and flexible model deployment via Ollama or OpenAI-compatible APIs. The system provides a web UI for knowledge base management, document ingestion, and persistent chat history, enabling rapid prototyping and local AI application development.
How It Works
AI LocalBase employs a RAG architecture, leveraging Qdrant as its vector database. Documents (TXT, Markdown, PDF, XLSX, CSV) are processed through text splitting, embedding, and indexing. User queries trigger vector searches in Qdrant, with relevant context dynamically injected into prompts sent to LLMs accessed via Ollama or OpenAI-compatible endpoints. The system features local persistence for chat history (SQLite) and configuration (JSON), alongside advanced retrieval strategies like MMR de-duplication and optional hybrid search for enhanced accuracy.
Quick Start & Requirements
The quickest way to get started is via Docker Compose, using docker compose up --build. For local development, prerequisites include Docker, Go, and Node.js. Users will need to configure chat and embedding models, with examples provided for Ollama and OpenAI-compatible APIs. Detailed setup, environment variables, and API explanations are available in docs/getting-started.md.
Highlighted Details
Maintenance & Community
The project includes contribution (CONTRIBUTING.md) and security (SECURITY.md) guidelines, indicating a structured approach to development. Specific details on active contributors, community channels (like Discord/Slack), or sponsorship are not explicitly detailed in the provided README.
Licensing & Compatibility
The project is open-source, with a LICENSE file present. Specific license type and compatibility notes for commercial use or closed-source linking are not detailed in the provided text. It is designed for local and self-hosted environments.
Limitations & Caveats
PDF support is described as "text" based, suggesting that advanced OCR capabilities might not be included or are basic. The README focuses on core functionality and setup; more in-depth details on specific retrieval optimizations, MCP capabilities, and deployment variations are located in separate documentation files.
13 hours ago
Inactive
tobi