Discover and explore top open-source AI tools and projects—updated daily.
On-premises RAG with configurable containers
Top 36.9% on SourcePulse
Minima provides an on-premises, containerized Retrieval Augmented Generation (RAG) solution for querying local documents. It targets users who need to keep their data private while integrating with popular LLMs like ChatGPT or Anthropic Claude, or operate entirely offline.
How It Works
Minima utilizes a containerized architecture, allowing flexible deployment modes. It supports fully isolated local operation using Ollama for LLM inference, or integration with external services like ChatGPT or Anthropic Claude. The system indexes local documents (PDF, XLS, DOCX, TXT, MD, CSV) using Sentence Transformer embedding models and a specified reranker, storing embeddings in Qdrant.
Quick Start & Requirements
docker compose
with specific files (docker-compose-ollama.yml
, docker-compose-chatgpt.yml
, docker-compose-mcp.yml
) and a .env
file. For Claude Desktop integration, npx -y @smithery/cli install minima --client claude
can be used.uv
(for MCP). Requires specifying LOCAL_FILES_PATH
, EMBEDDING_MODEL_ID
, EMBEDDING_SIZE
, OLLAMA_MODEL
(for local), RERANKER_MODEL
, USER_ID
, and PASSWORD
(for ChatGPT).Highlighted Details
http://localhost:3000
for the fully local setup.Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
The project is described as "open source RAG on-premises containers," implying it may still be under active development. Specific LLM and reranker compatibility beyond tested models (e.g., sentence-transformers/all-mpnet-base-v2
, BAAI rerankers) is not detailed.
1 month ago
1 day