LLM framework for deep document understanding and RAG
New!
Top 26.1% on SourcePulse
WeKnora is an LLM-powered framework designed for deep document understanding, semantic retrieval, and context-aware question answering, leveraging the RAG paradigm. It targets enterprise knowledge management, research, technical support, legal, and medical domains, offering efficient and controllable document Q&A pipelines.
How It Works
WeKnora employs a modular architecture for a complete document understanding and retrieval pipeline. It integrates multi-modal preprocessing, semantic vector indexing, intelligent retrieval, and LLM inference. The core retrieval process uses RAG, combining contextually relevant document snippets with LLMs for high-quality semantic answers. This approach allows for flexible configuration and extension of each component.
Quick Start & Requirements
.env.example
to .env
and configure it, then run ./scripts/start_all.sh
or make start-all
.http://localhost
, API at http://localhost:8080
.Highlighted Details
Maintenance & Community
The project welcomes community contributions via Issues and Pull Requests, with guidelines for bug fixes, new features, documentation improvements, and UI/UX optimizations. Code contributions should follow Go Code Review Comments and use gofmt
. Commit messages should adhere to Conventional Commits.
Licensing & Compatibility
The project is released under the MIT license, allowing free use, modification, and distribution with attribution.
Limitations & Caveats
The README does not explicitly detail limitations or known issues. The project is developed by Tencent, suggesting potential enterprise-grade backing but also a possible focus on internal use cases.
2 days ago
Inactive