wenda  by wenda-LLM

LLM platform for efficient, environment-specific content generation

created 2 years ago
6,242 stars

Top 8.4% on sourcepulse

GitHubView on GitHub
Project Summary

This project provides a platform for efficient content generation tailored for specific environments, addressing the computational resource limitations, knowledge security, and privacy concerns of individuals and small to medium-sized businesses. It integrates various large language models, local and online knowledge bases, and an extensible "Auto" scripting system for custom workflows.

How It Works

The platform supports a wide array of LLMs, including offline options like ChatGLM, RWKV, Llama, Baichuan, Aquila, and InternLM, as well as online APIs from OpenAI and ChatGLM. It features a robust knowledge base system that can connect to local offline vector databases, local search engines, and online search engines. The "Auto" scripting system, written in JavaScript, allows users to extend functionality by creating plugins for custom dialogue flows, external API access, and dynamic model switching.

Quick Start & Requirements

  • Installation: pip install -r requirements/requirements.txt (or use provided "lazy packages" for Windows).
  • Dependencies: Python, CUDA (for some model operations), specific model weights, and potentially compilers. Configuration is managed via config.yml.
  • Resources: Default parameters are stated to run well on 6GB VRAM GPUs. Pre-building indexes for the RTST knowledge base mode requires CUDA.
  • Links: Lazy package download links and QQ groups for discussion are provided.

Highlighted Details

  • Supports multiple LLMs with varying hardware requirements (CPU/GPU, quantization).
  • Knowledge base integrates with local vector stores (FAISS), local search (FESS), and online search (Bing).
  • "Auto" scripting system enables custom logic, API integrations, and UI extensions via JavaScript.
  • Offers features like dialogue history management, internal network deployment, and multi-user support.

Maintenance & Community

The project has active community engagement via QQ groups for general discussion, knowledge base usage, and Auto development. Specific user contributions for model fine-tuning are mentioned.

Licensing & Compatibility

The README does not explicitly state a license. Compatibility for commercial use or closed-source linking is not specified.

Limitations & Caveats

Some models (e.g., Llama, Moss) are noted as not recommended for Chinese users or require specific configurations (Baichuan with LoRA). Knowledge base data insertion has length and quantity limits, which can be mitigated by Auto scripts. Pre-building RTST indexes mandates CUDA, while runtime index building can use CPU for lower-VRAM systems.

Health Check
Last commit

6 months ago

Responsiveness

Inactive

Pull Requests (30d)
0
Issues (30d)
1
Star History
11 stars in the last 90 days

Explore Similar Projects

Starred by Chip Huyen Chip Huyen(Author of AI Engineering, Designing Machine Learning Systems) and Elie Bursztein Elie Bursztein(Cybersecurity Lead at Google DeepMind).

LightRAG by HKUDS

1.1%
19k
RAG framework for fast, simple retrieval-augmented generation
created 10 months ago
updated 1 day ago
Starred by Chip Huyen Chip Huyen(Author of AI Engineering, Designing Machine Learning Systems), Tim J. Baek Tim J. Baek(Founder of Open WebUI), and
2 more.

llmware by llmware-ai

0.2%
14k
Framework for enterprise RAG pipelines using small, specialized models
created 1 year ago
updated 1 week ago
Feedback? Help us improve.