wenda  by wenda-LLM

LLM platform for efficient, environment-specific content generation

Created 2 years ago
6,237 stars

Top 8.3% on SourcePulse

GitHubView on GitHub
Project Summary

This project provides a platform for efficient content generation tailored for specific environments, addressing the computational resource limitations, knowledge security, and privacy concerns of individuals and small to medium-sized businesses. It integrates various large language models, local and online knowledge bases, and an extensible "Auto" scripting system for custom workflows.

How It Works

The platform supports a wide array of LLMs, including offline options like ChatGLM, RWKV, Llama, Baichuan, Aquila, and InternLM, as well as online APIs from OpenAI and ChatGLM. It features a robust knowledge base system that can connect to local offline vector databases, local search engines, and online search engines. The "Auto" scripting system, written in JavaScript, allows users to extend functionality by creating plugins for custom dialogue flows, external API access, and dynamic model switching.

Quick Start & Requirements

  • Installation: pip install -r requirements/requirements.txt (or use provided "lazy packages" for Windows).
  • Dependencies: Python, CUDA (for some model operations), specific model weights, and potentially compilers. Configuration is managed via config.yml.
  • Resources: Default parameters are stated to run well on 6GB VRAM GPUs. Pre-building indexes for the RTST knowledge base mode requires CUDA.
  • Links: Lazy package download links and QQ groups for discussion are provided.

Highlighted Details

  • Supports multiple LLMs with varying hardware requirements (CPU/GPU, quantization).
  • Knowledge base integrates with local vector stores (FAISS), local search (FESS), and online search (Bing).
  • "Auto" scripting system enables custom logic, API integrations, and UI extensions via JavaScript.
  • Offers features like dialogue history management, internal network deployment, and multi-user support.

Maintenance & Community

The project has active community engagement via QQ groups for general discussion, knowledge base usage, and Auto development. Specific user contributions for model fine-tuning are mentioned.

Licensing & Compatibility

The README does not explicitly state a license. Compatibility for commercial use or closed-source linking is not specified.

Limitations & Caveats

Some models (e.g., Llama, Moss) are noted as not recommended for Chinese users or require specific configurations (Baichuan with LoRA). Knowledge base data insertion has length and quantity limits, which can be mitigated by Auto scripts. Pre-building RTST indexes mandates CUDA, while runtime index building can use CPU for lower-VRAM systems.

Health Check
Last Commit

7 months ago

Responsiveness

1 day

Pull Requests (30d)
0
Issues (30d)
0
Star History
2 stars in the last 30 days

Explore Similar Projects

Feedback? Help us improve.