Discover and explore top open-source AI tools and projects—updated daily.
LLM text generation server with multiple personalities
Top 89.9% on SourcePulse
LoLLMs Server provides a Flask-based API and WebSocket interface for generating text using various large language models. It's designed for developers and researchers to easily integrate LLM capabilities into applications, offering features like multiple personalities, real-time generation, and local data handling.
How It Works
The server acts as a unified interface to different LLM bindings, allowing users to select and load models from Hugging Face or local storage. It supports multiple "personalities" to influence generation style and uses WebSockets for real-time communication, enabling interactive chat applications. The safe_generate
method handles context window management, preventing overflow.
Quick Start & Requirements
pip install --upgrade lollms
or pip install --upgrade git+https://github.com/ParisNeo/lollms_legacy.git
conda create --name lollms python=3.10
, conda activate lollms
, conda install -c anaconda cudatoolkit
.lollms-settings
.Highlighted Details
llama_cpp_official
).Maintenance & Community
The project is hosted on GitHub at https://github.com/ParisNeo/lollms_legacy
. Contribution guidelines are available in CONTRIBUTING.md
.
Licensing & Compatibility
Licensed under the Apache 2.0 License. This license is permissive and generally compatible with commercial and closed-source applications.
Limitations & Caveats
The project is named lollms_legacy
, suggesting it might be an older or superseded version. Specific details on active maintenance or potential deprecation are not explicitly stated in the README.
1 month ago
Inactive