Open LLM by vivo AI Lab, featuring base and chat models
Top 40.4% on sourcepulse
BlueLM is an open-source suite of large language models developed by vivo AI Lab, offering 7B parameter models with both base and chat variants. It addresses the need for high-performance, multilingual LLMs with extended context capabilities, targeting researchers and developers. The models provide competitive performance on various benchmarks and support commercial applications under a permissive license.
How It Works
BlueLM models are trained on a 2.6 trillion token multilingual corpus, including Chinese, English, and some Japanese/Korean data. The 7B models are available in standard 2K context length and extended 32K context length versions. Quantized versions (4-bit AWQ/GPTQ) are also provided for reduced VRAM requirements. The project emphasizes strong performance on benchmarks like C-Eval and CMMLU, aiming for competitive results against other open-source models.
Quick Start & Requirements
pip install -r requirements.txt
.flash_attn
(version 2.3.3 or pre-compiled) is recommended.transformers
library, with provided CLI, web (Streamlit), and OpenAI-compatible API server demos.quant_cuda
installation.Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
The project statement strongly advises against using the models for activities that harm national social security or violate laws, and recommends thorough risk assessment for product applications. The developers disclaim responsibility for any issues arising from the use of the open-source models.
7 months ago
1 day