Bilingual chat LLM for research/commercial use (after registration)
Top 3.1% on sourcepulse
ChatGLM2-6B is an open-source, bilingual conversational large language model designed for efficient deployment and strong performance. It targets researchers and developers looking for a capable LLM that can run on consumer hardware, offering significant improvements over its predecessor in various benchmarks and extended context handling.
How It Works
ChatGLM2-6B is built upon the GLM architecture, featuring a mixed objective function and pre-training on 1.4T tokens. It incorporates FlashAttention for extended context windows (up to 32K) and Multi-Query Attention for faster inference and reduced memory usage. The model has undergone human preference alignment training, contributing to its competitive performance on benchmarks like MMLU, CEval, GSM8K, and BBH.
Quick Start & Requirements
pip install -r requirements.txt
Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
The model is not guaranteed to be accurate and can be easily misled. The project team has not developed any official applications for the model. The README warns of potential data security and public opinion risks due to model misuse. Compatibility with PyTorch versions below 2.0 may lead to higher memory usage.
1 year ago
1 day