Bilingual dialogue language model for research
Top 0.7% on sourcepulse
ChatGLM-6B is an open-source, bilingual (Chinese/English) dialogue language model based on the General Language Model (GLM) architecture. It offers a 6.2 billion parameter model optimized for Chinese question answering and dialogue, with the ability to run on consumer-grade GPUs with as little as 6GB VRAM through quantization.
How It Works
The model is trained on approximately 1T tokens of Chinese and English data, incorporating techniques like supervised fine-tuning, self-feedback, and Reinforcement Learning from Human Feedback (RLHF). This approach aims to generate human-preference aligned responses. For customization, it supports efficient parameter fine-tuning via P-Tuning v2, requiring minimal VRAM.
Quick Start & Requirements
pip install -r requirements.txt
(transformers version >= 4.23.1, recommended 4.27.1).gcc
and openmp
.Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
The model's 6B parameter size limits its factual recall, logical reasoning, and performance on complex tasks. It may generate biased or harmful content and has weaker English language capabilities compared to Chinese. The model is also susceptible to misdirection and has limited conversational robustness.
1 year ago
Inactive