GLM-130B  by zai-org

Bilingual model for research and evaluation

created 3 years ago
7,685 stars

Top 6.9% on sourcepulse

GitHubView on GitHub
Project Summary

GLM-130B is an open-source, 130-billion parameter, bilingual (English/Chinese) language model designed for researchers and developers working with large-scale NLP models. It offers strong performance on various benchmarks and supports efficient inference through quantization and optimized libraries.

How It Works

GLM-130B utilizes the General Language Model (GLM) pre-training approach, which combines autoregressive blank infilling with bidirectional context. This allows it to excel at both left-to-right generation and filling in masked segments of text. The model is designed for efficient deployment, supporting INT4 quantization to enable inference on consumer-grade hardware.

Quick Start & Requirements

Health Check
Last commit

2 years ago

Responsiveness

Inactive

Pull Requests (30d)
0
Issues (30d)
0
Star History
27 stars in the last 90 days

Explore Similar Projects

Starred by Chip Huyen Chip Huyen(Author of AI Engineering, Designing Machine Learning Systems), Jared Palmer Jared Palmer(Ex-VP of AI at Vercel; Founder of Turborepo; Author of Formik, TSDX), and
1 more.

mpt-30B-inference by abacaj

0%
575
CPU inference code for MPT-30B
created 2 years ago
updated 2 years ago
Starred by Lysandre Debut Lysandre Debut(Chief Open-Source Officer at Hugging Face), Chip Huyen Chip Huyen(Author of AI Engineering, Designing Machine Learning Systems), and
1 more.

AQLM by Vahe1994

0.1%
1k
PyTorch code for LLM compression via Additive Quantization (AQLM)
created 1 year ago
updated 2 months ago
Starred by Chip Huyen Chip Huyen(Author of AI Engineering, Designing Machine Learning Systems), Georgios Konstantopoulos Georgios Konstantopoulos(CTO, General Partner at Paradigm), and
2 more.

GPTQ-for-LLaMa by qwopqwop200

0.0%
3k
4-bit quantization for LLaMA models using GPTQ
created 2 years ago
updated 1 year ago
Feedback? Help us improve.