Chinese ALBERT model for self-supervised learning
Top 12.5% on sourcepulse
This repository provides implementations and pre-trained models for ALBERT, a lighter BERT variant optimized for self-supervised learning of language representations, particularly for Chinese. It offers significantly reduced parameter counts while maintaining competitive accuracy across various NLP tasks, making it suitable for resource-constrained environments or applications requiring faster inference.
How It Works
ALBERT's efficiency stems from three core architectural changes over BERT: factorized embedding parameterization, cross-layer parameter sharing, and a sentence-order prediction (SOP) loss that focuses on coherence rather than topic prediction. These modifications drastically reduce model size and computational requirements. The project also explores removing dropout for increased model capacity and utilizes the LAMB optimizer for large batch training.
Quick Start & Requirements
bash run_classifier_clue.sh
for an end-to-end test.Highlighted Details
albert_tiny_zh
achieves 85.4% on LCQMC with 10x faster inference than BERT-base and ~60MB memory footprint when converted to TensorFlow Lite.Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
2 years ago
Inactive