Chinese ELECTRA pre-trained language models
Top 29.2% on sourcepulse
This repository provides pre-trained ELECTRA models for the Chinese language, offering a more efficient alternative to BERT for various NLP tasks. It targets researchers and developers working with Chinese NLP, enabling them to leverage ELECTRA's smaller model size and strong performance.
How It Works
ELECTRA utilizes a novel pre-training approach involving a small generator network that replaces tokens in input text, and a larger discriminator network trained to detect these replaced tokens. This "Replaced Token Detection" (RTD) task is more sample-efficient than BERT's Masked Language Model (MLM), leading to better performance with fewer computational resources. The project focuses solely on the discriminator for downstream fine-tuning.
Quick Start & Requirements
AutoTokenizer.from_pretrained(MODEL_NAME)
) or PaddleHub (hub.Module(name=MODULE_NAME)
).Highlighted Details
Maintenance & Community
The project is from the Harbin Institute of Technology (HIT) & iFlytek Joint Lab (HFL). Recent activity includes the release of Chinese LLaMA/Alpaca models. Users are encouraged to check the FAQ before submitting issues.
Licensing & Compatibility
The repository does not explicitly state a license. The models are available for download via Google Drive and Baidu Netdisk. Compatibility with commercial or closed-source projects is not specified.
Limitations & Caveats
2 weeks ago
1 day