Discover and explore top open-source AI tools and projects—updated daily.
Pretrained ELECTRA model for Korean language tasks
Top 53.0% on SourcePulse
KoELECTRA provides pretrained ELECTRA models specifically for the Korean language, offering improved performance over BERT-like models by leveraging the Replaced Token Detection pre-training task. It is designed for researchers and developers working with Korean NLP tasks, enabling more effective text understanding and generation.
How It Works
KoELECTRA utilizes the ELECTRA architecture, which trains a discriminator model to distinguish between original and replaced tokens generated by a smaller generator model. This approach allows for learning from all input tokens, leading to greater efficiency and performance. The models are trained on a substantial Korean corpus (34GB) and are compatible with the Hugging Face Transformers library.
Quick Start & Requirements
monologg/koelectra-base-v3-discriminator
).Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
tf_model.h5
was removed due to issues, reverting to from_pt=True
loading.1 year ago
1 day