PyTorch toolkit for pre-training and fine-tuning NLP models
Top 15.9% on sourcepulse
UER-py is an open-source PyTorch framework for pre-training and fine-tuning Natural Language Processing (NLP) models. It targets researchers and practitioners seeking to leverage or extend universal encoder representations, offering modularity and a model zoo for various NLP tasks. The framework aims to reproduce state-of-the-art (SOTA) results and facilitate custom model development.
How It Works
UER-py employs a modular architecture, separating models into components like embeddings, encoders, and targets. This design allows users to combine different modules to construct custom pre-training models with flexibility. It supports various pre-training objectives and architectures, including BERT, GPT-2, ELMo, and T5, and facilitates distributed training across multiple GPUs.
Quick Start & Requirements
pip install -r requirements.txt
Highlighted Details
Maintenance & Community
The project is associated with Tencent and has multiple academic and industry contributors. A refactored, newer version, TencentPretrain, is available and supports multi-modal models and larger model training.
Licensing & Compatibility
The README does not explicitly state a license. Compatibility for commercial use or closed-source linking is not specified.
Limitations & Caveats
The project recommends TencentPretrain for multi-modal or large model training, suggesting UER-py is primarily suited for text models under one billion parameters. The README does not specify a license, which could impact commercial adoption.
1 year ago
Inactive