Chinese UniLM base model for NLU and NLG tasks
Top 69.1% on sourcepulse
This repository provides a Chinese pre-trained UniLM model, a versatile architecture capable of both Natural Language Understanding (NLU) and Natural Language Generation (NLG) tasks. It's intended for researchers and developers working with Chinese NLP, offering a strong baseline for tasks like text classification, reading comprehension, and summarization.
How It Works
UniLM is a unified pre-training framework that allows a single model to perform both NLU and NLG tasks by conditioning the self-attention mechanism. For NLU tasks, it functions similarly to BERT. For NLG tasks, it's fine-tuned using a sequence-to-sequence approach, enabling generation. This dual capability reduces the need for separate models for understanding and generation.
Quick Start & Requirements
Highlighted Details
unilm_base
.bert_base
in ROUGE scores.type_token_id
guidance for NLU ([0,1]) and NLG ([4,5]) tasks.Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
The project relies on older versions of PyTorch (1.4.0) and Transformers (2.6.0), which may pose compatibility issues with current libraries. The lack of explicit licensing and community support could be a barrier to adoption.
3 years ago
Inactive