Discover and explore top open-source AI tools and projects—updated daily.
Chinese summary generation model based on GPT2
Top 71.5% on SourcePulse
This project provides a Chinese text summarization model based on the GPT-2 architecture. It is designed for researchers and developers working with Chinese NLP tasks who need a pre-trained model for generating concise summaries from longer texts. The benefit is a readily available, fine-tuned GPT-2 model for Chinese summarization.
How It Works
The model leverages the GPT-2 architecture, specifically fine-tuning the GPT2-Chinese and GPT2-chitchat models. It processes training data by concatenating text and its summary, then trains the model on this sequence. This approach aims to adapt the generative capabilities of GPT-2 for the specific task of Chinese text summarization.
Quick Start & Requirements
pip install transformers==2.1.1 pytorch==1.3.1
summary_model
folder and run interact.py
.Highlighted Details
GPT2-nlpcc-summary
and GPT2-wiki
.Maintenance & Community
The project appears to be a personal or academic endeavor with no explicit mention of active maintenance, community channels (like Discord/Slack), or a roadmap. It acknowledges contributions from the GPT2-Chinese and GPT2-chitchat projects.
Licensing & Compatibility
The README does not explicitly state a license. Given its reliance on GPT2-Chinese and GPT2-chitchat, users should verify the licenses of those underlying projects for compatibility, especially for commercial use.
Limitations & Caveats
The project is presented as an exploration and may not be suitable for production without retraining the general and summary models. The README notes that the summarization quality might be better in vertical domains than general news. The sampling-based decoding means generated summaries can vary.
2 years ago
Inactive