Code for extractive summarization via fine-tuned BERT
Top 28.2% on sourcepulse
This repository provides the code for fine-tuning BERT for extractive summarization, targeting researchers and practitioners in Natural Language Processing. It offers improved ROUGE scores over baseline models on the CNN/Dailymail dataset by integrating BERT with different decoder architectures.
How It Works
BertSum leverages BERT's contextual embeddings to identify salient sentences for extractive summarization. It explores three encoder-decoder configurations: a simple classifier, a Transformer, and an RNN. The BERT+Transformer variant, in particular, achieves state-of-the-art results by using BERT for sentence encoding and a Transformer for sequence decoding, allowing for effective capture of long-range dependencies.
Quick Start & Requirements
Highlighted Details
Maintenance & Community
The project is associated with a research paper, indicating a focus on academic contributions. No specific community channels or active maintenance signals are explicitly mentioned in the README.
Licensing & Compatibility
The repository's license is not explicitly stated in the provided README. Compatibility for commercial use or closed-source linking would require clarification of the licensing terms.
Limitations & Caveats
The data preparation process is complex and requires external tools like Stanford CoreNLP. The README does not specify the exact BERT model used (e.g., base, large) or provide pre-trained models for direct use, necessitating custom training.
3 years ago
Inactive