BERT fine-tuning for named entity recognition
Top 32.0% on sourcepulse
This repository provides a fine-tuned BERT model for Named Entity Recognition (NER) on the CoNLL-2003 dataset. It's designed for researchers and practitioners looking to implement and experiment with BERT for sequence labeling tasks, offering a cleaner, updated version of an earlier implementation.
How It Works
The project leverages Google's BERT architecture for NER, fine-tuning it on the CoNLL-2003 dataset. The approach involves data preprocessing and custom layer design, with specific suggestions to modify the CRF or softmax layers for potential performance improvements. The use of a cased BERT model is recommended for better accuracy, aligning with findings from Google's research.
Quick Start & Requirements
bash run_ner.sh
.conlleval.pl
script for evaluation.Highlighted Details
crf_layer
or softmax_layer
for further tuning.Maintenance & Community
The repository appears to be a personal project with no explicit mention of active maintenance, community channels, or notable contributors beyond the author.
Licensing & Compatibility
The README does not explicitly state a license. The project relies on Google's BERT, which is typically Apache 2.0 licensed. Compatibility for commercial use would depend on the licensing of the BERT model weights and the CoNLL-2003 dataset.
Limitations & Caveats
The project relies on external downloads for BERT model weights and the dataset. Performance might require further tuning beyond the default parameters, as suggested by the author. The README does not specify Python version requirements or explicit dependency management beyond the core BERT components.
3 years ago
Inactive