TensorFlow code for Chinese NER using BERT fine-tuning with BiLSTM-CRF
Top 10.5% on sourcepulse
This repository provides a TensorFlow implementation for Named Entity Recognition (NER) using a BERT-BiLSTM-CRF architecture. It's designed for researchers and developers working with Chinese NLP tasks, offering fine-tuning capabilities with Google's BERT models and a deployable service for NER and text classification.
How It Works
The core of the project combines BERT's contextual embeddings with a Bidirectional Long Short-Term Memory (BiLSTM) network and a Conditional Random Field (CRF) layer. BERT provides powerful pre-trained representations, which are then fed into the BiLSTM to capture sequential dependencies. The CRF layer models the dependencies between predicted labels, leading to more accurate NER tagging. This hybrid approach leverages the strengths of both pre-trained language models and sequence labeling architectures.
Quick Start & Requirements
pip install bert-base==0.0.9 -i https://pypi.org/simple
or clone and python3 setup.py install
.chinese_L-12_H-768_A-12.zip
). Training data from specified GitHub repos or custom formats.Highlighted Details
Maintenance & Community
The project is maintained by macanv. Community interaction is encouraged via GitHub issues.
Licensing & Compatibility
Limitations & Caveats
The project primarily targets Chinese NER, though adaptation for other languages is mentioned as requiring minor code changes. The README contains some outdated links and references to older versions. The service cannot run NER and text classification simultaneously; separate instances are required.
4 years ago
Inactive