Benchmark for Chinese biomedical language understanding
Top 44.9% on sourcepulse
CBLUE provides a comprehensive benchmark for Chinese biomedical language understanding, targeting researchers and developers in the medical AI field. It offers datasets, baseline models, and an evaluation platform for eight distinct Natural Language Understanding (NLU) tasks, aiming to accelerate AI research and application in healthcare by establishing standardized evaluation metrics and facilitating model comparison.
How It Works
The benchmark comprises eight Chinese biomedical NLU tasks: Named Entity Recognition (NER), Relation Extraction (RE), Diagnosis Normalization, Sentence Classification (two tasks), Sentence Similarity, Natural Language Inference (two tasks). It supports various Chinese pre-trained language models (e.g., BERT, RoBERTa, ALBERT, MacBERT) and provides task-specific data processors, datasets, and trainers built on PyTorch and Huggingface Transformers.
Quick Start & Requirements
pip install -r requirements.txt
.examples/run_{task}.sh
) for training and prediction, or adapt the baseline scripts (baselines/run_classifier.py
, etc.).Highlighted Details
Maintenance & Community
The project is associated with the ACL 2022 conference. Further community engagement details (e.g., Discord/Slack) are not explicitly mentioned in the README.
Licensing & Compatibility
The repository's license is not explicitly stated in the README. Compatibility for commercial use or closed-source linking would require clarification of the license terms.
Limitations & Caveats
The README indicates that state-of-the-art neural models perform significantly worse than the human ceiling on these tasks, suggesting substantial room for improvement. Specific licensing terms for commercial use are not detailed.
2 years ago
1 day