Pre-trained weights for biomedical text mining
Top 50.0% on sourcepulse
This repository provides pre-trained weights for BioBERT, a BERT-based language representation model specifically tailored for biomedical text mining tasks. It offers researchers and practitioners a powerful tool for applications like named entity recognition, relation extraction, and question answering within the biomedical domain, leveraging extensive biomedical corpora for enhanced performance.
How It Works
BioBERT is built upon Google's original BERT architecture, utilizing a WordPiece vocabulary derived from BERT-base-Cased. This approach allows for effective representation of novel biomedical terms through subword tokenization. The model has been pre-trained on large biomedical text datasets, including PubMed abstracts and PubMed Central full texts, resulting in specialized language understanding capabilities for the biomedical field.
Quick Start & Requirements
Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
5 years ago
1 day