BioGPT  by microsoft

BioGPT is a generative pre-trained transformer for biomedical text

created 3 years ago
4,447 stars

Top 11.2% on sourcepulse

GitHubView on GitHub
Project Summary

BioGPT provides generative pre-trained transformer models specifically for biomedical text generation and mining. It is designed for researchers and developers working with biomedical data who need to perform tasks like relation extraction, question answering, and text generation within this domain. The models offer specialized capabilities for understanding and generating biomedical language.

How It Works

BioGPT is based on the Transformer architecture, leveraging a GPT-style generative model. It is pre-trained on a large corpus of biomedical literature, enabling it to capture domain-specific language patterns and knowledge. The implementation utilizes PyTorch and the fairseq library, with specific dependencies on older versions of fairseq (v0.12.0) and PyTorch (1.12.0), along with external tools like Moses and fastBPE for tokenization and BPE encoding.

Quick Start & Requirements

  • Installation: Requires cloning fairseq (v0.12.0), mosesdecoder, and fastBPE, setting environment variables MOSES and FASTBPE, and installing sacremoses and scikit-learn.
  • Prerequisites: Python 3.10, PyTorch 1.12.0, fairseq 0.12.0, CUDA (implied for GPU usage).
  • Models: Pre-trained and fine-tuned checkpoints are available via direct download or Hugging Face Hub.
  • Usage: Examples provided for direct PyTorch integration and Hugging Face pipeline usage.
  • Documentation: Links to Hugging Face documentation for further details.

Highlighted Details

  • Offers pre-trained BioGPT and BioGPT-Large models.
  • Provides fine-tuned checkpoints for specific tasks: Question Answering (PubMedQA), Relation Extraction (BC5CDR, DDI, KD-DTI), and Document Classification (HoC).
  • Integrated into Hugging Face transformers library for easier access and use.
  • Demos available on Hugging Face Spaces for Text Generation and Question Answering.

Maintenance & Community

  • Developed by Microsoft.
  • Follows Microsoft Open Source Code of Conduct.
  • Contributions welcome via a Contributor License Agreement (CLA).

Licensing & Compatibility

  • MIT License.
  • The license applies to pre-trained models as well, permitting commercial use and linking with closed-source projects.

Limitations & Caveats

The installation process requires specific, older versions of PyTorch (1.12.0) and fairseq (0.12.0), which may present compatibility challenges with newer environments. The setup involves manual cloning and compilation of several external dependencies.

Health Check
Last commit

1 year ago

Responsiveness

Inactive

Pull Requests (30d)
0
Issues (30d)
0
Star History
52 stars in the last 90 days

Explore Similar Projects

Feedback? Help us improve.