LLM pretraining and fine-tuning for medical dialogue
Top 95.3% on sourcepulse
This repository provides a framework for pre-training and fine-tuning Large Language Models (LLMs), specifically demonstrating a medical dialogue model. It targets researchers and developers working with LLMs in specialized domains, offering a structured approach to leverage DeepSpeed for efficient training on limited hardware.
How It Works
The project utilizes DeepSpeed with Zero-3, CPU offload, and FP16 for memory optimization, enabling the training of large models on multi-GPU setups (e.g., 8x A6000). It outlines data processing strategies for both pre-training and fine-tuning, including text concatenation, slicing, and special token usage, referencing established practices from models like ChatHome and Llama 2. The fine-tuning process includes masking input labels for supervised learning and mentions experimental support for LoRA.
Quick Start & Requirements
Model_Bloom_Pretrain.py
), fine-tuning (Model_Bloom_Sft.py
), model conversion, inference, and API serving.Highlighted Details
Maintenance & Community
The project was updated on 20230926, announcing the open-sourcing of a WiNGPT model by Weining Health. Specific community links or active maintenance signals are not detailed in the README.
Licensing & Compatibility
The README does not explicitly state a license. Compatibility for commercial use or closed-source linking is not specified.
Limitations & Caveats
The README notes that LoRA fine-tuning currently only supports single-GPU setups and may have issues with multi-GPU configurations. It also highlights that many existing open-source medical LLMs may not meet the demands of practical applications.
1 year ago
Inactive