Chat model trained via LoRA, using ChatGPT-generated dialogs
Top 15.6% on sourcepulse
Baize-chatbot provides open-source chat models fine-tuned on ChatGPT-generated self-chat data, targeting researchers and developers seeking to deploy custom conversational AI. It offers parameter-efficient fine-tuning (PEFT) with LoRA, enabling training and inference on a single GPU, significantly reducing resource requirements compared to full model fine-tuning.
How It Works
Baize leverages LoRA for parameter-efficient fine-tuning of LLaMA base models. It uses 100k dialogs generated by ChatGPT conversing with itself, augmented with Alpaca's dataset. This approach allows for rapid training and deployment of capable chatbots with substantially less VRAM and time than traditional fine-tuning methods.
Quick Start & Requirements
pip install git+https://github.com/lm-sys/FastChat.git
python3 -m fastchat.model.apply_lora --base huggyllama/llama-7b --target ./model_weights/baize-7b --lora project-baize/baize-lora-7B
python -m fastchat.serve.cli --model-path ./model_weights/baize-7b
Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
Commercial use of model weights and data is strictly prohibited due to licensing. The project relies on access to LLaMA base models, which have their own usage restrictions. Training is recommended on A100-80G GPUs, though smaller GPUs can be used with reduced batch sizes.
1 year ago
Inactive