Instruction tuning code extends synthetic training
Top 80.3% on sourcepulse
This repository provides code and pretrained models for instruction tuning existing language models, specifically Flan-T5, using datasets like Alpaca and GPT4-Alpaca. It aims to make instruction-following capabilities more accessible and cost-effective, targeting researchers and developers working with large language models.
How It Works
The project leverages synthetic instruction data, generated by larger models like GPT-3, to fine-tune smaller, more accessible models such as Flan-T5. This approach allows for the transfer of instruction-following capabilities without the licensing constraints or computational demands of models like LLaMA. The code supports various data sources (Alpaca, GPT4-Alpaca, GPT4All, ShareGPT) and offers training scripts for different model sizes, including XL (3B) and XXL (11B) variants.
Quick Start & Requirements
conda create -n paca python=3.8 -y
, conda activate paca
, pip install -r requirements.txt
.alpaca_data.json
, alpaca_data_cleaned.json
, alpaca_gpt4_data.json
from releases.transformers
, torch
, pytorch-lightning
. Training requires at least one A6000 GPU (4x A6000 for XXL models with FSDP).Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
The project relies on synthetic data, which may contain noise. The README does not detail specific performance benchmarks against other instruction-tuned models beyond claims about Flacuna. Licensing for the code and models requires clarification for commercial applications.
2 years ago
Inactive