SDK for LLM tuning and Sample Design Engineering (SDE)
Top 37.7% on sourcepulse
This repository provides tools and tutorials for efficiently fine-tuning Large Language Models (LLMs) using Sample Design Engineering (SDE). It targets developers and researchers aiming to improve LLM performance on downstream tasks with minimal data and computational resources. The core contribution is the SDE methodology, which empirically identifies effective sample design strategies for fine-tuning.
How It Works
The project introduces Sample Design Engineering (SDE) as a systematic approach to optimize fine-tuning datasets. It explores various sample design strategies, uncovering patterns consistent across different LLMs. The ES-SDE approach integrates the most effective options, demonstrating superiority over baseline methods in empirical studies. This method focuses on the quality and structure of training samples rather than solely on model architecture or training algorithms.
Quick Start & Requirements
pip install transformers datasets accelerate sentencepiece tensorboard peft
tokenize.sh
, followed by training via train.sh
. Specific Python scripts are used for different models (e.g., chatglm_lora_tuning.py
, baichuan_lora_tuning.py
).Highlighted Details
Maintenance & Community
peft
library and references the ChatGLM-Tuning
and LLaMA-Efficient-Tuning
projects.Licensing & Compatibility
Limitations & Caveats
1 year ago
Inactive