trlX is a distributed training framework for fine-tuning large language models using Reinforcement Learning from Human Feedback (RLHF). It supports models up to 20B parameters via Hugging Face Accelerate and larger models using NVIDIA NeMo, offering PPO and ILQL algorithms.
How It Works
trlX leverages Hugging Face Accelerate for efficient distributed training of models up to 20B parameters, and NVIDIA NeMo for scaling beyond that. It supports training with either a custom reward function or a reward-labeled dataset, abstracting the complexities of RL algorithms like PPO and ILQL.
Quick Start & Requirements
pip install -e .
after cloning the repository.pip install torch --extra-index-url https://download.pytorch.org/whl/cu118
).accelerate launch
.Highlighted Details
Maintenance & Community
trl
library.Licensing & Compatibility
Limitations & Caveats
1 year ago
1 day