LLaMA implementation for pretraining, finetuning, and inference
Top 8.6% on sourcepulse
Lit-LLaMA provides an independent, Apache 2.0 licensed implementation of the LLaMA language model, built on nanoGPT. It targets researchers and developers seeking to use, fine-tune, or pre-train LLaMA-compatible models without the GPL restrictions of the original Meta implementation, enabling broader integration and open-source AI development.
How It Works
This project offers a simplified, single-file implementation of LLaMA, prioritizing correctness and optimization for consumer hardware and scalable deployments. It leverages techniques like flash attention, INT8 and GPTQ 4-bit quantization for reduced memory footprint, and parameter-efficient fine-tuning methods such as LoRA and LLaMA-Adapter.
Quick Start & Requirements
pip install -e ".[all]"
Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
This repository is explicitly marked as "Not Actively Maintained." Users are directed to the LitGPT project for updated features and support.
1 month ago
Inactive