lit-llama  by Lightning-AI

LLaMA implementation for pretraining, finetuning, and inference

Created 2 years ago
6,079 stars

Top 8.5% on SourcePulse

GitHubView on GitHub
Project Summary

Lit-LLaMA provides an independent, Apache 2.0 licensed implementation of the LLaMA language model, built on nanoGPT. It targets researchers and developers seeking to use, fine-tune, or pre-train LLaMA-compatible models without the GPL restrictions of the original Meta implementation, enabling broader integration and open-source AI development.

How It Works

This project offers a simplified, single-file implementation of LLaMA, prioritizing correctness and optimization for consumer hardware and scalable deployments. It leverages techniques like flash attention, INT8 and GPTQ 4-bit quantization for reduced memory footprint, and parameter-efficient fine-tuning methods such as LoRA and LLaMA-Adapter.

Quick Start & Requirements

Highlighted Details

  • Supports flash attention, INT8, and GPTQ 4-bit quantization.
  • Enables LoRA and LLaMA-Adapter fine-tuning.
  • Includes scripts for pre-training on datasets like RedPajama.
  • Designed for numerical equivalence to the original LLaMA model.

Maintenance & Community

  • Warning: This repository is no longer actively maintained. The successor project is LitGPT.
  • Community: Discord server available for engagement. https://discord.gg/VptPCZkGNa

Licensing & Compatibility

  • License: Apache 2.0.
  • Compatibility: Permissive license allows integration with closed-source projects.

Limitations & Caveats

This repository is explicitly marked as "Not Actively Maintained." Users are directed to the LitGPT project for updated features and support.

Health Check
Last Commit

2 months ago

Responsiveness

1 day

Pull Requests (30d)
0
Issues (30d)
0
Star History
6 stars in the last 30 days

Explore Similar Projects

Starred by Tobi Lutke Tobi Lutke(Cofounder of Shopify), Chip Huyen Chip Huyen(Author of "AI Engineering", "Designing Machine Learning Systems"), and
6 more.

xTuring by stochasticai

0.0%
3k
SDK for fine-tuning and customizing open-source LLMs
Created 2 years ago
Updated 1 day ago
Starred by Junyang Lin Junyang Lin(Core Maintainer at Alibaba Qwen), Vincent Weisser Vincent Weisser(Cofounder of Prime Intellect), and
25 more.

alpaca-lora by tloen

0.0%
19k
LoRA fine-tuning for LLaMA
Created 2 years ago
Updated 1 year ago
Feedback? Help us improve.