chatllama  by henrywoo

Open-source implementation for LLaMA-based ChatGPT, runnable on a single GPU

created 2 years ago
1,204 stars

Top 33.2% on sourcepulse

GitHubView on GitHub
Project Summary

ChatLLaMA provides an open-source implementation for fine-tuning Meta's LLaMA models into ChatGPT-like conversational agents using Reinforcement Learning from Human Feedback (RLHF). It targets researchers and developers aiming to build cost-effective, single-GPU deployable chatbots with faster training than original ChatGPT.

How It Works

ChatLLaMA implements the RLHF training pipeline for LLaMA models. It leverages DeepSpeed ZERO for efficient, distributed fine-tuning, enabling faster training on smaller hardware. The approach supports all LLaMA model sizes (7B to 65B), allowing users to balance training time and inference performance.

Quick Start & Requirements

  • Install via pip: pip install chatllama
  • Requires Meta's LLaMA model weights (apply via Meta's form).
  • Requires a custom dataset or generation via provided scripts.
  • Supports all LLaMA architectures (7B, 13B, 33B, 65B).
  • Official documentation and examples are available.

Highlighted Details

  • Claims a 15x faster training process compared to ChatGPT.
  • Enables single-GPU inference for LLaMA models.
  • Built-in support for DeepSpeed ZERO for accelerated fine-tuning.
  • Compatible with all LLaMA model sizes.

Maintenance & Community

  • Project appears to be a personal or small-team effort.
  • No explicit links to community channels (Discord, Slack) or roadmaps are provided in the README.

Licensing & Compatibility

  • The README does not explicitly state a license for the chatllama library itself.
  • Compatibility with Meta's LLaMA model weights is subject to Meta's terms of use.

Limitations & Caveats

The repository does not include model weights, requiring users to obtain them separately from Meta. The README implies a focus on the algorithmic implementation of RLHF rather than a fully packaged, ready-to-deploy solution.

Health Check
Last commit

6 months ago

Responsiveness

1 day

Pull Requests (30d)
0
Issues (30d)
0
Star History
2 stars in the last 90 days

Explore Similar Projects

Starred by Chip Huyen Chip Huyen(Author of AI Engineering, Designing Machine Learning Systems), Jeff Hammerbacher Jeff Hammerbacher(Cofounder of Cloudera), and
3 more.

LLaMA-Adapter by OpenGVLab

0.0%
6k
Efficient fine-tuning for instruction-following LLaMA models
created 2 years ago
updated 1 year ago
Starred by Andrej Karpathy Andrej Karpathy(Founder of Eureka Labs; Formerly at Tesla, OpenAI; Author of CS 231n), Georgios Konstantopoulos Georgios Konstantopoulos(CTO, General Partner at Paradigm), and
2 more.

lit-llama by Lightning-AI

0.1%
6k
LLaMA implementation for pretraining, finetuning, and inference
created 2 years ago
updated 1 month ago
Starred by Chip Huyen Chip Huyen(Author of AI Engineering, Designing Machine Learning Systems), Ying Sheng Ying Sheng(Author of SGLang), and
9 more.

alpaca-lora by tloen

0.0%
19k
LoRA fine-tuning for LLaMA
created 2 years ago
updated 1 year ago
Feedback? Help us improve.