Llama-X  by AetherCortex

Open academic research project improving LLaMA to SOTA LLM

created 2 years ago
1,619 stars

Top 26.5% on sourcepulse

GitHubView on GitHub
Project Summary

Llama-X is an open academic research project aiming to systematically improve the LLaMA model to state-of-the-art performance through community collaboration. It targets researchers and developers interested in advancing LLM capabilities by providing transparent access to code, data, model weights, and experimental details, fostering rapid, iterative improvements.

How It Works

The project follows a versioned improvement strategy, with each new model release expected to demonstrate significant performance gains over its predecessor on a defined set of benchmarks. It focuses on ten key research areas, including instruction tuning, RLHF, data quality, long context, multi-modal, multilingual capabilities, infrastructure efficiency, evaluation, interpretability, and LLM actions. The approach emphasizes full fine-tuning rather than parameter-efficient methods like LoRA.

Quick Start & Requirements

  • Install: Conda environment setup with PyTorch 1.12.0, Transformers 4.31.0, and CUDA 11.3.
  • Prerequisites: Python 3.10, Conda, DeepSpeed, Hugging Face Transformers, and LLaMA model weights.
  • Setup: Requires downloading LLaMA weights and converting them to Hugging Face format. Training commands are provided for single and multi-node setups using DeepSpeed Zero-3.
  • Demo: Available via a demo page.

Highlighted Details

  • Claims to achieve comparable performance to gpt-35-turbo with version 3.5.0 and is closing the gap with GPT-4.
  • Reports efficient training: 7B model fine-tuned in 1 hour on 8x V100 GPUs with a batch size of 64.
  • Focuses on full fine-tuning, not LoRA.
  • Comprehensive evaluation plan includes Type-A (MMLU, HumanEval, GSM-8K, etc.) and Type-B benchmarks.

Maintenance & Community

  • Actively seeking contributors for code, data, papers, and computing resources.
  • Communication channels include GitHub Issues and email (llama-x@mail.com).
  • Inspired by Meta AI LLaMA, Huggingface Transformers, and Llama Alpaca.

Licensing & Compatibility

  • Resources are limited to academic research and prohibited for commercial purposes.
  • No explicit license is stated, but the commercial use restriction is a significant caveat.

Limitations & Caveats

The project explicitly states that the use of its resources is limited to academic research and prohibits commercial use. The accuracy of model output is not guaranteed due to inherent LLM randomness.

Health Check
Last commit

1 year ago

Responsiveness

1 day

Pull Requests (30d)
0
Issues (30d)
1
Star History
2 stars in the last 90 days

Explore Similar Projects

Starred by Chip Huyen Chip Huyen(Author of AI Engineering, Designing Machine Learning Systems), Elie Bursztein Elie Bursztein(Cybersecurity Lead at Google DeepMind), and
10 more.

open_llama by openlm-research

0.0%
8k
Open-source reproduction of LLaMA models
created 2 years ago
updated 2 years ago
Starred by Chip Huyen Chip Huyen(Author of AI Engineering, Designing Machine Learning Systems), Woosuk Kwon Woosuk Kwon(Author of vLLM), and
11 more.

WizardLM by nlpxucan

0.1%
9k
LLMs built using Evol-Instruct for complex instruction following
created 2 years ago
updated 1 month ago
Starred by George Hotz George Hotz(Author of tinygrad; Founder of the tiny corp, comma.ai), Calvin French-Owen Calvin French-Owen(Coounder of Segment), and
12 more.

StableLM by Stability-AI

0.0%
16k
Language models by Stability AI
created 2 years ago
updated 1 year ago
Starred by Chip Huyen Chip Huyen(Author of AI Engineering, Designing Machine Learning Systems), Ying Sheng Ying Sheng(Author of SGLang), and
9 more.

alpaca-lora by tloen

0.0%
19k
LoRA fine-tuning for LLaMA
created 2 years ago
updated 1 year ago
Starred by Andrej Karpathy Andrej Karpathy(Founder of Eureka Labs; Formerly at Tesla, OpenAI; Author of CS 231n), John Yang John Yang(Author of SWE-bench, SWE-agent), and
13 more.

stanford_alpaca by tatsu-lab

0.1%
30k
Instruction-following LLaMA model training and data generation
created 2 years ago
updated 1 year ago
Feedback? Help us improve.