LLaMA2-Accessory  by Alpha-VLLM

Open-source toolkit for LLM development, pretraining, finetuning, and deployment

created 2 years ago
2,786 stars

Top 17.5% on sourcepulse

GitHubView on GitHub
Project Summary

LLaMA2-Accessory is an open-source toolkit for developing, finetuning, and deploying large language models (LLMs) and multimodal LLMs (MLLMs). It extends the LLaMA-Adapter project with advanced features, including the SPHINX MLLM, which supports diverse training tasks, data domains, and visual embeddings, aiming to provide a comprehensive solution for LLM practitioners.

How It Works

The toolkit supports parameter-efficient finetuning methods like Zero-init Attention and Bias-norm Tuning, alongside distributed training strategies such as Fully Sharded Data Parallel (FSDP) and optimizations like Flash Attention 2 and QLoRA. It integrates various visual encoders (CLIP, Q-Former, ImageBind, DINOv2) and supports a wide range of LLMs including LLaMA, LLaMA2, CodeLlama, InternLM, Falcon, and Mixtral-8x7B. This modular design allows for flexible customization and efficient scaling of LLM development.

Quick Start & Requirements

  • Installation: Refer to the Environment Setup documentation.
  • Prerequisites: Python, PyTorch, Hugging Face libraries, and potentially CUDA for GPU acceleration. Specific model requirements may vary.
  • Documentation: Comprehensive guides for model pretraining, finetuning, and inference are available.

Highlighted Details

  • SPHINX-MoE achieves state-of-the-art performance on MMVP (49.33%) and AesBench.
  • Supports finetuning on a wide array of datasets for both single-modal and multi-modal tasks.
  • Includes efficient quantization with OmniQuant for reduced model size and faster inference.
  • Offers demos for various LLM applications, including chatbots and multimodal interactions.

Maintenance & Community

  • Active development with recent updates in early 2024, including releases for Large-DiT-T2I and SPHINX-Tiny.
  • The project is associated with the General Vision Group at Shanghai AI Lab, with hiring announcements for researchers.
  • Community engagement is encouraged via WeChat.

Licensing & Compatibility

  • The LLaMA 2 models are licensed under the LLAMA 2 Community License.
  • The toolkit itself appears to be open-source, but specific licensing for all components and datasets should be verified. Compatibility with commercial or closed-source applications depends on the underlying model licenses.

Limitations & Caveats

  • The project is heavily reliant on the LLaMA 2 Community License, which may have restrictions on commercial use.
  • While extensive, the breadth of supported models and datasets means specific configurations might require careful setup and dependency management.
Health Check
Last commit

6 months ago

Responsiveness

1 day

Pull Requests (30d)
0
Issues (30d)
0
Star History
24 stars in the last 90 days

Explore Similar Projects

Starred by Chip Huyen Chip Huyen(Author of AI Engineering, Designing Machine Learning Systems), Jeff Hammerbacher Jeff Hammerbacher(Cofounder of Cloudera), and
3 more.

LLaMA-Adapter by OpenGVLab

0.0%
6k
Efficient fine-tuning for instruction-following LLaMA models
created 2 years ago
updated 1 year ago
Feedback? Help us improve.