self-llm  by datawhalechina

LLM guide for Chinese users on Linux

Created 1 year ago
24,409 stars

Top 1.6% on SourcePulse

GitHubView on GitHub
Project Summary

This repository provides a comprehensive tutorial for Chinese beginners on deploying and fine-tuning open-source Large Language Models (LLMs) and Multimodal Large Language Models (MLLMs) within a Linux environment. It aims to simplify the process of using and applying these models, making them more accessible to students and researchers.

How It Works

The tutorial covers the entire lifecycle of working with open-source LLMs, from initial environment configuration tailored to specific model requirements, to deploying and using popular models like LLaMA, ChatGLM, and InternLM. It also details various fine-tuning techniques, including full parameter fine-tuning and efficient methods like LoRA and P-tuning, enabling users to customize models for their specific needs.

Quick Start & Requirements

  • Installation: Primarily involves environment setup and model downloads via Hugging Face, ModelScope, or git-lfs. Specific commands depend on the chosen model and deployment method (e.g., vLLM, FastApi, LMStudio, Ollama).
  • Prerequisites: Linux environment, Python, potentially CUDA for GPU acceleration, and specific model dependencies.
  • Resources: Requires significant disk space for models and potentially powerful GPUs for efficient fine-tuning and inference.
  • Documentation: Comprehensive guides are available within the repository.

Highlighted Details

  • Supports a wide array of popular LLMs including Qwen, Kimi, Llama, Gemma, DeepSeek, and more.
  • Provides detailed tutorials for various deployment methods like FastApi, vLLM, and web demos.
  • Includes practical examples and case studies, such as creating a "Zhen Huan" chatbot or a math-focused LLM.
  • Offers guidance on integrating LLMs with frameworks like LangChain.

Maintenance & Community

The project is actively maintained by Datawhale members and contributors, with a clear structure for issues and pull requests. Contact information is provided for deeper involvement.

Licensing & Compatibility

The repository itself appears to be open-source, but the licensing of the individual models covered varies. Users should verify the license of each model they intend to use, especially for commercial applications.

Limitations & Caveats

The tutorial is primarily focused on Linux environments, and setup on other operating systems might require adaptation. While it covers many models, the rapid pace of LLM development means new models may not be immediately included.

Health Check
Last Commit

2 weeks ago

Responsiveness

1 day

Pull Requests (30d)
6
Issues (30d)
5
Star History
1,143 stars in the last 30 days

Explore Similar Projects

Starred by Andrej Karpathy Andrej Karpathy(Founder of Eureka Labs; Formerly at Tesla, OpenAI; Author of CS 231n), Gabriel Almeida Gabriel Almeida(Cofounder of Langflow), and
2 more.

torchchat by pytorch

0.1%
4k
PyTorch-native SDK for local LLM inference across diverse platforms
Created 1 year ago
Updated 1 week ago
Starred by Casper Hansen Casper Hansen(Author of AutoAWQ), Yineng Zhang Yineng Zhang(Inference Lead at SGLang; Research Scientist at Together AI), and
5 more.

xtuner by InternLM

0.5%
5k
LLM fine-tuning toolkit for research
Created 2 years ago
Updated 1 day ago
Starred by Andrej Karpathy Andrej Karpathy(Founder of Eureka Labs; Formerly at Tesla, OpenAI; Author of CS 231n), Stefan van der Walt Stefan van der Walt(Core Contributor to scientific Python ecosystem), and
12 more.

litgpt by Lightning-AI

0.1%
13k
LLM SDK for pretraining, finetuning, and deploying 20+ high-performance LLMs
Created 2 years ago
Updated 6 days ago
Feedback? Help us improve.