genv  by run-ai

GPU environment/cluster manager with LLM support

Created 3 years ago
639 stars

Top 52.0% on SourcePulse

GitHubView on GitHub
Project Summary

Genv is an open-source system for managing GPU environments and clusters, designed to simplify GPU resource allocation and sharing for data scientists and ML engineers. It allows users to easily control, configure, monitor, and enforce GPU usage across machines or clusters, enabling efficient collaboration and resource utilization, particularly for LLM development and deployment.

How It Works

Genv operates by creating isolated GPU environments, inspired by tools like pyenv and Conda. Users can activate specific environments with defined GPU counts and memory allocations, abstracting away the underlying hardware. This approach allows seamless switching between GPU resources without modifying code, facilitating fair resource distribution, quota enforcement, and efficient sharing of GPUs for tasks like serving local LLMs.

Quick Start & Requirements

  • Install: pip install genv or conda install -c conda-forge genv
  • Prerequisites: NVIDIA GPU with compatible drivers. CUDA 11.4 is shown in the example, but specific version requirements are not explicitly stated beyond driver compatibility.
  • Setup: Installation is straightforward via package managers. Activating an environment involves commands like genv activate --name my-env --gpus 1.
  • Docs: Genv documentation site

Highlighted Details

  • Facilitates sharing GPUs among teammates and pooling resources from multiple machines.
  • Enforces GPU quotas (count and memory) for equitable resource allocation.
  • Integrates with Ollama for managing and serving local LLMs on cluster GPUs.
  • Offers monitoring capabilities via Grafana dashboards for administrators.

Maintenance & Community

  • Developed by Run.ai Labs.
  • Community support and feature discussion available on their Discord server.

Licensing & Compatibility

  • Licensed under AGPLv3. Run.ai intends for AGPLv3 obligations to be interpreted broadly, particularly regarding "work based on the Program" and "Corresponding Source."
  • The broad interpretation of AGPLv3 terms may impose significant obligations on derivative works and linked code, potentially restricting commercial use or integration into closed-source projects.

Limitations & Caveats

The AGPLv3 license, with its broad interpretation clause, presents a significant consideration for commercial adoption or integration into proprietary software due to potential copyleft requirements. Specific CUDA version compatibility beyond the example is not detailed.

Health Check
Last Commit

1 year ago

Responsiveness

Inactive

Pull Requests (30d)
0
Issues (30d)
0
Star History
7 stars in the last 30 days

Explore Similar Projects

Starred by Chip Huyen Chip Huyen(Author of "AI Engineering", "Designing Machine Learning Systems"), Johannes Hagemann Johannes Hagemann(Cofounder of Prime Intellect), and
4 more.

S-LoRA by S-LoRA

0.2%
2k
System for scalable LoRA adapter serving
Created 1 year ago
Updated 1 year ago
Starred by Andrej Karpathy Andrej Karpathy(Founder of Eureka Labs; Formerly at Tesla, OpenAI; Author of CS 231n), Jason Knight Jason Knight(Director AI Compilers at NVIDIA; Cofounder of OctoML), and
3 more.

gpu.cpp by AnswerDotAI

0%
4k
C++ library for portable GPU computation using WebGPU
Created 1 year ago
Updated 2 months ago
Starred by Chip Huyen Chip Huyen(Author of "AI Engineering", "Designing Machine Learning Systems") and Ying Sheng Ying Sheng(Coauthor of SGLang).

fastllm by ztxz16

0.4%
4k
High-performance C++ LLM inference library
Created 2 years ago
Updated 1 week ago
Feedback? Help us improve.