BentoDiffusion  by bentoml

Diffusion model deployment examples using BentoML

created 2 years ago
373 stars

Top 77.1% on sourcepulse

GitHubView on GitHub
Project Summary

BentoDiffusion provides a streamlined approach to self-hosting and deploying Stable Diffusion models, specifically targeting developers and researchers who need to integrate advanced image generation capabilities into their applications. It simplifies the process of serving models like SDXL Turbo, enabling rapid prototyping and scalable deployment.

How It Works

This project leverages BentoML to package and serve diffusion models. BentoML's framework handles model serialization, dependency management, and API endpoint creation, abstracting away much of the complexity typically associated with deploying large machine learning models. This allows users to focus on the diffusion models themselves rather than the intricacies of serving infrastructure.

Quick Start & Requirements

  • Install: pip install -r requirements.txt within the model's subdirectory (e.g., sdxl-turbo).
  • Prerequisites: Nvidia GPU with at least 12GB VRAM recommended. Python 3.11.
  • Run: bentoml serve in the project directory.
  • Docs: Stable Diffusion XL Turbo

Highlighted Details

  • Demonstrates serving various Stable Diffusion models, including SDXL Turbo, FLUX.1, Stable Diffusion 3 Medium, and ControlNet.
  • Offers integration with BentoCloud for managed deployment and scalability.
  • Generates OCI-compliant images for custom infrastructure deployments.
  • Provides example client code (cURL and Python) for interacting with the served models.

Maintenance & Community

This repository is part of the BentoML ecosystem. Further community and support information can be found via BentoML's official channels.

Licensing & Compatibility

The specific license for this repository is not explicitly stated in the provided README. Compatibility for commercial use or closed-source linking would depend on the licenses of the underlying diffusion models and BentoML itself.

Limitations & Caveats

The README strongly recommends an Nvidia GPU with 12GB VRAM, indicating potential performance issues or inability to run on hardware with less VRAM. The project focuses on specific diffusion models, and support for others may require manual adaptation.

Health Check
Last commit

3 months ago

Responsiveness

1 week

Pull Requests (30d)
0
Issues (30d)
1
Star History
13 stars in the last 90 days

Explore Similar Projects

Starred by Chip Huyen Chip Huyen(Author of AI Engineering, Designing Machine Learning Systems), Alexey Milovidov Alexey Milovidov(Cofounder of Clickhouse), and
7 more.

OpenLLM by bentoml

0.2%
12k
SDK for running open-source LLMs as OpenAI-compatible APIs
created 2 years ago
updated 3 days ago
Feedback? Help us improve.