AI infrastructure SDK for serverless workloads
Top 36.1% on sourcepulse
This project provides a Python-based framework for deploying and managing AI workloads across distributed infrastructure, including bare-metal servers and cloud GPUs. It targets developers seeking a simplified, serverless experience for AI inference and computation, offering autoscaling, secure communication, and fleet management.
How It Works
Beta9 utilizes Python decorators (@endpoint
, @function
, @task_queue
, @schedule
) to define and deploy AI workloads. These workloads are containerized and exposed as load-balanced HTTP endpoints. The system manages distributed clusters, allowing users to connect arbitrary GPU-enabled machines via a simple agent installation script. It leverages a Tailscale-powered service mesh for secure, encrypted communication (WireGuard) and provides a centralized control plane for fleet management.
Quick Start & Requirements
make setup
(requires Docker, k3d).make setup-sdk
to install.dnsmasq
for local domain resolution.Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
The project is in "beta9," suggesting potential instability or breaking changes. The lack of a specified license poses a significant adoption blocker for commercial use.
1 day ago
1 day