CLI tool for simplifying LLMs and ML workflows on AWS SageMaker
Top 69.0% on sourcepulse
Sagify simplifies the deployment and management of Large Language Models (LLMs) and machine learning workflows on AWS SageMaker. It offers a unified interface, the LLM Gateway, to interact with both proprietary (OpenAI) and open-source models for tasks like chat completions, image generation, and embeddings, abstracting away infrastructure complexities for ML engineers and researchers.
How It Works
Sagify utilizes a modular architecture centered around an LLM Gateway, a FastAPI application that provides a consistent API for various LLM providers. It supports deploying open-source models directly to AWS SageMaker endpoints, allowing for fine-grained control over instance types and scaling. Alternatively, it can interface with OpenAI's API for proprietary models. The gateway acts as a central hub, routing requests to the appropriate backend service, whether it's a SageMaker endpoint or the OpenAI API.
Quick Start & Requirements
pip install sagify
sh sagify cloud foundation-model-deploy --model-id model-txt2img-stabilityai-stable-diffusion-v2-1-base --model-version 1.* -n 1 -e ml.p3.2xlarge --aws-region us-east-1 --aws-profile sagemaker-dev
sagify llm gateway --image sagify-llm-gateway:v0.1.0 --start-local
) or deployed to AWS Fargate.Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
3 days ago
1 day