ML library for distributed training, model serving, and federated learning
Top 12.7% on sourcepulse
FedML is a unified, scalable machine learning library designed for distributed training, model serving, and federated learning across diverse hardware environments. It targets developers and researchers needing to run AI jobs efficiently on any GPU cloud or on-premise cluster, with TensorOpera AI offering a complementary platform for generative AI and LLMs.
How It Works
FedML provides a unified MLOps layer with Studio for accessing and fine-tuning foundational models, and a Job Store for pre-built AI tasks. Its scheduler, TensorOpera Launch, optimizes GPU resource allocation and automates job execution across various compute topologies. The compute layer includes platforms for scalable model serving (Deploy), large-scale distributed training (Train), and federated learning (Federate), leveraging FedML's core library for cross-device and cross-cloud operations.
Quick Start & Requirements
pip install fedml
Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
The project is heavily integrated with the TensorOpera AI platform, suggesting potential vendor lock-in or a focus on their ecosystem for advanced features. The README mentions "world’s first FLOps" which may indicate early-stage or experimental features within the federated learning component.
3 weeks ago
Inactive