Inference server for machine learning models
Top 43.7% on sourcepulse
MLServer is an open-source inference server designed for deploying machine learning models via REST and gRPC interfaces. It targets ML engineers and data scientists seeking a standardized, scalable, and framework-agnostic solution for model deployment, offering compatibility with the KFServing V2 Dataplane spec.
How It Works
MLServer utilizes an "inference runtime" system, acting as a bridge between the server and various ML frameworks. This modular design allows for easy integration of new frameworks and custom model serving logic. It supports multi-model serving within a single process, parallel inference across worker pools, and adaptive batching for request optimization, all while adhering to the V2 Inference Protocol for broad compatibility.
Quick Start & Requirements
pip install mlserver
pip install mlserver-<framework>
(e.g., mlserver-sklearn
)Highlighted Details
Maintenance & Community
./hack/update-version.sh
.make test
, tox -e py3 -- <test_file>
.Licensing & Compatibility
Limitations & Caveats
Python versions 3.7, 3.13, and earlier are unsupported. The README notes that associated software may have different licensing terms, requiring careful review for commercial use.
5 days ago
1 day