Inference library for Mistral models
Top 4.9% on sourcepulse
This repository provides the official inference library for Mistral AI's large language models, enabling users to run and interact with models like Mistral 7B, Mixtral 8x7B, and Codestral. It's designed for researchers and developers who need direct control over model execution and integration into custom applications.
How It Works
The library offers a reference implementation for running Mistral models, leveraging PyTorch for efficient computation. It supports various model architectures and features like function calling and multimodal capabilities. The core design prioritizes minimal dependencies for straightforward integration, while also providing options for multi-GPU inference and deployment via vLLM.
Quick Start & Requirements
pip install mistral-inference
xformers
dependency.Highlighted Details
mistral-demo
, mistral-chat
) for easy testing and interaction.Maintenance & Community
Licensing & Compatibility
codestral-22B-v0.1.tar
and mistral-large-instruct-2407.tar
are subject to custom non-commercial licenses (MNPL and MRL respectively).Limitations & Caveats
4 months ago
Inactive