Discover and explore top open-source AI tools and projects—updated daily.
lucasjinrealHigh-performance Rust inference engine
Top 93.1% on SourcePulse
Crane is a high-performance, pure Rust inference engine for LLMs, VLMs, TTS, and OCR, built on the Candle framework. It targets developers seeking a simpler, faster alternative to C++-based solutions like llama.cpp, offering significant speedups and ease of deployment, especially on Apple Silicon.
How It Works
Crane leverages Rust's Candle framework to achieve blazing-fast inference speeds on both CPUs and GPUs. Its core design emphasizes eliminating C++ complexity while maintaining native performance, enabling hardware-agnostic execution across CPU, CUDA, and Metal (for Apple Silicon). This approach simplifies model integration, allowing new models to be added with minimal code.
Quick Start & Requirements
cargo run --bin qwenchat --release (after downloading models).Highlighted Details
2 days ago
Inactive
trymirai
baidu-research
NVIDIA
openvinotoolkit