Discover and explore top open-source AI tools and projects—updated daily.
alexziskind1LLM inference hardware calculator
Top 98.5% on SourcePulse
A web-based calculator estimates hardware requirements for Large Language Model (LLM) inference. It targets engineers and researchers needing to plan VRAM, system RAM, and GPU configurations, simplifying hardware procurement and deployment decisions.
How It Works
Built with React, TypeScript, and Vite, the tool employs a direct calculation approach. It takes user inputs for model size (parameters), quantization method (e.g., FP32, FP16, INT8, INT4), context length, and KV cache settings to derive hardware needs. The architecture supports estimations for both discrete GPU setups and unified memory systems, providing a clear overview of resource demands.
Quick Start & Requirements
npm install and npm run dev for development, or npm run build for production..env file from .env.example, set PORT, then run docker compose up -d --build.Highlighted Details
Maintenance & Community
No specific details on contributors, community channels, or roadmap are provided in the README.
Licensing & Compatibility
The project is released under the MIT License, which is highly permissive for commercial use and integration into closed-source projects.
Limitations & Caveats
All calculations are approximations and actual requirements may vary based on specific LLM implementations and runtime environments. The tool's discrete GPU estimations are based on a fixed 24GB VRAM card assumption.
5 days ago
Inactive
Mega4alik
ztxz16
AlexsJones