Discover and explore top open-source AI tools and projects—updated daily.
RayFernando1337LLM RAM calculator for inference optimization
Top 99.3% on SourcePulse
Instantly calculate the maximum size of quantized language models that can fit in your available RAM, helping you optimize models for inference. This interactive React + TypeScript + Vite application targets engineers and researchers needing to quickly estimate LLM deployment feasibility based on hardware constraints.
How It Works
This project utilizes a modern frontend stack: React for an interactive user interface, TypeScript for type safety, and Vite for rapid development with Hot Module Replacement (HMR). The core calculation converts available RAM and estimated OS overhead from gigabytes to bytes. It then subtracts memory required for the specified context window size and converts the chosen quantization level (bits per parameter) into bytes per parameter. This allows for a precise estimation of the maximum number of model parameters that can fit within the remaining RAM.
Quick Start & Requirements
bun installbun run devhttp://localhost:5173 in your browser.bun run build (output in dist directory).Highlighted Details
Maintenance & Community
Contributions are welcome via Pull Requests. The README does not specify maintainers, sponsorships, or community channels like Discord or Slack.
Licensing & Compatibility
Licensed under the MIT License, which permits commercial use and integration into closed-source projects.
Limitations & Caveats
The tool provides an estimation based on user-provided inputs for RAM, OS usage, and context window. Actual model performance and memory footprint may vary due to specific LLM architectures, runtime efficiencies, and other system-level factors not included in this calculation. This is a frontend calculator, not an inference engine.
2 weeks ago
Inactive
Mega4alik
lyogavin