free-coding-models  by vava-nessa

CLI tool for finding fastest coding LLMs

Created 5 days ago

New!

314 stars

Top 86.2% on SourcePulse

GitHubView on GitHub
Project Summary

This project provides a real-time terminal-based tool to discover and select the fastest coding-focused Large Language Models (LLMs) from numerous providers. It targets developers and researchers seeking to optimize their AI coding assistants by offering live performance data, including latency and uptime, for over 100 models across nine different services. The primary benefit is enabling users to make informed decisions about which LLM offers the best performance for their specific coding tasks, directly integrating with popular tools like OpenCode and OpenClaw.

How It Works

The tool operates by pinging a comprehensive list of coding LLMs from providers such as NVIDIA NIM, Groq, Cerebras, and others, simultaneously in parallel. It utilizes a Text User Interface (TUI) that continuously monitors and re-pings these models every two seconds, displaying live "Latest," "Avg" (rolling average), and "Up%" (uptime) columns. This approach allows users to observe real-time performance differences and identify the most responsive models without manual testing or complex setup. It also features keyless latency testing, showing server reachability even without an API key.

Quick Start & Requirements

  • Installation: Install globally via npm, pnpm, or bun: npm i -g free-coding-models. Can also be run directly using npx free-coding-models.
  • Prerequisites: Node.js 18+ is required. Users need at least one free API key from supported providers like NVIDIA NIM, Groq, Cerebras, SambaNova, OpenRouter, Codestral, Hyperbolic, Scaleway, or Google AI. Optional installations of OpenCode or OpenClaw are supported for direct integration.
  • Links: NVIDIA NIM (build.nvidia.com), Groq (console.groq.com), Cerebras (cloud.cerebras.ai), SambaNova (cloud.sambanova.ai), OpenRouter (openrouter.ai), Mistral Codestral (codestral.mistral.ai), Hyperbolic (app.hyperbolic.ai), Scaleway (console.scaleway.com), Google AI Studio (aistudio.google.com).

Highlighted Details

  • Supports 101 coding-focused LLMs from 9 distinct providers.
  • Features real-time animation, smart ranking of top models, and continuous monitoring.
  • Provides uptime tracking, rolling averages, and an auto-retry mechanism for models.
  • Offers interactive selection and direct integration with OpenCode CLI/Desktop and OpenClaw.
  • Includes keyless latency testing to evaluate provider reachability before API key acquisition.
  • Models are ranked using SWE-bench verified scores.

Maintenance & Community

The project is actively maintained, with contributions welcomed via GitHub issues and pull requests. A Discord server is available for community discussion (https://discord.gg/5MbTnDC3Md). GitHub Actions automate publishing to npm.

Licensing & Compatibility

The project is released under the MIT license, permitting broad use, modification, and distribution, including for commercial purposes, with standard attribution requirements.

Limitations & Caveats

The tool is explicitly labeled as a BETA TUI and may contain bugs or experience crashes. Users should proceed with caution. Integration with OpenClaw requires a specific patching script (patch-openclaw.js) to fully populate the model allowlist due to OpenClaw's strict configuration. Free tiers from API providers have inherent usage limits.

Health Check
Last Commit

1 day ago

Responsiveness

Inactive

Pull Requests (30d)
10
Issues (30d)
10
Star History
322 stars in the last 5 days

Explore Similar Projects

Feedback? Help us improve.