ChatLLM-Web  by Ryan-yang125

Browser-based chat app for local LLM inference

created 2 years ago
634 stars

Top 53.3% on sourcepulse

GitHubView on GitHub
Project Summary

This project provides a web-based interface for interacting with large language models (LLMs) like Vicuna directly within the browser, leveraging WebGPU for acceleration. It targets users seeking a private, serverless, and easily deployable solution for LLM chat, offering a seamless user experience with features like local data storage and PWA support.

How It Works

The application runs entirely client-side, utilizing WebGPU for hardware-accelerated LLM inference. The LLM model operates within a web worker, preventing UI blocking and ensuring a smooth user experience. Model weights are cached locally after an initial download, enabling faster subsequent loads.

Quick Start & Requirements

  • Install: npm install && npm run dev
  • Prerequisites: Browser with WebGPU support (Chrome 113+), GPU with ~6.4GB VRAM recommended.
  • Initial Model Download: ~4GB.
  • More details: mlc.ai/web-llm

Highlighted Details

  • Runs entirely in the browser with WebGPU acceleration.
  • Web worker for non-blocking UI.
  • One-click Vercel deployment.
  • Local model caching.
  • Multi-conversation support with local data storage.
  • Markdown and streaming response support (math, code highlighting).
  • PWA support for offline use.

Maintenance & Community

The project is actively developed by Ryan-yang125. Further community or roadmap details are not explicitly provided in the README.

Licensing & Compatibility

  • License: MIT
  • Compatibility: Suitable for commercial use and integration with closed-source applications due to the permissive MIT license.

Limitations & Caveats

Requires a modern browser with WebGPU support and a GPU with sufficient VRAM for optimal performance; performance may degrade on lower-spec hardware. The README indicates support for Vicuna-7b and RedPajama-INCITE-Chat-3B, but only Vicuna-7b is marked as fully supported.

Health Check
Last commit

1 year ago

Responsiveness

1 week

Pull Requests (30d)
0
Issues (30d)
0
Star History
3 stars in the last 90 days

Explore Similar Projects

Starred by Chip Huyen Chip Huyen(Author of AI Engineering, Designing Machine Learning Systems), Simon Willison Simon Willison(Author of Django), and
1 more.

Lumos by andrewnguonly

0.1%
2k
Chrome extension for local LLM web RAG co-piloting
created 1 year ago
updated 6 months ago
Feedback? Help us improve.