In-browser chatbot for local, private LLM inference
Top 18.3% on sourcepulse
This project provides a fully private, in-browser Large Language Model (LLM) chatbot, enabling users to interact with models like Llama 3 and Mistral without any server-side components. It targets users seeking privacy and offline LLM capabilities, offering a ChatGPT-like interface directly within their web browser.
How It Works
The chatbot leverages the WebLLM inference engine, which allows LLMs to run directly in the browser using WebGPU. This approach eliminates the need for server infrastructure, ensuring all conversation data remains on the user's local machine for maximum privacy. It supports quantized models, reducing memory footprint and enabling larger models to run efficiently within browser constraints.
Quick Start & Requirements
yarn install
, then yarn dev
.Highlighted Details
Maintenance & Community
The project is seeking contributors for interface improvements, model support, and bug fixes. A Discord server is available for community interaction.
Licensing & Compatibility
The repository's license is not explicitly stated in the README. Compatibility for commercial use or closed-source linking is not specified.
Limitations & Caveats
WebGPU support is required, which may necessitate manual configuration in some browsers like Firefox and Safari. Performance and model availability are dependent on the user's hardware and browser capabilities.
1 year ago
1 day