Local PDF chatbot for document interaction
Top 24.7% on sourcepulse
This project provides a fully local, client-side chat-over-documents solution, targeting users who want to query PDFs without uploading data to external servers. It leverages WebAssembly and browser-based LLM inference for privacy and offline functionality.
How It Works
The application processes uploaded PDFs entirely within the browser. It chunks the document, creates vector embeddings using Transformers.js (or optionally Ollama), and stores them in a WASM-based vector store (Voy). Retrieval-Augmented Generation (RAG) is then performed using LangChain.js and LangGraph.js, interacting with a locally hosted LLM.
Quick Start & Requirements
OLLAMA_ORIGINS=https://webml-demo.vercel.app OLLAMA_HOST=127.0.0.1:11435 ollama serve
OLLAMA_HOST=127.0.0.1:11435 ollama pull mistral
Highlighted Details
Maintenance & Community
The project acknowledges contributors from Voy, Ollama, WebLLM, and Transformers.js. The author is active on Twitter (@Hacubu).
Licensing & Compatibility
The repository does not explicitly state a license in the README. Compatibility for commercial use or closed-source linking is not specified.
Limitations & Caveats
The Gemini Nano integration is experimental and may yield variable results as the model is not chat-tuned. The project is a Next.js app, and deployment details beyond local setup are not provided.
4 months ago
1 day