Chrome extension for local LLM web RAG co-piloting
Top 27.9% on sourcepulse
Lumos is a Chrome extension that acts as a Retrieval-Augmented Generation (RAG) co-pilot for web browsing, leveraging local Large Language Models (LLMs) via Ollama. It enables users to summarize web content, ask questions about articles or documentation, and interact with LLMs directly on their machine without relying on remote servers, offering enhanced privacy and control.
How It Works
Lumos integrates with a local Ollama server for LLM inference and embedding. It fetches web page content, chunks it, and uses an embedding model to create vector representations for efficient retrieval. When a user queries Lumos, it retrieves relevant chunks from the vector store and feeds them, along with the query, to the local LLM to generate a contextually aware response. This RAG approach allows LLMs to access and process information from the web.
Quick Start & Requirements
ollama pull llama2
OLLAMA_ORIGINS=chrome-extension://* ollama serve
dist
folder in the project directory or download a pre-built package from the Releases page.Highlighted Details
Maintenance & Community
The project is inspired by Web LLM and LangChain examples. Further details on community or maintenance are not explicitly provided in the README.
Licensing & Compatibility
The README does not specify a license. Compatibility for commercial use or closed-source linking is not detailed.
Limitations & Caveats
Lumos requires a local Ollama server to be running for functionality. While it aims for local-only inference, it relies on a local server for LLM operations due to Chrome extension security constraints. Content highlighted by the user is not cached in the vector store, requiring re-embedding for each subsequent prompt.
6 months ago
Inactive