Search portal using LLMs and RAG for comprehensive question answering
Top 50.7% on sourcepulse
YaCy Expert aims to create a question-answering search engine by combining Large Language Models (LLMs) with Retrieval Augmented Generation (RAG). It targets users with large text corpora, such as those acquired through YaCy web crawling, enabling them to build domain-specific "expert" chatbots. The project leverages LLMs for semantic understanding and RAG for context delivery, moving beyond traditional keyword-based search.
How It Works
The system architecture consists of a web interface acting as a wrapper for two backend services: an LLM (inference engine) and a RAG system (knowledge base). The LLM is designed as a drop-in replacement for the OpenAI chat API, utilizing llama.cpp
. The RAG system is embedded within the YaCy Expert web interface's backend and uses Faiss for efficient vector similarity search on user-provided data dumps. This approach allows for semantically relevant context retrieval from large text corpora to augment LLM responses.
Quick Start & Requirements
.jsonlist
), placing it in the knowledge
directory, and running python3 knowledge_indexing.py .
(approx. 1 hour per 10,000 entries). Custom BERT models can be specified via .ini
files.llama.cpp
(Docker recommended).Highlighted Details
llama.cpp
for OpenAI API compatibility, enabling self-hosted LLMs.Maintenance & Community
yacy
organization.Licensing & Compatibility
yacy
project is typically licensed under GPLv2.Limitations & Caveats
6 months ago
Inactive