LLM answer engine for Perplexity-style search
Top 10.3% on sourcepulse
This project provides a Perplexity-inspired answer engine, targeting developers interested in NLP and search. It enables users to build a sophisticated application that returns answers, sources, images, videos, and follow-up questions based on user queries, leveraging a combination of LLMs and search APIs.
How It Works
The engine integrates multiple LLMs (Groq, Mistral, OpenAI) and search services (Brave, Serper) within a Next.js framework. It utilizes Langchain.js for text processing and OpenAI embeddings for semantic search. User queries are processed to retrieve relevant information from search engines, which is then fed to an LLM for answer generation. The architecture supports streaming responses and optional features like rate limiting and semantic caching via Upstash Redis.
Quick Start & Requirements
git clone https://github.com/developersdigest/llm-answer-engine.git
followed by npm install
or bun install
(non-Docker) or docker compose up -d
(Docker)..env
and app/config.tsx
.Highlighted Details
Maintenance & Community
developersdigest
.Licensing & Compatibility
Limitations & Caveats
Ollama support for follow-up questions is not yet implemented, and using local Ollama models for inference and embeddings can result in long time-to-first-token. Function calling is currently in beta.
1 month ago
1 day