sebastian  by celaraze

LLM assistant with permanent memory

created 4 years ago
909 stars

Top 40.8% on sourcepulse

GitHubView on GitHub
Project Summary

Sebastian is an LLM assistant designed to overcome the context limitations of traditional chatbots by providing permanent memory. It targets users who need a conversational AI that can recall past interactions and evolving user preferences across extended periods, offering a more personalized and context-aware experience akin to a digital Jarvis.

How It Works

Sebastian employs Retrieval Augmented Generation (RAG) and embeddings to manage memory, rather than relying on context window tokens. This approach allows it to store and retrieve information indefinitely, enabling features like automatic memory discovery, iteration (updating stored information), association (linking related facts), and consolidation (periodically reviewing and optimizing stored knowledge).

Quick Start & Requirements

  • Deployment: Docker Compose is recommended.
  • Prerequisites: Docker and Docker Compose installed.
  • LLM Inference: Requires API keys for either Qwen2-Max (via AliCloud) or GPT-4o (via OpenAI). Local Ollama support is planned.
  • Setup: Edit docker-compose.yml to set LANGUAGE, DASHSCOPE_API_KEY (for Chinese) or OPENAI_API_KEY (for English), and a TOKEN. Run docker compose up -d.
  • API Interaction: Use curl commands to interact via text or audio endpoints.
  • Documentation: Official Docker website for installation.

Highlighted Details

  • Permanent memory breaks traditional LLM context limitations.
  • Automatic memory discovery, iteration, association, and consolidation.
  • Supports private deployment for data security.
  • Handles both text and voice conversations.

Maintenance & Community

The project welcomes contributions via issues and pull requests. Sponsorship is available via Afdian.net. Acknowledgments are given to Gitee AI, Alibaba Cloud, and JetBrains.

Licensing & Compatibility

The README does not explicitly state a license. Compatibility for commercial use or closed-source linking is not specified.

Limitations & Caveats

Currently relies on cloud-based LLM inference services (Qwen2-Max, GPT-4o), requiring API keys and incurring potential costs. Local Ollama deployment is planned but not yet available. The specific license is not mentioned, which may impact commercial adoption.

Health Check
Last commit

5 months ago

Responsiveness

Inactive

Pull Requests (30d)
0
Issues (30d)
0
Star History
4 stars in the last 90 days

Explore Similar Projects

Starred by Chip Huyen Chip Huyen(Author of AI Engineering, Designing Machine Learning Systems).

LangBot by langbot-app

0.8%
13k
IM bot platform for the LLM era
created 2 years ago
updated 6 hours ago
Starred by Andrej Karpathy Andrej Karpathy(Founder of Eureka Labs; Formerly at Tesla, OpenAI; Author of CS 231n), Alex Cheema Alex Cheema(Cofounder of EXO Labs), and
3 more.

Perplexica by ItzCrazyKns

0.3%
23k
AI-powered search engine alternative
created 1 year ago
updated 1 day ago
Feedback? Help us improve.