Discover and explore top open-source AI tools and projects—updated daily.
Fine-tune LLM agents without LLM weight updates
Top 26.0% on SourcePulse
Memento addresses the challenge of continually improving Large Language Model (LLM) agents without the need for computationally expensive fine-tuning of the LLM weights themselves. It offers a memory-based, continual-learning framework that enables agents to learn from experience, making them more adaptable and efficient. The target audience includes researchers and developers working with LLM agents who need to enhance their performance over time through practical application. The primary benefit is achieving agent improvement with significantly reduced computational cost and complexity.
How It Works
Memento reframes continual learning as memory-based online reinforcement learning within a memory-augmented Markov Decision Process (MDP). It employs a case-based reasoning (CBR) approach where a neural case-selection policy guides actions. Experiences, represented as successful or failed trajectories, are stored in a "Case Bank" and efficiently retrieved for reuse. This memory-augmented learning allows the agent to steer its planning and execution based on past experiences, facilitating low-cost, transferable, and online continual learning. The architecture features a two-stage planner–executor loop: a Meta-Planner decomposes tasks and retrieves relevant cases, while an Executor runs subtasks using a unified MCP (Meta-Controller Protocol) interface, orchestrating various tools and logging outcomes.
Quick Start & Requirements
conda create -n Memento python=3.11 -y
, conda activate Memento
), navigate to the client
directory, create a .env
file for API keys, and install dependencies (pip install -r requirements.txt
, pip install -U crawl4ai crawl4ai-setup crawl4ai-doctor playwright install
)..env
file and potentially running a SearxNG Docker container.Highlighted Details
Maintenance & Community
The project acknowledges contributions from Camel-AI for some toolkits and interpreters. Information on specific maintainers, community channels (like Discord/Slack), or a public roadmap is not detailed in the provided README. Contributing guidelines are mentioned.
Licensing & Compatibility
The README does not explicitly state the license type. Compatibility for commercial use or closed-source linking is not specified.
Limitations & Caveats
Long-horizon tasks, particularly GAIA Level-3, remain challenging due to compounding errors. Performance on frontier knowledge tasks is limited by the current tooling. Open-source coverage for executor validation in fully open pipelines is limited.
1 week ago
Inactive