beyondllm  by aiplanethub

RAG toolkit for LLM app building, evaluation, and observation

Created 1 year ago
288 stars

Top 91.2% on SourcePulse

GitHubView on GitHub
1 Expert Loves This Project
Project Summary

BeyondLLM is an open-source toolkit designed for building, evaluating, and observing Retrieval-Augmented Generation (RAG) applications. It targets developers and researchers working with Large Language Models (LLMs), aiming to simplify RAG system development, reduce hallucinations, and enhance reliability through automated integration and customizable evaluation.

How It Works

BeyondLLM streamlines RAG pipelines by abstracting complex data ingestion, retrieval, and generation steps. It supports various data sources (like YouTube videos) and allows integration with different LLMs and embedding models (e.g., OpenAI). The framework includes built-in evaluation metrics for both embeddings (hit rate, MRR) and LLM responses (context relevancy, answer relevancy, groundness), providing a comprehensive approach to RAG system quality assurance. An observability feature monitors LLM performance, including latency and cost for supported models.

Quick Start & Requirements

  • Primary install: pip install beyondllm
  • Prerequisites: OpenAI API key (for custom LLM/embedding examples), Google API key (for YouTube data source example).
  • Documentation: beyondllm.aiplanet.com
  • Demo: Available on Google Colab.

Highlighted Details

  • Build RAG applications in as few as 5 lines of code.
  • Supports custom LLMs and embedding models.
  • Includes automated evaluation for retrieval and generation components.
  • Observability feature for monitoring LLM performance (latency, cost).

Maintenance & Community

  • Active community engagement encouraged via Discord.
  • Contributions are welcomed for features, infrastructure, and documentation.
  • Acknowledgements include HuggingFace, LlamaIndex, OpenAI, and Google Gemini.

Licensing & Compatibility

  • Licensed under the Apache License, version 2.0.
  • Permissive license suitable for commercial use and integration into closed-source projects.

Limitations & Caveats

The observability feature currently only monitors OpenAI LLM models. While the project supports various LLMs and embeddings, specific integrations might require custom implementation or may not be fully optimized.

Health Check
Last Commit

7 months ago

Responsiveness

Inactive

Pull Requests (30d)
0
Issues (30d)
0
Star History
1 stars in the last 30 days

Explore Similar Projects

Starred by Morgan Funtowicz Morgan Funtowicz(Head of ML Optimizations at Hugging Face), Luis Capelo Luis Capelo(Cofounder of Lightning AI), and
7 more.

lighteval by huggingface

2.6%
2k
LLM evaluation toolkit for multiple backends
Created 1 year ago
Updated 1 day ago
Starred by Pawel Garbacki Pawel Garbacki(Cofounder of Fireworks AI), Shizhe Diao Shizhe Diao(Author of LMFlow; Research Scientist at NVIDIA), and
14 more.

SWE-bench by SWE-bench

2.3%
4k
Benchmark for evaluating LLMs on real-world GitHub issues
Created 1 year ago
Updated 17 hours ago
Starred by Luis Capelo Luis Capelo(Cofounder of Lightning AI), Chip Huyen Chip Huyen(Author of "AI Engineering", "Designing Machine Learning Systems"), and
6 more.

opik by comet-ml

1.7%
14k
Open-source LLM evaluation framework for RAG, agents, and more
Created 2 years ago
Updated 11 hours ago
Starred by Chip Huyen Chip Huyen(Author of "AI Engineering", "Designing Machine Learning Systems"), Elvis Saravia Elvis Saravia(Founder of DAIR.AI), and
2 more.

awesome-llm-apps by Shubhamsaboo

2.6%
69k
LLM app collection with AI agents and RAG examples
Created 1 year ago
Updated 3 days ago
Feedback? Help us improve.