Python SDK for LLM response evaluation
Top 91.7% on sourcepulse
Athina-evals provides a Python SDK for evaluating Large Language Model (LLM) responses, offering over 50 preset evaluations and support for custom ones. It's designed for AI teams focused on observability and experimentation, serving as a companion to the Athina IDE for prototyping, running experiments, and comparing datasets.
How It Works
The SDK allows programmatic execution of evaluations, with results visualized and managed within the Athina IDE. This integrated approach facilitates side-by-side dataset comparison and experiment tracking, streamlining the LLM development lifecycle.
Quick Start & Requirements
pip install athina-evals
CodeExecutionV2
evaluations, install e2b-code-interpreter
.Highlighted Details
Maintenance & Community
No specific contributor or community details are provided in the README.
Licensing & Compatibility
The README does not specify a license.
Limitations & Caveats
The README does not detail any limitations or caveats.
1 month ago
Inactive