moonshot  by aiverify-foundation

LLM app evaluation and red-teaming tool

Created 1 year ago
280 stars

Top 92.9% on SourcePulse

GitHubView on GitHub
Project Summary

Moonshot is a modular tool designed for evaluating and red-teaming LLM applications, targeting AI developers, compliance teams, and system owners. It aims to simplify the complex processes of benchmarking LLM performance and identifying vulnerabilities through adversarial testing, offering a unified platform for comprehensive AI system assessment.

How It Works

Moonshot integrates benchmarking and red-teaming capabilities, allowing users to test LLMs against predefined competency metrics and probe for vulnerabilities using adversarial prompts. It supports various interfaces, including a web UI, CLI, and library APIs for MLOps integration. The tool utilizes "recipes" – collections of datasets and metrics – which can be curated or customized, and supports prompt templates and optional grading scales for standardized evaluation.

Quick Start & Requirements

  • Installation: pip install "aiverify-moonshot[all]" and python -m moonshot -i moonshot-data -i moonshot-ui.
  • Prerequisites: Python 3.11, Git, virtual environment recommended. For Web UI: Node.js 20.11.1 LTS+. Requires test assets from moonshot-data and moonshot-ui.
  • Running: Web UI: python -m moonshot web. CLI: python -m moonshot cli interactive.
  • Documentation: Installation Guide

Highlighted Details

  • Supports integration with popular LLM providers (OpenAI, Anthropic, HuggingFace) via API keys, with custom connector options.
  • Includes a range of benchmarks covering capability, quality, and trust & safety, incorporating community standards like BigBench and MLCommons safety benchmarks.
  • Features "cookbooks" for selecting relevant tests and allows custom recipe creation for unique use cases.
  • Offers automated red-teaming capabilities using research-backed attack modules to scale adversarial testing.

Maintenance & Community

Developed by the AI Verify Foundation. Links to user guides and documentation are provided.

Licensing & Compatibility

Licensed under the Apache Software License 2.0. This license is permissive and generally compatible with commercial and closed-source applications.

Limitations & Caveats

The project is at version 0.6.2, indicating potential for ongoing development and changes. While Python 3.11 is specified, compatibility with later releases is not guaranteed. The tool requires external data and UI components for full functionality.

Health Check
Last Commit

2 months ago

Responsiveness

1 day

Pull Requests (30d)
0
Issues (30d)
1
Star History
6 stars in the last 30 days

Explore Similar Projects

Starred by Shizhe Diao Shizhe Diao(Author of LMFlow; Research Scientist at NVIDIA), Pawel Garbacki Pawel Garbacki(Cofounder of Fireworks AI), and
3 more.

promptbench by microsoft

0.1%
3k
LLM evaluation framework
Created 2 years ago
Updated 3 weeks ago
Feedback? Help us improve.