LLM unlearning framework for unifying evaluation benchmarks
Top 83.3% on sourcepulse
This repository provides a unified and extensible framework for evaluating Large Language Model (LLM) unlearning. It supports popular benchmarks like TOFU and MUSE, offering implementations for multiple unlearning methods, datasets, evaluation metrics, and LLM architectures, making it a valuable resource for researchers and practitioners in LLM privacy and security.
How It Works
Open-Unlearning leverages Hydra for flexible experiment configuration, allowing users to easily select and combine benchmarks, unlearning methods, datasets, and model architectures. It provides streamlined implementations of unlearning algorithms and evaluation metrics, facilitating reproducible research and benchmarking. The framework is designed for extensibility, encouraging community contributions of new components.
Quick Start & Requirements
conda create -n unlearning python=3.11
conda activate unlearning
pip install .
pip install --no-build-isolation flash-attn==2.6.3
python setup_data.py
to download evaluation log files.Highlighted Details
Maintenance & Community
contributing.md
.Licensing & Compatibility
Limitations & Caveats
The repository is a replacement for the no-longer-maintained locuslab/tofu
repository. Users must run setup_data.py
after merging the latest version to refresh evaluation log files.
1 week ago
1 day