Discover and explore top open-source AI tools and projects—updated daily.
GAIR-NLPGenerative judge for evaluating LLM alignment
Top 100.0% on SourcePulse
Generative Judge for Evaluating Alignment (Auto-J) is an open-source tool designed to evaluate Large Language Models (LLMs) on their alignment with human preferences. It addresses the need for reliable, interpretable LLM assessment by providing a generative judge capable of handling diverse real-world scenarios. Auto-J is beneficial for researchers and developers seeking to benchmark LLM performance and identify areas for improvement in alignment.
How It Works
Auto-J operates as a generative model trained on a broad dataset of real-world user queries and LLM responses across 58 scenarios. It offers flexibility by supporting both pairwise response comparison (determining which of two responses is superior) and single-response evaluation (providing critiques and ratings). The core innovation lies in its ability to generate detailed, natural language critiques, which enhance the transparency and reliability of the evaluation process and facilitate human involvement.
Quick Start & Requirements
pip install -r requirements.txt. Creating a virtual environment (e.g., with conda) is recommended.torch>=2.0.1+cu118) is necessary, and GPU(s) are essential for running the models.Highlighted Details
autoj-bilingual-6b) is available for multilingual evaluation.autoj-13b-GPTQ-4bits) is provided, reducing VRAM requirements to approximately 8GB.Maintenance & Community
The project acknowledges computing resource support from Shanghai AI Lab and contributions for the bilingual version and human annotation. It builds upon the PKU-Alignment/safe-rlhf and vllm-project/vllm projects. No explicit community channels (e.g., Discord, Slack) or a public roadmap are detailed in the provided information.
Licensing & Compatibility
Auto-J (13B) and Auto-J-Scenario-Classifier (13B) models are licensed under Llama 2.Auto-J-Bilingual (6B) model is released under the Yi License.Limitations & Caveats
The 4-bit quantized version may exhibit behavioral differences compared to the original model. The bilingual version has known issues, including occasional code-switching and limitations in mathematical and coding capabilities. Model deployment requires specific tensor_parallel_size configurations (e.g., 1, 2, 4, 8) due to the underlying vLLM implementation.
2 years ago
Inactive
braintrustdata
lmarena