Discover and explore top open-source AI tools and projects—updated daily.
cvs-healthLLM bias and fairness assessment SDK
Top 99.9% on SourcePulse
<2-3 sentences summarising what the project addresses and solves, the target audience, and the benefit.> LangFair is a Python library designed for use-case-level bias and fairness assessments of Large Language Models (LLMs). It addresses the limitations of static benchmarks by enabling users to conduct evaluations tailored to their specific LLM applications through a "Bring Your Own Prompts" (BYOP) approach. This ensures that fairness metrics accurately reflect real-world performance and risks, focusing on practical, output-based measures suitable for governance audits. The library is targeted at engineers, researchers, and power users needing to rigorously evaluate LLM behavior.
How It Works
LangFair's core methodology centers on a "Bring Your Own Prompts" (BYOP) strategy, empowering users to define evaluation datasets specific to their LLM use cases. The framework prioritizes output-based metrics, which are practical for governance audits and real-world testing without requiring access to internal model states. Key functionalities include generating LLM responses, computing toxicity, stereotype associations, and counterfactual fairness metrics. For streamlined assessments in text generation and summarization, the AutoEval class automates multiple evaluation steps.
Quick Start & Requirements
pip install langfairChatVertexAI). A virtual environment is recommended. GPU support is optional for accelerating toxicity computation (torch.device("cuda")).Highlighted Details
AutoEval class provides a semi-automated evaluation for text generation and summarization use cases, consolidating toxicity, stereotype, and counterfactual metrics.Maintenance & Community
The project lists a development team comprising Dylan Bouchard, Mohit Singh Chauhan, David Skarbrevik, Viren Bajaj, and Zeya Ahmad. No external community channels (e.g., Discord, Slack) or roadmaps are explicitly linked in the provided text. Internal commit history is not made public.
Licensing & Compatibility
The README does not explicitly state the software's license. This omission prevents an immediate assessment of compatibility for commercial use or closed-source linking.
Limitations & Caveats
The AutoEval class is specifically highlighted for text generation and summarization use cases; users may need to manually apply individual metrics for other LLM applications. The absence of a stated license is a critical adoption blocker. Internal commit history is not public, potentially limiting transparency.
2 weeks ago
Inactive
braintrustdata
huggingface
microsoft
UKGovernmentBEIS
mshumer