Published November 16, 2025
| Version v1.0.0
Report
Open
FairEval: Human-Aligned Evaluation Framework for Generative Models
Description
FairEval is an open-source evaluation framework designed to measure human alignment, bias, fairness and toxicity in generative models. It integrates rubric-based LLM-as-Judge scoring, category-wise toxicity detection, fairness auditing and human–model agreement metrics. The system provides dashboards for transparency and supports real-world model analysis across different tasks and demographic groups.
This technical report introduces the design, methodology and benchmarking results of FairEval, along with reproducible experiments and UI components.
Project GitHub Repository: https://github.com/kritibehl/FairEval
Files
FairEval__Human_Aligned_Evaluation_for_Generative_Models.pdf
Files
(493.5 kB)
| Name | Size | Download all |
|---|---|---|
|
md5:8352dae8d42f0059f30530f8b0cdf04e
|
493.5 kB | Preview Download |
Additional details
Related works
- Is supplement to
- Publication: https://medium.com/@kriti0608/faireval-a-human-aligned-evaluation-framework-for-generative-models-d822bfd5c99d (URL)
Software
- Repository URL
- https://github.com/kritibehl/FairEval
- Programming language
- Python
- Development Status
- Active