Human-Aligned-LLM-Evaluation-Audit
PublicA data-driven audit of AI judge reliability using MT-Bench human annotations. This project analyzes 3,500+ model comparisons across 6 LLMs and 8 task categories to measure how well GPT-4 evaluations align with human judgment. Includes Python workflow, disagreement metrics, and a Power BI dashboard for insights.