A Leaderboard to Benchmark Ethical Biases in LLMs
Gomez-Vazquez M., Morales S., Castignani G., Clarisó R., Conrardy A., Deladiennee L., Renault S., Cabot J.
CEUR Workshop Proceedings, vol. 3744, 2024
This paper introduces a public leaderboard that comprehensively assesses and benchmarks Large Language Models (LLMs) according to a set of ethical biases and test metrics. The initiative aims to raise awareness about the status of the latest advances in development of ethical AI, and foster its alignment to recent regulations in order to guardrail its societal impacts.