Robuta

https://mlcommons.org/2025/04/ail-benchmarking-policy/
Apr 16, 2025 - The MLCommons AILuminate Benchmarking Policy outlines the policy for systems under test (SUTs) for inclusion in the AILuminate benchmark.
announcingailuminatebenchmarkingpolicy
https://mlcommons.org/ailuminate/
Oct 15, 2025 - The v1.0 AILuminate benchmark from the MLCommons AI Risk & Reliability working group is the first AI risk assessment benchmark developed with broad...
ailuminate
https://mlcommons.org/2025/04/ailuminate-french-datasets/
Apr 16, 2025 - MLCommons announces the release of two French language datasets for the AILuminate benchmark. A 1,200 prompt Creative-Commons licensed version, and 12,000...
releasesfrenchailuminatebenchmarkdemo
https://mlcommons.org/ailuminate/jailbreak/
Dec 9, 2025 - The v1.0 AILuminate benchmark from the MLCommons AI Risk & Reliability working group is the first AI risk assessment benchmark developed with broad...
ailuminatejailbreak
https://mlcommons.org/2025/05/nasscom/
Jun 26, 2025 - MLCommons is partnering with NASSCOM to bring AILuminate’s globally recognized AI reliability benchmarks to South Asia.
announcesexpansionindustryleadingailuminate
https://mlcommons.org/ailuminate/safety/
Dec 2, 2025 - The v1.0 AILuminate Safety benchmark from the MLCommons AI Risk & Reliability working group is the first AI risk assessment benchmark developed with broad...
ailuminatesafety