MLCommons
An open AI engineering consortium that builds industry-standard benchmarks and datasets to measure and improve AI accuracy, safety, speed, and efficiency.
At a Glance
Pricing
Free access to benchmarks, datasets, and research resources
Engagement
Available On
About MLCommons
MLCommons is an open AI engineering consortium that brings together industry leaders, academics, and researchers to build trusted, safe, and efficient AI systems. The organization develops industry-standard benchmarks and open datasets that measure quality, performance, and risk in machine learning systems, helping companies and universities worldwide build better AI that benefits society.
-
MLPerf Benchmarks provide neutral, consistent measurements of AI system accuracy, speed, and efficiency across training, inference, storage, and specialized domains like automotive, mobile, and tiny ML applications.
-
AILuminate offers comprehensive AI safety evaluation tools including safety benchmarks, jailbreak testing, and agentic AI assessment methodologies to help developers build more reliable AI systems.
-
Open Datasets include People's Speech, Multilingual Spoken Words, Dollar Street, and other large-scale, diverse datasets that improve AI model training and evaluation.
-
Croissant Metadata Standard serves as today's standard vocabulary for ML datasets, making machine learning work easier to reproduce and replicate across the research community.
-
AI Risk & Reliability Working Group brings together a global consortium of AI industry leaders, practitioners, researchers, and civil society experts committed to building a harmonized approach for safer AI.
-
Collaborative Research supports scientific advancement through shared infrastructure and diverse community participation, enabling new breakthroughs in AI through working groups focused on algorithms, data-centric ML, and scientific applications.
To get started with MLCommons, organizations can join as members or affiliates to participate in working groups, contribute to benchmark development, access datasets, and collaborate on research initiatives. The consortium operates on principles of open collaboration, consensus-driven decision-making, and inclusive participation from startups, large companies, academics, and non-profits globally.
Community Discussions
Be the first to start a conversation about MLCommons
Share your experience with MLCommons, ask questions, or help others learn from your insights.
Pricing
Open Source
Free access to benchmarks, datasets, and research resources
- Access to MLPerf benchmark results
- Open datasets including People's Speech and Multilingual Spoken Words
- Croissant metadata standard
- Research publications and documentation
- Community participation
Capabilities
Key Features
- MLPerf Training benchmarks
- MLPerf Inference benchmarks
- MLPerf Storage benchmarks
- MLPerf Automotive benchmarks
- MLPerf Mobile benchmarks
- MLPerf Tiny benchmarks
- MLPerf Client benchmarks
- AILuminate safety benchmarks
- AILuminate jailbreak testing
- AILuminate agentic AI evaluation
- Croissant metadata standard
- Open ML datasets
- AlgoPerf training algorithms benchmark
- AI Risk & Reliability working group
- Medical AI working group
- MLCube containerization
