Covering Scientific & Technical AI | Tuesday, December 24, 2024

MLCommons Science Working Group Invites Researchers to Run New Benchmarks 

June 1, 2023 -- The MLCommons Science Working Group encourages and supports the curation of large-scale experimental and scientific datasets and the engineering of machine learning (ML) benchmarks operating on those datasets. Tirelessly working to advance ML and artificial intelligence (AI) within the research community, the Working Group has recently released four testing benchmarks focused on  cloudmasking, geophysics, scanning transmission electron microscopes (STEMDL) and cancer distributed learning environments (CANDLE UNO).

“We are thrilled to invite researchers to run these benchmarks and report back to us on the systems that they used and how long it takes them to run,” said Christine Kirkpatrick, the director of Research Data Services at the San Diego Supercomputer Center at UC San Diego and an active participant in MLCommons. “For instance, we are really interested in better understanding how ML can be more efficient based on the choices made with hardware and software.”

Kirkpatrick leads several initiatives focused on findable, accessible, interoperable and reusable (FAIR) data and other research data management concepts. In this role, she looks for where data intersects with ML. Her most recent efforts involve the FAIR in ML, AI Readiness, AI Reproducibility (FARR) project, which has been funded by the National Science Foundation to promote better practices for AI, improve efficiency and reproducibility, and explore research gaps and priorities for data-centric AI.

“The MLCommons benchmarks provide a wealth of computer science data to mine with the ultimate goal of finding new efficiencies,” Kirkpatrick said. “This aligns perfectly with the efforts of the FARR project and we are eager to learn how applying information science techniques to these benchmarks makes the data more (re)usable.”

Geoffrey Fox, a computer science professor at the University of Virginia and co-founder of the Working Group noted, “These benchmarks represent useful tutorials on methods that can be used across many disciplines in AI for Science.”

To access the MLCommons Science Working Group benchmarks and to participate or learn more about the group and its work, visit the website.

About MLCommons

MLCommons is an open engineering consortium with a mission to make machine learning better for everyone through benchmarks and data. The foundation for MLCommons began with the MLPerf benchmark in 2018, which rapidly scaled as a set of industry metrics to measure machine learning performance and promote transparency of machine learning techniques. In collaboration with its 50+ founding partners – global technology providers, academics and researchers, MLCommons is focused on collaborative engineering work that builds tools for the entire machine learning industry through benchmarks and metrics, public datasets and best practices.


Source: Kimberly Mann Bruch, SDSC

AIwire