Doorgaan naar hoofdnavigatie Doorgaan naar zoeken Ga verder naar hoofdinhoud

DataPerf: Benchmarks for Data-Centric AI Development

  • Mark Mazumder
  • , Colby Banbury
  • , Xiaozhe Yao
  • , Bojan Karlaš
  • , William Gaviria Rojas
  • , Sudnya Diamos
  • , Greg Diamos
  • , Lynn He
  • , Alicia Parrish
  • , Hannah Rose Kirk
  • , Jessica Quaye
  • , Charvi Rastogi
  • , Douwe Kiela
  • , David Jurado
  • , David Kanter
  • , Rafael Mosquera
  • , Juan Ciro
  • , Lora Aroyo
  • , Bilge Acun
  • , Lingjiao Chen
  • Mehul Smriti Raje, Max Bartolo, Sabri Eyuboglu, Amirata Ghorbani, Emmett Goodman, Oana Inel, Tariq Kane, Christine R. Kirkpatrick, Tzu Sheng Kuo, Jonas Mueller, Tristan Thrush, Joaquin Vanschoren, Margaret Warren, Adina Williams, Serena Yeung, Newsha Ardalani, Praveen Paritosh, Ce Zhang, James Zou, Carole Jean Wu, Cody Coleman, Andrew Ng, Peter Mattson, Vijay Janapa Reddi

Onderzoeksoutput: Hoofdstuk in Boek/Rapport/CongresprocedureConferentiebijdrageAcademicpeer review

35 Downloads (Pure)

Samenvatting

Machine learning research has long focused on models rather than datasets, and prominent datasets are used for common ML tasks without regard to the breadth, difficulty, and faithfulness of the underlying problems. Neglecting the fundamental importance of data has given rise to inaccuracy, bias, and fragility in real-world applications, and research is hindered by saturation across existing dataset benchmarks. In response, we present DataPerf, a community-led benchmark suite for evaluating ML datasets and data-centric algorithms. We aim to foster innovation in data-centric AI through competition, comparability, and reproducibility. We enable the ML community to iterate on datasets, instead of just architectures, and we provide an open, online platform with multiple rounds of challenges to support this iterative development. The first iteration of DataPerf contains five benchmarks covering a wide spectrum of data-centric techniques, tasks, and modalities in vision, speech, acquisition, debugging, and diffusion prompting, and we support hosting new contributed benchmarks from the community. The benchmarks, online evaluation platform, and baseline implementations are open source, and the MLCommons Association will maintain DataPerf to ensure long-term benefits to academia and industry.

Originele taal-2Engels
Titel37th Conference on Neural Information Processing Systems, NeurIPS 2023
UitgeverijNeural information processing systems foundation
Aantal pagina's28
StatusGepubliceerd - 2023
Evenement37th Conference on Neural Information Processing Systems, NeurIPS 2023 - New Orleans, Verenigde Staten van Amerika
Duur: 10 dec. 202316 dec. 2023
Congresnummer: 37

Publicatie series

NaamAdvances in Neural Information Processing Systems
Volume36
ISSN van geprinte versie1049-5258

Congres

Congres37th Conference on Neural Information Processing Systems, NeurIPS 2023
Verkorte titelNeurIPS 2023
Land/RegioVerenigde Staten van Amerika
StadNew Orleans
Periode10/12/2316/12/23

Vingerafdruk

Duik in de onderzoeksthema's van 'DataPerf: Benchmarks for Data-Centric AI Development'. Samen vormen ze een unieke vingerafdruk.

Citeer dit