Research Working Group
Dynabench Working Group
- Training Working Group
- Inference Working Group
- Datasets Working Group
- Best Practices Working Group
- Research Working Group
Accelerate machine innovation and increase scientific rigor in machine learning by providing a flexible ML benchmarking platform.
Dynabench is a research platform for dynamic data collection and benchmarking. In particular, Dynabench challenges existing ML benchmarking dogma by embracing dynamic dataset generation. Benchmarks for machine learning solutions based on static datasets have well-known issues: they saturate quickly, are susceptible to overfitting, contain exploitable annotator artifacts and have unclear or imperfect evaluation metrics. In this sense, Dynabench enables a scientific experiment: is it possible to make faster progress if data is collected dynamically, with humans and models in the loop, rather than in the old-fashioned static way? Further, DynaBench enables an ecosystem of other ML benchmarks in areas such as algorithmic efficiency.
Roadmap for Dynabench development
Dynabench benchmarking platform
Dynabench community support
Weekly on Thursday from 10:30-11:00pm Pacific.
Working Group Resources
Working Group Chair Emails
Douwe Kiela email@example.com
Adina Williams firstname.lastname@example.org
Working Group Chair Bios
Douwe Kiela is a Research Scientist at Facebook AI Research, working on natural language processing and multimodal reasoning and understanding. His work has mainly been focused on representation learning, grounded language learning and multi-agent communication. Recently, he has become interested in improving the way we evaluate AI systems.
Adina Williams is an AI Research Scientist in the Facebook Artificial Intelligence Research (FAIR) Group in New York City. She received her PhD in Linguistics under the supervision of Liina Pylkkänen in the fall of 2018 from New York University, where she also contributed to the Machine Learning for Language Laboratory in the Center for Data Science with the support of Sam Bowman. Her research aims to bridge the gap between linguistics, cognitive science, and NLP. She is currently working on projects involving natural language inference, evaluating model biases, and information theoretic approaches to computational morphology.