Co-located with 35th International Conference on Machine Learning (ICML 2018)
15 July 2018, Stockholm, Sweden
1 May 2018, 23:59 Anywhere on Earth (AoE)
This workshop aims to bring together a growing community of researchers, practitioners, and policymakers concerned with fairness, accountability, and transparency in machine learning.
The past several years have seen growing recognition that machine learning raises new ethical, legal, and technical challenges. In particular, policymakers, regulators, and advocates have expressed fears about the potential discriminatory impact of machine learning models, with many calling for research into how we can use automated decision-making tools without inadvertently encoding and perpetuating societal biases. Concurrently, there has been increasing concern that the complexity of machine learning models limits their use in critical applications involving humans, such as loan approval to recidivism prediction. Most recently, there is emerging concern that the standard emphasis in machine learning on prediction rather than causation inhibits the ability of data-driven tools to produce meaningful, actionable recommendations.
The goal of this workshop is to provide researchers with a venue to explore how to characterize and address these issues in ways that are computationally rigorous and scientifically defensible. We seek contributions that attempt to measure and mitigate bias in machine learning, to audit and evaluate machine learning models, and to render such models more interpretable and their decisions more explainable.
This year, the workshop is co-located with ICML, and will consist of invited talks, invited panels, contributed talks, as well as a poster session. We welcome paper submissions from researchers and practitioners that address any issue of fairness, accountability, and transparency related to machine learning. In particular, we will place a special emphasis on causal inference to address questions of fairness, and to create recommendation systems directed at altering causal factors. We will also focus on issues surrounding the collection, measurement, and mitigation of biased data.
Papers must be limited to 4 pages, including figures and tables, and should use a standard 2-column, 11pt format. An additional 5th page containing only cited references is permitted. We recommend using the ICML template.
Accepted papers will be posted on the workshop website and should also be posted by the authors to arXiv. Note that the workshop's proceedings will be considered non-archival, meaning that contributors are free to publish their work in archival journals or conferences. Accepted papers will be either presented as a talk or poster (to be determined by the workshop organizers). We only wish to consider papers that have not yet been published elsewhere. Dual submissions are allowed.
All papers must must be anonymized for double-blind reviewing, and submitted using via Easy Chair.
22 May 2018
1 July 2018