Skip to main content

Why Are We Using Black Box Models in AI When We Don’t Need To? A Lesson From An Explainable AI Competition

Forthcoming. Read the abstract to learn more!
Why Are We Using Black Box Models in AI When We Don’t Need To? A Lesson From An Explainable AI Competition
·
Contributors (2)
Published
Nov 01, 2019

Abstract

In 2018, a landmark challenge in artificial intelligence took place, namely the Explainable Machine Learning Challenge. The goal of the competition was to create a complicated black box model for the dataset and explain how it worked. One team did not follow the rules. Instead of sending in a black box, they created a model that was fully interpretable. This leads to the question of whether the real-world of machine learning is similar to the Explainable Machine Learning Challenge, where black box models are used even when they are not needed. We discuss this team’s thought processes during the competition and their implications, which reach far beyond the competition itself.

Keywords: Interpretability, Explainability, Machine Learning, Finance

Article forthcoming.

Comments
0
comment

No comments here