Manipulating and measuring model interpretability

Video description

Machine learning is increasingly used to make decisions that affect people’s lives in critical domains like criminal justice, fair lending, and medicine. While most of the research in machine learning focuses on improving the performance of models on held-out datasets, this is seldom enough to convince end users that these models are trustworthy and reliable in the wild. To address this problem, a new line of research has emerged that focuses on developing interpretable machine learning methods and helping end users make informed decisions. Despite the growing body of work in developing interpretable models, there is still no consensus on the definition and quantification of interpretability.

Forough Poursabzi-Sangdeh (Microsoft Research NYC) argues that to understand interpretability, we need to bring humans in the loop and run human-subject experiments. Forough approaches the problem of interpretability from an interdisciplinary perspective built on decades of research in psychology, cognitive science, and social science to understand human behavior and trust. She describes a set of controlled user experiments in which researchers manipulated various design factors in models that are commonly thought to make them more or less interpretable and measured their influence on users’ behavior. The findings emphasize the importance of studying how models are presented to people and empirically verifying that interpretable models achieve their intended effects on end users.

This session was recorded at the 2019 O'Reilly Artificial Intelligence Conference in New York.

Product information

  • Title: Manipulating and measuring model interpretability
  • Author(s): Forough Poursabzi-Sangdeh
  • Release date: October 2019
  • Publisher(s): O'Reilly Media, Inc.
  • ISBN: 0636920339700