Institutional Scholarship

Understanding machine learning models through Shapley values and the associated uncertainties

Show simple item record

dc.contributor.advisor Friedler, Sorelle Ngo, Jason 2021-07-12T11:57:41Z 2021-07-12T11:57:41Z 2021
dc.description.abstract To explain black-box machine learning models, one popular method is to calculate the feature importance using Shapley value, a solution to distribute the payoff among the players in a coalition game. Though popular, Shapley-based feature importance methods suffer from interpretation issues when there are correlated features and out-of-distribution perturbed data points. Kumar et al. 2020b introduces Shapley residuals to quantify the information lost during the computation of Shapley values, e.g. feature interaction in the model; meanwhile, Slack et al. 2020b introduces Bayesian SHAP to calculate the variance of the Shapley values. With these uncertainty measures, one important question arises: how do practitioners interpret the Shapley values and does that interpretation match the underlying machine learning model? This thesis aims to answer that question and contribute to the literature by (1) introducing a linear mental model that represents how practitioners interpret Shapley values, and (2) introducing deviation as a way to measure how much this mental model deviates from the machine learning model. Our experiments show that this deviation measure is related to Shapley values and sampling variance, but also captures another aspect of the model: its nonlinearity.
dc.description.sponsorship Haverford College. Department of Computer Science
dc.language.iso eng
dc.title Understanding machine learning models through Shapley values and the associated uncertainties
dc.type Thesis
dc.rights.access Dark Archive until 2022-01-01, afterwards Haverford users only.

Files in this item

This item appears in the following Collection(s)

Show simple item record Except where otherwise noted, this item's license is described as



My Account