Explaining the unexplainable Part II: SHAP and SAGE
Image courtesy of iancovert.com Academic's take This is a follow-up to our post on LIME , and a post I'm writing on request. Having discovered an excellent write-up that explains both SHAP (SHapley Additive exPlanations) and SAGE (Shapley Additive Global importancE), I will focus on the why questions, possible links to counterfactuals and causality, and the implications for data centricity. Why do we need SHAP and SAGE? The need for explainability methods stems clearly from the fact that most ensemble methods are black boxes in nature: they minimize prediction error but they obscure how the predictions are made. This is problematic because some predictions are only as good as the underlying conditions are favorable. For example, Zillow's iBuying was a notorious failure , likely due to a lack of clarity and transparency in how the predictions were made. Could Zillow have done better if the data science team had paid more attention to explaining the mod...