Contents:
Overview of the 4 goal-oriented quadrants of additive Feature Importances, from datajms.com, courtesy of the author.
You have probably heard of Feature Importance methods: there are many of them around and they can be very useful for variable selection and model explanation. But there is more: the additive Variable Importance landscape has recently become structured and made systematically optimal.
This article is not just an other SHAP article, it presents similar notions which shares an important component: Shapley values. A structured 2 by 2 matrix is presented to better think about Variable Importances in terms of their goals and scopes. Focused on additive Feature attribution methods, the 4 identified quadrants are presented along with their “optimal” method: SHAP, SHAPLEY EFFECTS, SHAPloss and the very recent SAGE. Then, we will look into Shapley values and their properties, which make the 4 methods theoretically optimal. Finally, I will share my thoughts on the perspectives concerning Variable Importance methods.
1. Which purpose for Variable Importance?
So, what are Variable Importances and which properties should they have? We will focus on Variable Importances with the 2 following requirements:
While Feature attribution property is the essence of Variable Importance, the Additive importance requirement is more challenging. More well known Variable Importance methods break it: the Breiman Random Forest variable importance, Feature ablation, Permutation importance, etc. Let’s focus on Variable Importances with these 2 properties.
#data-science #shap #data analysis #data analysis