Machine learning models with distinct Shapley value explanations decouple feature attribution and interpretation for chemical compound predictions
Explaining black box predictions of machine learning (ML) models is a topical issue in artificial intelligence (AI) research. For the identification of features determining predictions, the Shapley value formalism originally developed in game theory is widely used in different fields. Typically, Shapley values quantifying feature contributions to predictions need to be approximated in machine learning. We introduce a framework for the calculation of exact Shapley values for 4 kernel functions used in support vector machine (SVM) models and analyze consistently accurate compound activity predictions based on exact Shapley values. Dramatic changes in feature contributions are detected depending on the kernel function, leading to mostly distinct explanations of predictions of the same test compounds. Very different feature contributions yield comparable predictions, which complicate numerical and graphical model explanation and decouple feature attribution and human interpretability.
- Published in:
Cell Reports Physical Science - Type:
Article - Authors:
Roth, Jannik P.; Bajorath, Jürgen - Year:
2024 - Source:
https://www.cell.com/cell-reports-physical-science/fulltext/S2666-3864(24)00388-6
Citation information
Roth, Jannik P.; Bajorath, Jürgen: Machine learning models with distinct Shapley value explanations decouple feature attribution and interpretation for chemical compound predictions, Cell Reports Physical Science, 2024, 5, https://www.cell.com/cell-reports-physical-science/fulltext/S2666-3864(24)00388-6, Roth.Bajorath.2024b,
@Article{Roth.Bajorath.2024b,
author={Roth, Jannik P.; Bajorath, Jürgen},
title={Machine learning models with distinct Shapley value explanations decouple feature attribution and interpretation for chemical compound predictions},
journal={Cell Reports Physical Science},
volume={5},
url={https://www.cell.com/cell-reports-physical-science/fulltext/S2666-3864(24)00388-6},
year={2024},
abstract={Explaining black box predictions of machine learning (ML) models is a topical issue in artificial intelligence (AI) research. For the identification of features determining predictions, the Shapley value formalism originally developed in game theory is widely used in different fields. Typically, Shapley values quantifying feature contributions to predictions need to be approximated in machine...}}