Machine learning models with distinct Shapley value explanations decouple feature attribution and interpretation for chemical compound predictions

Explaining black box predictions of machine learning (ML) models is a topical issue in artificial intelligence (AI) research. For the identification of features determining predictions, the Shapley value formalism originally developed in game theory is widely used in different fields. Typically, Shapley values quantifying feature contributions to predictions need to be approximated in machine learning. We introduce a framework for the calculation of exact Shapley values for 4 kernel functions used in support vector machine (SVM) models and analyze consistently accurate compound activity predictions based on exact Shapley values. Dramatic changes in feature contributions are detected depending on the kernel function, leading to mostly distinct explanations of predictions of the same test compounds. Very different feature contributions yield comparable predictions, which complicate numerical and graphical model explanation and decouple feature attribution and human interpretability.

Informationen zur Zitierung

Roth, Jannik P.; Bajorath, Jürgen: Machine learning models with distinct Shapley value explanations decouple feature attribution and interpretation for chemical compound predictions, Cell Reports Physical Science, 2024, 5, https://www.cell.com/cell-reports-physical-science/fulltext/S2666-3864(24)00388-6, Roth.Bajorath.2024b,

Assoziierte Lamarr-ForscherInnen

lamarr institute person Bajorath Juergen - Lamarr Institute for Machine Learning (ML) and Artificial Intelligence (AI)

Prof. Dr. Jürgen Bajorath

Area Chair Life Sciences zum Profil