Explaining Black Box Drug Target Prediction Through Model Agnostic Counterfactual Samples.

Journal: IEEE/ACM transactions on computational biology and bioinformatics
Published Date:

Abstract

Many high-performance DTA deep learning models have been proposed, but they are mostly black-box and thus lack human interpretability. Explainable AI (XAI) can make DTA models more trustworthy, and allows to distill biological knowledge from the models. Counterfactual explanation is one popular approach to explaining the behaviour of a deep neural network, which works by systematically answering the question "How would the model output change if the inputs were changed in this way?". We propose a multi-agent reinforcement learning framework, Multi-Agent Counterfactual Drug-target binding Affinity (MACDA), to generate counterfactual explanations for the drug-protein complex. Our proposed framework provides human-interpretable counterfactual instances while optimizing both the input drug and target for counterfactual generation at the same time. We benchmark the proposed MACDA framework using the Davis and PDBBind dataset and find that our framework produces more parsimonious explanations with no loss in explanation validity, as measured by encoding similarity. We then present a case study involving ABL1 and Nilotinib to demonstrate how MACDA can explain the behaviour of a DTA model in the underlying substructure interaction between inputs in its prediction, revealing mechanisms that align with prior domain knowledge.

Authors

  • Tri Minh Nguyen
  • Thomas P Quinn
    Centre for Pattern Recognition and Data Analytics, Deakin University, Geelong, Australia.
  • Thin Nguyen
    Applied Artificial Intelligence Institute (A2I2), Deakin University, Geelong, Australia.
  • Truyen Tran
    Centre for Pattern Recognition and Data Analytics, School of Information Technology, Deakin University, Geelong, Victoria, Australia.