ExplaiNN: interpretable and transparent neural networks for genomics.

Journal: Genome biology
PMID:

Abstract

Deep learning models such as convolutional neural networks (CNNs) excel in genomic tasks but lack interpretability. We introduce ExplaiNN, which combines the expressiveness of CNNs with the interpretability of linear models. ExplaiNN can predict TF binding, chromatin accessibility, and de novo motifs, achieving performance comparable to state-of-the-art methods. Its predictions are transparent, providing global (cell state level) as well as local (individual sequence level) biological insights into the data. ExplaiNN can serve as a plug-and-play platform for pretrained models and annotated position weight matrices. ExplaiNN aims to accelerate the adoption of deep learning in genomic sequence analysis by domain experts.

Authors

  • Gherman Novakovsky
    Centre for Molecular Medicine and Therapeutics, BC Children's Hospital Research Institute, Vancouver, BC, V5Z 4H4, Canada.
  • Oriol Fornes
    Centre for Molecular Medicine and Therapeutics, BC Children's Hospital Research Institute, Vancouver, BC, V5Z 4H4, Canada. oriol@cmmt.ubc.ca.
  • Manu Saraswat
    Institut de Biologie Computationnelle, Montpellier, France.
  • Sara Mostafavi
    Department of Statistics, University of British Columbia, Vancouver, BC V6T 1Z4, Canada; cb@hms.harvard.edu saram@stat.ubc.ca.
  • Wyeth W Wasserman
    Centre for Molecular Medicine and Therapeutics, Child and Family Research Institute, Department of Medical Genetics, University of British Columbia Vancouver, British Columbia V5Z 4H4, Canada. Electronic address: wyeth@cmmt.ubc.ca.