Quantifying interpretation reproducibility in Vision Transformer models with TAVAC.

Journal: Science advances
Published Date:

Abstract

Deep learning algorithms can extract meaningful diagnostic features from biomedical images, promising improved patient care in digital pathology. Vision Transformer (ViT) models capture long-range spatial relationships and offer robust prediction power and better interpretability for image classification tasks than convolutional neural network models. However, limited annotated biomedical imaging datasets can cause ViT models to overfit, leading to false predictions due to random noise. To address this, we introduce Training Attention and Validation Attention Consistency (TAVAC), a metric for evaluating ViT model overfitting and quantifying interpretation reproducibility. By comparing high-attention regions between training and testing, we tested TAVAC on four public image classification datasets and two independent breast cancer histological image datasets. Overfitted models showed significantly lower TAVAC scores. TAVAC also distinguishes off-target from on-target attentions and measures interpretation generalization at a fine-grained cellular level. Beyond diagnostics, TAVAC enhances interpretative reproducibility in basic research, revealing critical spatial patterns and cellular structures of biomedical and other general nonbiomedical images.

Authors

  • Yue Zhao
    The Affiliated Eye Hospital, Nanjing Medical University, Nanjing, China.
  • Dylan Agyemang
    Department of Mathematics and Statistics, University of North Carolina at Chapel Hill, Chapel Hill, NC, USA.
  • Yang Liu
    Department of Computer Science, Hong Kong Baptist University, Hong Kong, China.
  • Matt Mahoney
    The Jackson Laboratory for Mouse Genetics, Bar Harbor, ME, USA.
  • Sheng Li
    School of Data Science, University of Virginia, Charlottesville, VA, United States.