Detecting shortcut learning for fair medical AI using shortcut testing.

Journal: Nature communications
Published Date:

Abstract

Machine learning (ML) holds great promise for improving healthcare, but it is critical to ensure that its use will not propagate or amplify health disparities. An important step is to characterize the (un)fairness of ML models-their tendency to perform differently across subgroups of the population-and to understand its underlying mechanisms. One potential driver of algorithmic unfairness, shortcut learning, arises when ML models base predictions on improper correlations in the training data. Diagnosing this phenomenon is difficult as sensitive attributes may be causally linked with disease. Using multitask learning, we propose a method to directly test for the presence of shortcut learning in clinical ML systems and demonstrate its application to clinical tasks in radiology and dermatology. Finally, our approach reveals instances when shortcutting is not responsible for unfairness, highlighting the need for a holistic approach to fairness mitigation in medical AI.

Authors

  • Alexander Brown
    Chelsea and Westminster Hospital NHS Foundation Trust, 369 Fulham Road, London, SW10 9NH, UK.
  • Nenad Tomasev
    DeepMind, London, EC4A 3TW, UK.
  • Jan Freyberg
    Google Research, Mountain View, CA, USA.
  • Yuan Liu
    Department of General Surgery, Wuxi People's Hospital Affiliated to Nanjing Medical University, Wuxi, China.
  • Alan Karthikesalingam
    Department of Outcomes Research, St George's Vascular Institute, London, SW17 0QT, United Kingdom.
  • Jessica Schrouff
    Laboratory of Behavioral & Cognitive Neuroscience, Stanford Human Intracranial Cognitive Electrophysiology Program (SHICEP), Department of Neurology & Neurological Sciences, Stanford University, Stanford, CA, USA; Department of Computer Science, University College London, United Kingdom.