Mitigating data bias and ensuring reliable evaluation of AI models with shortcut hull learning.
Journal:
Nature communications
Published Date:
Jul 1, 2025
Abstract
Shortcut learning poses a significant challenge to both the interpretability and robustness of artificial intelligence, arising from dataset biases that lead models to exploit unintended correlations, or shortcuts, which undermine performance evaluations. Addressing these inherent biases is particularly difficult due to the complex, high-dimensional nature of data. Here, we introduce shortcut hull learning, a diagnostic paradigm that unifies shortcut representations in probability space and utilizes diverse models with different inductive biases to efficiently learn and identify shortcuts. This paradigm establishes a comprehensive, shortcut-free evaluation framework, validated by developing a shortcut-free topological dataset to assess deep neural networks' global capabilities, enabling a shift from Minsky and Papert's representational analysis to an empirical investigation of learning capacity. Unexpectedly, our experimental results suggest that under this framework, convolutional models-typically considered weak in global capabilities-outperform transformer-based models, challenging prevailing beliefs. By enabling robust and bias-free evaluation, our framework uncovers the true model capabilities beyond architectural preferences, offering a foundation for advancing AI interpretability and reliability.