HD-Former: A hierarchical dependency Transformer for medical image segmentation.

Journal: Computers in biology and medicine
Published Date:

Abstract

Medical image segmentation is a compelling fundamental problem and an important auxiliary tool for clinical applications. Recently, the Transformer model has emerged as a valuable tool for addressing the limitations of convolutional neural networks by effectively capturing global relationships and numerous hybrid architectures combining convolutional neural networks (CNNs) and Transformer have been devised to enhance segmentation performance. However, they suffer from multilevel semantic feature gaps and fail to account for multilevel dependencies between space and channel. In this paper, we propose a hierarchical dependency Transformer for medical image segmentation, named HD-Former. First, we utilize a Compressed Bottleneck (CB) module to enrich shallow features and localize the target region. We then introduce the Dual Cross Attention Transformer (DCAT) module to fuse multilevel features and bridge the feature gap. In addition, we design the broad exploration network (BEN) that cascades convolution and self-attention from different percepts to capture hierarchical dense contextual semantic features locally and globally. Finally, we exploit uncertain multitask edge loss to adaptively map predictions to a consistent feature space, which can optimize segmentation edges. The extensive experiments conducted on medical image segmentation from ISIC, LiTS, Kvasir-SEG, and CVC-ClinicDB datasets demonstrate that our HD-Former surpasses the state-of-the-art methods in terms of both subjective visual performance and objective evaluation. Code: https://github.com/barcelonacontrol/HD-Former.

Authors

  • Haifan Wu
    School of Mathematics and Computer Sciences, Nanchang University, Nanchang, 330031, China. Electronic address: wuhaifan@email.ncu.edu.cn.
  • Weidong Min
    School of Information Engineering, Nanchang University, Nanchang 330031, China. minweidong@ncu.edu.cn.
  • Di Gai
    School of Mathematics and Computer Sciences, Nanchang University, Nanchang, 330031, China; Institute of Metaverse, Nanchang University, Nanchang, 330031, China; Jiangxi Key Laboratory of Virtual Reality, Nanchang, 330031, China. Electronic address: gaidi@ncu.edu.cn.
  • Zheng Huang
    Institute of Laser and Optoelectronics Technology, Fujian Provincial Key Laboratory for Photonics Technology, Key Laboratory of OptoElectronic Science and Technology for Medicine of Ministry of Education, Fujian Normal University, Fuzhou, China.
  • Yuhan Geng
    School of Public Health, University of Michigan, Ann Arbor, MI, 48105, USA. Electronic address: gengyh@umich.edu.
  • Qi Wang
    Biotherapeutics Discovery Research Center, Shanghai Institute of Materia Medica, Chinese Academy of Sciences, Shanghai, 201203, China.
  • Ruibin Chen
    School of Mathematics and Computer Sciences, Nanchang University, Nanchang, 330031, China; Information Department, The First Affiliated Hospital of Nanchang University, Nanchang, 330096, China. Electronic address: Ruibin.Chen@email.ncu.edu.cn.