MixUNETR: A U-shaped network based on W-MSA and depth-wise convolution with channel and spatial interactions for zonal prostate segmentation in MRI.

Journal: Neural networks : the official journal of the International Neural Network Society
PMID:

Abstract

Magnetic resonance imaging (MRI) plays a pivotal role in diagnosing and staging prostate cancer. Precise delineation of the peripheral zone (PZ) and transition zone (TZ) within prostate MRI is essential for accurate diagnosis and subsequent artificial intelligence-driven analysis. However, existing segmentation methods are limited by ambiguous boundaries, shape variations and texture complexities between PZ and TZ. Moreover, they suffer from inadequate modeling capabilities and limited receptive fields. To address these challenges, we propose a Enhanced MixFormer, which integrates window-based multi-head self-attention (W-MSA) and depth-wise convolution with parallel design and cross-branch bidirectional interaction. We further introduce MixUNETR, which use multiple Enhanced MixFormers as encoder to extract features from both PZ and TZ in prostate MRI. This augmentation effectively enlarges the receptive field and enhances the modeling capability of W-MSA, ultimately improving the extraction of both global and local feature information from PZ and TZ, thereby addressing mis-segmentation and challenges in delineating boundaries between them. Extensive experiments were conducted, comparing MixUNETR with several state-of-the-art methods on the Prostate158, ProstateX public datasets and private dataset. The results consistently demonstrate the accuracy and robustness of MixUNETR in MRI prostate segmentation. Our code of methods is available at https://github.com/skyous779/MixUNETR.git.

Authors

  • Quanyou Shen
    School of Automation, Guangdong University of Technology, Guangzhou, Guangdong 510006, China (W.L., Q.S., J.T.); Guangdong Provincial Key Laboratory of Intelligent Decision and Cooperative Control, Guangzhou, Guangdong 510006, China (W.L., Q.S., J.T.); Guangdong-Hong Kong Joint Laboratory for Intelligent Decision and Cooperative Control, Guangzhou, Guangdong 510006, China (W.L., Q.S., J.T.).
  • Bowen Zheng
    Department of Mechanical Engineering, University of California, Berkeley, CA, 94720, USA.
  • Wenhao Li
    Flight Control Research Institute, Nanjing University of Aeronautics and Astronautics, Nanjing, Jiangsu, China.
  • Xiaoran Shi
    The Ministry of Education, Key Laboratory of Electronic Information Counter-measure and Simulation, Xidian University, Xi'an 710071, China; School of Electronic Engineering, Xidian University, Xi'an 710071, China.
  • Kun Luo
    Department of Orthopedic Surgery, First Affiliated Hospital of Nanchang University, Nanchang, China; Institute of Spine and Spinal Cord, Nanchang University, Nanchang, China.
  • Yuqian Yao
    School of Automation, Guangdong University of Technology, Guangzhou, 510006, China; Guangdong Provincial Key Laboratory of Intelligent Decision and Cooperative Control, Guangzhou, 510006, China; Guangdong-Hong Kong Joint Laboratory for Intelligent Decision and Cooperative Control, Guangzhou, 510006, China.
  • Xinyan Li
    Institute for Brain Research, Wuhan Center of Brain Science, Huazhong University of Science and Technology, Wuhan, 430030, China. lixinyan1026@163.com.
  • Shidong Lv
    Department of Urology, Nanfang Hospital, Southern Medical University, Guangzhou, Guangdong 510515, China (B.Z., F.M., X.S., S.L., Q.W.).
  • Jie Tao
    Guangdong Provincial Key Laboratory of Intelligent Decision and Cooperative Control, School of Automation, Guangdong University of Technology, Guangzhou 510006, China. Electronic address: jtao@iipc.zju.edu.cn.
  • Qiang Wei
    School of Biomedical Informatics, The University of Texas Health Science Center at Houston, Houston, TX, USA.