Deep Learning Model for Real-Time Nuchal Translucency Assessment at Prenatal US.

Journal: Radiology. Artificial intelligence
Published Date:

Abstract

Purpose To develop and evaluate an artificial intelligence-based model for real-time nuchal translucency (NT) plane identification and measurement in prenatal US assessments. Materials and Methods In this retrospective multicenter study conducted from January 2022 to October 2023, the Automated Identification and Measurement of NT (AIM-NT) model was developed and evaluated using internal and external datasets. NT plane assessment, including identification of the NT plane and measurement of NT thickness, was independently conducted by AIM-NT and experienced radiologists, with the results subsequently audited by radiology specialists and accuracy compared between groups. To assess alignment of artificial intelligence with radiologist workflow, discrepancies between the AIM-NT model and radiologists in NT plane identification time and thickness measurements were evaluated. Results The internal dataset included a total of 3959 NT images from 3153 fetuses, and the external dataset included 267 US videos from 267 fetuses. On the internal testing dataset, AIM-NT achieved an area under the receiver operating characteristic curve of 0.92 for NT plane identification. On the external testing dataset, there was no evidence of differences between AIM-NT and radiologists in NT plane identification accuracy (88.8% vs 87.6%, = .69) or NT thickness measurements on standard and nonstandard NT planes ( = .29 and .59). AIM-NT demonstrated high consistency with radiologists in NT plane identification time, with 1-minute discrepancies observed in 77.9% of cases, and NT thickness measurements, with a mean difference of 0.03 mm and mean absolute error of 0.22 mm (95% CI: 0.19, 0.25). Conclusion AIM-NT demonstrated high accuracy in identifying the NT plane and measuring NT thickness on prenatal US images, showing minimal discrepancies with radiologist workflow. Ultrasound, Fetus, Segmentation, Feature Detection, Diagnosis, Convolutional Neural Network (CNN) © RSNA, 2025 See also commentary by Horii in this issue.

Authors

  • Yuanji Zhang
    Department of Ultrasound, Luohu People's Hospital, Shenzhen, China.
  • Xin Yang
    Department of Oral Maxillofacial-Head Neck Oncology, Ninth People's Hospital, College of Stomatology, Shanghai Jiao Tong University School of Medicine, National Clinical Research Center for Oral Diseases, Shanghai Key Laboratory of Stomatology & Shanghai Research Institute of Stomatology, Shanghai, China.
  • Chunya Ji
    Center for Medical Ultrasound, The Affiliated Suzhou Hospital of Nanjing Medical University, Suzhou Municipal Hospital, Gusu School, Nanjing Medical University, Suzhou, Jiangsu, China.
  • Xindi Hu
    Shenzhen RayShape Medical Technology Co. Ltd., Shenzhen, China.
  • Yan Cao
    School of Pharmacy, Second Military Medical University, 325 Guohe Road, Shanghai, 200433, China.
  • Chaoyu Chen
    National-Regional Key Technology Engineering Laboratory for Medical Ultrasound, School of Biomedical Engineering, Health Science Center, Shenzhen University, Shenzhen, China; Medical Ultrasound Image Computing (MUSIC) Laboratory, Shenzhen University, Shenzhen, China.
  • He Sui
  • Binghan Li
    National-Regional Key Technology Engineering Laboratory for Medical Ultrasound, School of Biomedical Engineering, Health Science Center, Shenzhen University, Shenzhen, Guangdong, 518073, China.
  • Chaojiong Zhen
    Department of Medical Ultrasonics, First People's Hospital of Foshan, Foshan, PR China.
  • Weijun Huang
  • Xuedong Deng
    Center for Medical Ultrasound, Nanjing Medical University Affiliated Suzhou Hospital, Suzhou, China.
  • Linliang Yin
    Center for Medical Ultrasound, The Affiliated Suzhou Hospital of Nanjing Medical University, Suzhou Municipal Hospital, Gusu School, Nanjing Medical University, Suzhou, Jiangsu, China.
  • Dong Ni