An orchestration learning framework for ultrasound imaging: Prompt-Guided Hyper-Perception and Attention-Matching Downstream Synchronization.

Journal: Medical image analysis
Published Date:

Abstract

Ultrasound imaging is pivotal in clinical diagnostics due to its affordability, portability, safety, real-time capability, and non-invasive nature. It is widely utilized for examining various organs, such as the breast, thyroid, ovary, cardiac, and more. However, the manual interpretation and annotation of ultrasound images are time-consuming and prone to variability among physicians. While single-task artificial intelligence (AI) solutions have been explored, they are not ideal for scaling AI applications in medical imaging. Foundation models, although a trending solution, often struggle with real-world medical datasets due to factors such as noise, variability, and the incapability of flexibly aligning prior knowledge with task adaptation. To address these limitations, we propose an orchestration learning framework named PerceptGuide for general-purpose ultrasound classification and segmentation. Our framework incorporates a novel orchestration mechanism based on prompted hyper-perception, which adapts to the diverse inductive biases required by different ultrasound datasets. Unlike self-supervised pre-trained models, which require extensive fine-tuning, our approach leverages supervised pre-training to directly capture task-relevant features, providing a stronger foundation for multi-task and multi-organ ultrasound imaging. To support this research, we compiled a large-scale Multi-task, Multi-organ public ultrasound dataset (M-US), featuring images from 9 organs and 16 datasets, encompassing both classification and segmentation tasks. Our approach employs four specific prompts-Object, Task, Input, and Position-to guide the model, ensuring task-specific adaptability. Additionally, a downstream synchronization training stage is introduced to fine-tune the model for new data, significantly improving generalization capabilities and enabling real-world applications. Experimental results demonstrate the robustness and versatility of our framework in handling multi-task and multi-organ ultrasound image processing, outperforming both specialist models and existing general AI solutions. Compared to specialist models, our method improves segmentation from 82.26% to 86.45%, classification from 71.30% to 79.08%, while also significantly reducing model parameters.

Authors

  • Zehui Lin
    National-Regional Key Technology Engineering Laboratory for Medical Ultrasound, Guangdong Key Laboratory for Biomedical Measurements and Ultrasound Imaging, School of Biomedical Engineering, Health Science Center, Shenzhen University, Shenzhen, 518060, China.
  • Shuo Li
    Shanghai Mental Health Center, Shanghai Jiao Tong University School of Medicine, Shanghai, China.
  • Shanshan Wang
    Key Laboratory of Agri-food Safety and Quality, Institute of Quality Standard and Testing Technology for Agro-Products, Chinese Academy of Agricultural Sciences, Ministry of Agriculture of China, Beijing, 100081, PR China.
  • Zhifan Gao
    School of Biomedical Engineering, Sun Yat-sen University, Shenzhen, China.
  • Yue Sun
    Department of Rheumatology, The First Affiliated Hospital of Anhui University of Chinese Medicine, Hefei, Anhui, China.
  • Chan-Tong Lam
    Faculty of Applied Sciences, Macao Polytechnic University, 999078, Macao Special Administrative Region of China.
  • Xindi Hu
    Shenzhen RayShape Medical Technology Co. Ltd., Shenzhen, China.
  • Xin Yang
    Department of Oral Maxillofacial-Head Neck Oncology, Ninth People's Hospital, College of Stomatology, Shanghai Jiao Tong University School of Medicine, National Clinical Research Center for Oral Diseases, Shanghai Key Laboratory of Stomatology & Shanghai Research Institute of Stomatology, Shanghai, China.
  • Dong Ni
  • Tao Tan
    Faculty of Applied Sciences, Macao Polytechnic University, Macao, China.