A method for the automated classification of benign and malignant masses on digital breast tomosynthesis images using machine learning and radiomic features.

Journal: Radiological physics and technology
Published Date:

Abstract

In digital mammography, which is used for the early detection of breast tumors, oversight may occur due to overlap between normal tissues and lesions. However, since digital breast tomosynthesis can acquire three-dimensional images, tissue overlapping is reduced, and, therefore, the shape and distribution of the lesions can be easily identified. However, it is often difficult to distinguish between benign and malignant breast lesions on images, and the diagnostic accuracy can be reduced due to complications from radiological interpretations, owing to acquisition of a higher number of images. In this study, we developed an automated classification method for diagnosing breast lesions on digital breast tomosynthesis images using radiomics to comprehensively analyze the radiological images. We extracted an analysis area centered on the lesion and calculated 70 radiomic features, including the shape of the lesion, existence of spicula, and texture information. The accuracy was compared by inputting the obtained radiomic features to four classifiers (support vector machine, random forest, naïve Bayes, and multi-layer perceptron), and the final classification result was obtained as an output using a classifier with high accuracy. To confirm the effectiveness of the proposed method, we used 24 cases with confirmed pathological diagnosis on biopsy. We also compared the classification results based on the presence or absence of dimension reduction using least absolute shrinkage and a selection operator (LASSO). As a result, when the support vector machine was used as a classifier, the correct identification rate of the benign tumors was 55% and that of malignant tumors was 84%, with best results. These results indicate that the proposed method may help in more accurately diagnosing cases that are difficult to classify on images.

Authors

  • Ayaka Sakai
    School of Medical Sciences, Fujita Health University, 1-98 Dengakugakubo, Kutsukake, Toyoake, Aichi, 470-1192, Japan.
  • Yuya Onishi
    Graduate School of Health Sciences, Fujita Health University, 1-98 Dengakugakubo, Kutsukake cho, Toyoake City, Aichi, 470-1192, Japan.
  • Misaki Matsui
    Aoyama Hospital, 100-1 Douji, Kosakai, Toyokawa, Aichi, 441-0195, Japan.
  • Hidetoshi Adachi
    Aoyama Hospital, 100-1 Douji, Kosakai, Toyokawa, Aichi, 441-0195, Japan.
  • Atsushi Teramoto
    Faculty of Radiological Technology, School of Health Sciences, Fujita Health University, 1-98 Dengakugakubo, Kutsukake, Toyoake, Aichi 470-1192, Japan.
  • Kuniaki Saito
    Graduate School of Health Sciences, Fujita Health University, 1-98 Dengakugakubo, Kutsukake cho, Toyoake City, Aichi, 470-1192, Japan.
  • Hiroshi Fujita
    Department of Intelligent Image Information, Division of Regeneration and Advanced Medical Sciences, Graduate School of Medicine, Gifu University, 1-1 Yanagido, Gifu 501-1194, Japan.