Medical Image Captioning Using Optimized Deep Learning Model.

Journal: Computational intelligence and neuroscience
Published Date:

Abstract

Medical image captioning provides the visual information of medical images in the form of natural language. It requires an efficient approach to understand and evaluate the similarity between visual and textual elements and to generate a sequence of output words. A novel show, attend, and tell model (ATM) is implemented, which considers a visual attention approach using an encoder-decoder model. But the show, attend, and tell model is sensitive to its initial parameters. Therefore, a Strength Pareto Evolutionary Algorithm-II (SPEA-II) is utilized to optimize the initial parameters of the ATM. Finally, experiments are considered using the benchmark data sets and competitive medical image captioning techniques. Performance analysis shows that the SPEA-II-based ATM performs significantly better as compared to the existing models.

Authors

  • Arjun Singh
    School of Computing and IT, Manipal University Jaipur, Jaipur, Rajasthan, India.
  • Jaya Krishna Raguru
    Manipal University Jaipur, Jaipur, India.
  • Gaurav Prasad
    Computer and Communication Engineering, Manipal University Jaipur, Jaipur, India.
  • Surbhi Chauhan
    Computer Science and Engineering, Jaipur Institute of Engineering and Management, Jaipur, India.
  • Pradeep Kumar Tiwari
    Manipal University Jaipur, Jaipur, India.
  • Atef Zaguia
    Department of Computer Science, College of Computers and Information Technology, Taif University, Taif, Saudi Arabia.
  • Mohammad Aman Ullah
    Department of Computer Science and Engineering, International Islamic University Chittagong, Chittagong, Bangladesh.