Multimodal Artificial Intelligence in Medicine.

Journal: Kidney360
Published Date:

Abstract

Traditional medical artificial intelligence models that are approved for clinical use restrict themselves to single-modal data ( e.g ., images only), limiting their applicability in the complex, multimodal environment of medical diagnosis and treatment. Multimodal transformer models in health care can effectively process and interpret diverse data forms, such as text, images, and structured data. They have demonstrated impressive performance on standard benchmarks, like United States Medical Licensing Examination question banks, and continue to improve with scale. However, the adoption of these advanced artificial intelligence models is not without challenges. While multimodal deep learning models like transformers offer promising advancements in health care, their integration requires careful consideration of the accompanying ethical and environmental challenges.

Authors

  • Conor S Judge
    HRB-Clinical Research Facility, University of Galway, Galway, Ireland.
  • Finn Krewer
    HRB-Clinical Research Facility, University of Galway, Galway, Ireland.
  • Martin J O'Donnell
    HRB-Clinical Research Facility, University of Galway, Galway, Ireland.
  • Lisa Kiely
    HRB-Clinical Research Facility, University of Galway, Galway, Ireland.
  • Donal Sexton
    Department of Medicine, Trinity College Dublin, Dublin, Ireland.
  • Graham W Taylor
    University of Guelph, 50 Stone Road East, N1G 2W1, Guelph, ON, Canada. gwtaylor@uoguelph.ca.
  • Joshua August Skorburg
    Duke University, Durham, North Carolina, USA.
  • Bryan Tripp
    Department of Systems Design Engineering, University of Waterloo, Canada; Centre for Theoretical Neuroscience, University of Waterloo, Canada. Electronic address: bptripp@uwaterloo.ca.