Biomedical Image Processing with Containers and Deep Learning: An Automated Analysis Pipeline: Data architecture, artificial intelligence, automated processing, containerization, and clusters orchestration ease the transition from data acquisition to insights in medium-to-large datasets.

Journal: BioEssays : news and reviews in molecular, cellular and developmental biology
PMID:

Abstract

Here, a streamlined, scalable, laboratory approach is discussed that enables medium-to-large dataset analysis. The presented approach combines data management, artificial intelligence, containerization, cluster orchestration, and quality control in a unified analytic pipeline. The unique combination of these individual building blocks creates a new and powerful analysis approach that can readily be applied to medium-to-large datasets by researchers to accelerate the pace of research. The proposed framework is applied to a project that counts the number of plasmonic nanoparticles bound to peripheral blood mononuclear cells in dark-field microscopy images. By using the techniques presented in this article, the images are automatically processed overnight, without user interaction, streamlining the path from experiment to conclusions.

Authors

  • Germán González
    1 Sierra Research, Alicante, Spain.
  • Conor L Evans
    Wellman Center for Photomedicine, Harvard Medical School, Massachusetts General Hospital, CNY149-3, 13th St, Charlestown, MA, 02129, USA.