Multimodal contrastive learning for enhanced explainability in pediatric brain tumor molecular diagnosis.

in Scientific reports by Sara Ketabi, Matthias W Wagner, Cynthia Hawkins, Uri Tabori, Birgit Betina Ertl-Wagner, Farzad Khalvati

TLDR

  • The study improves explainability and performance of CNNs in brain tumor diagnosis by integrating radiology reports and tumor location in a contrastive learning framework, achieving a test classification performance of 87.7% and improved explainability measures.

Abstract

Despite the promising performance of convolutional neural networks (CNNs) in brain tumor diagnosis from magnetic resonance imaging (MRI), their integration into the clinical workflow has been limited. That is mainly due to the fact that the features contributing to a model's prediction are unclear to radiologists and hence, clinically irrelevant, i.e., lack of explainability. As the invaluable sources of radiologists' knowledge and expertise, radiology reports can be integrated with MRI in a contrastive learning (CL) framework, enabling learning from image-report associations, to improve CNN explainability. In this work, we train a multimodal CL architecture on 3D brain MRI scans and radiology reports to learn informative MRI representations. Furthermore, we integrate tumor location, salient to several brain tumor analysis tasks, into this framework to improve its generalizability. We then apply the learnt image representations to improve explainability and performance of genetic marker classification of pediatric Low-grade Glioma, the most prevalent brain tumor in children, as a downstream task. Our results indicate a Dice score of 31.1% between the model's attention maps and manual tumor segmentation (as an explainability measure) with test classification performance of 87.7%, significantly outperforming the baselines. These enhancements can build trust in our model among radiologists, facilitating its integration into clinical practices for more efficient tumor diagnosis.

Overview

  • The study aims to improve the integration of convolutional neural networks (CNNs) in brain tumor diagnosis from magnetic resonance imaging (MRI) by increasing explainability through a multimodal contrastive learning (CL) framework.
  • The framework combines 3D brain MRI scans and radiology reports to learn informative MRI representations and improve generalizability by integrating tumor location.
  • The study focuses on genetic marker classification of pediatric Low-grade Glioma, the most prevalent brain tumor in children, as a downstream task.

Comparative Analysis & Findings

  • The multimodal CL architecture achieves a Dice score of 31.1% between the model's attention maps and manual tumor segmentation, indicating improved explainability.
  • The model achieves a test classification performance of 87.7%, significantly outperforming the baselines.
  • The study demonstrates the potential for Explainable AI (XAI) in medical imaging, building trust among radiologists and facilitating its integration into clinical practices.

Implications and Future Directions

  • The study's results have implications for improving the trustworthiness of AI systems in medicine by providing insights into the decision-making process.
  • Future research directions include incorporating additional modalities, such as clinical data or genomic information, to further improve the framework's generalizability and performance.
  • The study's findings can also be applied to other medical imaging tasks, enabling Explainable AI solutions for a broader range of applications.