Explainable Deep Learning Framework for Brain Tumor Segmentation Using Vision Transformer and Conditional Random Fields

Read the full article See related articles

Listed in

This article is not in any list yet, why not save it to one of your lists.
Log in to save this article

Abstract

Background/Objectives: Brain tumor segmentation is critical for accurate identification and treatment planning. However, delineating intricate tumor areas like Edema, Tumor Core, and Enhancing Tumor remains challenging owing to diversity in dimensions, morphology, and delineation. Methods: This introduces an innovative a novel deep learning model integrating a Pretrained Vision Transformer (ViT) for global context representation, Atrous Spatial Pyramid Pooling (ASPP) in both encoder and merging stages, and ReLU6. Additionally, attention mechanisms and Conditional Random Fields (CRF) were incorporated for refining feature emphasis and segmentation boundaries. Results: The model was trained and tested on the BraTS 2020 dataset and compared with state-of-the-art methods. The suggested model obtained the highest Dice coefficients across all tumor regions, with improvements of up to 4% for Enhancing Tumor and Edema. Grad-CAM heatmaps illustrated the model's proficiency in precisely identifying tumor areas, hence enhancing the interpretability of predictions. Conclusions: The incorporation of multi-scale feature extraction, attention mechanism, and global context representation markedly improved segmentation accuracy and interpretability. The findings underscore the proposed model's capacity to enhance clinical processes and establish a standard for brain tumor segmentation tasks. Future endeavors involve testing on varied datasets and investigating lightweight structures for real-time applications.

Article activity feed