Radiomic Stability and Segmentation Performance of Six Deep Learning Models for Ovarian Tumor Ultrasound Images
Listed in
This article is not in any list yet, why not save it to one of your lists.Abstract
Recent advances in deep learning have produced highly robust and efficient models for medical image segmentation. Despite this, few studies assess the radiomic stability of these state-of-the-art approaches for applications in ovarian tumor segmentation. This study compared six deep learning architectures with unique backbones by evaluating their performance and radiomic stability on 1469 two-dimensional (2D) ultrasound images from the Multi-Modality Ovarian Tumor Ultrasound (MMOTU) dataset. The tested models were U-Net++, DeepLabV3+, U-Net, SegFormer, FPN, and LinkNet. Standard performance was assessed using Dice and IoU scores, precision, recall, specificity, size error, and 95th percentile Hausdorff distance (HD95). Radiomic stability evaluation involved extracting 102 quantitative features, followed by a Pearson correlation analysis between the predicted and ground truth masks. All models performed well, with average Dice scores of 0.9311, 0.9259, 0.9277, 0.9271, 0.9280, and 0.9213 for U-Net++, DeepLabV3+, U-Net, SegFormer, FPN, and LinkNet, respectively. Radiomic correlations were similarly high, ranging from 0.9400 to 0.9518. Small but significant differences were observed across models (p < 0.05). U-Net++ achieved the best Dice and IoU scores, precision, HD95, and size error, while SegFormer obtained the strongest radiomic stability. These results provide a comparative benchmark for ovarian tumor segmentation, underscoring how model choice can influence mask quality and radiomic stability.