Abstract
Accurate multi-organ segmentation in abdominal CT is essential for diagnosis, treatment planning, and disease monitoring. However, it suffers from inaccurate segmentation due to intricate spatial relationships and varying organ shapes. Existing deep learning methods often struggle with imbalanced classes, size bias, and ambiguous boundaries, hindering segmentation accuracy. To address these challenges, this paper proposes Guided-nnUNet, a two-stage segmentation framework that decomposes abdominal multi-organ segmentation into organ localization and then localization-guided fine segmentation. In the first stage, a ResNet-50 model generates a low-dimensional localization map guiding organ locations. This spatial guidance is then fed into the second stage, where a 3D U-Net with dynamic affine feature-map transform performs the fine-grained segmentation by integrating spatial context from the localization map. Our evaluation on the publicly available AMOS and BTCV datasets demonstrates the effectiveness of the model. Guided-nnUNet achieves an average improvement of 7% and 9% on the AMOS and BTCV datasets, respectively, compared to the baseline nnUNet model. Additionally, our model outperforms the state-of-the-art MedNeXt by 3.6% and 5.3% on the AMOS and BTCV datasets, respectively. These results suggest that our two-stage solution offers a promising approach for accurate abdominal organ segmentation, particularly for overcoming the challenges associated with complex organ structures.