1. Intracerebral Hemorrhage Segmentation on Noncontrast Computed Tomography Using a Masked Loss Function U-Net Approach
- Author
-
Nadine A, Coorens, Kevin Groot, Lipman, Sanjith P, Krishnam, Can Ozan, Tan, Lejla, Alic, Rajiv, Gupta, TechMed Centre, Robotics and Mechatronics, and Magnetic Detection and Imaging
- Subjects
brain extraction mask ,2023 OA procedure ,hemorrhagic stroke ,Radiology, Nuclear Medicine and imaging ,dice coefficient ,convolutional neural model - Abstract
Objective Intracerebral hemorrhage (ICH) volume is a strong predictor of outcome in patients presenting with acute hemorrhagic stroke. It is necessary to segment the hematoma for ICH volume estimation and for computerized extraction of features, such as spot sign, texture parameters, or extravasated iodine content at dual-energy computed tomography. Manual and semiautomatic segmentation methods to delineate the hematoma are tedious, user dependent, and require trained personnel. This article presents a convolutional neural network to automatically delineate ICH from noncontrast computed tomography scans of the head. Methods A model combining a U-Net architecture with a masked loss function was trained on standard noncontrast computed tomography images that were down sampled to 256 × 256 size. Data augmentation was applied to prevent overfitting, and the loss score was calculated using the soft Dice loss function. The Dice coefficient and the Hausdorff distance were computed to quantitatively evaluate the segmentation performance of the model, together with the sensitivity and specificity to determine the ICH detection accuracy. Results The results demonstrate a median Dice coefficient of 75.9% and Hausdorff distance of 2.65 pixels in segmentation performance, with a detection sensitivity of 77.0% and specificity of 96.2%. Conclusions The proposed masked loss U-Net is accurate in the automatic segmentation of ICH. Future research should focus on increasing the detection sensitivity of the model and comparing its performance with other model architectures.
- Published
- 2023