PERSONAL Sign in with your SPIE account to access your personal subscriptions or to use specific features such as save to my library, sign up for alerts, save searches, etc.
Deep learning networks are gaining popularity in many medical image analysis tasks due to their generalized ability to automatically extract relevant features from raw images. However, this can make the learning problem unnecessarily harder requiring network architectures of high complexity. In case of anomaly detection, in particular, there is often sufficient regional difference between the anomaly and the surrounding parenchyma that could be easily highlighted through bottom-up saliency operators. In this paper we propose a new hybrid deep learning network using a combination of raw image and such regional maps to more accurately learn the anomalies using simpler network architectures. Specifically, we modify a deep learning network called U-Net using both the raw and pre-segmented images as input to produce joint encoding (contraction) and expansion paths (decoding) in the U-Net. We present results of successfully delineating subdural and epidural hematomas in brain CT imaging and liver hemangioma in abdominal CT images using such network.
Alex Karargyros andTanveer Syeda-Mahmood
"Saliency U-Net: A regional saliency map-driven hybrid deep learning network for anomaly segmentation", Proc. SPIE 10575, Medical Imaging 2018: Computer-Aided Diagnosis, 105751T (27 February 2018); https://doi.org/10.1117/12.2293976
ACCESS THE FULL ARTICLE
INSTITUTIONAL Select your institution to access the SPIE Digital Library.
PERSONAL Sign in with your SPIE account to access your personal subscriptions or to use specific features such as save to my library, sign up for alerts, save searches, etc.
The alert did not successfully save. Please try again later.
Alex Karargyros, Tanveer Syeda-Mahmood, "Saliency U-Net: A regional saliency map-driven hybrid deep learning network for anomaly segmentation," Proc. SPIE 10575, Medical Imaging 2018: Computer-Aided Diagnosis, 105751T (27 February 2018); https://doi.org/10.1117/12.2293976