DIKUL - logo
E-viri
  • Hough-CNN: Deep learning fo...
    Milletari, Fausto; Ahmadi, Seyed-Ahmad; Kroll, Christine; Plate, Annika; Rozanski, Verena; Maiostre, Juliana; Levin, Johannes; Dietrich, Olaf; Ertl-Wagner, Birgit; Bötzel, Kai; Navab, Nassir

    Computer vision and image understanding, November 2017, 2017-11-00, Letnik: 164
    Journal Article

    •We propose Hough-CNN, a novel segmentation approach based on a voting strategy. We show that the method is multi-modal, multi-region, robust and implicitly encoding priors on anatomical shape and appearance. Hough-CNN delivers results comparable or superior to other state-of-the-art approaches while being entirely registration-free. In particular, it outperforms methods based on voxel-wise, semantic classification.•Hough-CNN is scalable to different modalities with little change in parameterisation. We demonstrate multi-region segmentation in MRI and midbrain segmentation in 3D freehand transcranial ultrasound (TCUS).•We propose and evaluate several different CNN architectures, with varying numbers of layers and convolutional kernels per layer. In this way we acquire insights on how different network architectures cope with the amount of variability present in medical volumes and image modalities.•We evaluate the impact of the number of annotated training examples on the final segmentations by training the networks with different amounts of data. In particular, we show how complex networks with higher parameter number cope with relatively small training datasets.•We adapted the Caffe framework to perform convolutions of volumetric data, preserving its third dimension across the whole network. We compare CNN performance using 3D convolution to the more common 2D convolution, as well as to a recent 2.5D approach. In this work we propose a novel approach to perform segmentation by leveraging the abstraction capabilities of convolutional neural networks (CNNs). Our method is based on Hough voting, a strategy that allows for fully automatic localisation and segmentation of the anatomies of interest. This approach does not only use the CNN classification outcomes, but it also implements voting by exploiting the features produced by the deepest portion of the network. We show that this learning-based segmentation method is robust, multi-region, flexible and can be easily adapted to different modalities. In the attempt to show the capabilities and the behaviour of CNNs when they are applied to medical image analysis, we perform a systematic study of the performances of six different network architectures, conceived according to state-of-the-art criteria, in various situations. We evaluate the impact of both different amount of training data and different data dimensionality (2D, 2.5D and 3D) on the final results. We show results on both MRI and transcranial US volumes depicting respectively 26 regions of the basal ganglia and the midbrain.