Visual Saliency Based Active Learning for Prostate MRI Segmentation

  • Dwarikanath MahapatraEmail author
  • Joachim M. Buhmann
Conference paper
Part of the Lecture Notes in Computer Science book series (LNCS, volume 9352)


We propose an active learning (AL) approach for prostate segmentation from magnetic resonance (MR) images. Our label query strategy is inspired from the principles of visual saliency that has similar considerations for choosing the most salient region. These similarities are encoded in a graph using classification maps and low level features. Random walks identify the most informative node which is equivalent to the label query sample in AL. Experimental results on the MICCAI 2012 Prostate segmentation challenge show the superior performance of our approach to conventional methods using fully supervised learning.


Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.


  1. 1.
    Cancer facts and figures (2014). amer. cancer soc.
  2. 2.
    Birkbeck, N., et. al.: Region-specific hierarchical segmentation of MR prostate using discriminative learning. In: Proc. MICCAI PROMISE 2012, pp. 4–11 (2012)Google Scholar
  3. 3.
    Breiman, L.: Random forests. Machine Learning 45(1), 5–32 (2001)CrossRefGoogle Scholar
  4. 4.
    Chapelle, O., Scholkopf, B., Zien, A.: Semi-Supervised Learning. MIT Press, Cambridge (2006)CrossRefGoogle Scholar
  5. 5.
    Cohen, M., et al.: Rapid and effective correction of RF inhomogeneity for high field magnetic resonance imaging. Human Brain Map 10(4), 204–211 (2000)CrossRefGoogle Scholar
  6. 6.
    Criminisi, A., Shotton., J.: Decision Forests for Computer Vision and Medical Image Analysis. SpringerGoogle Scholar
  7. 7.
    Gao, Q., et. al.: An automatic multi-atlas based prostate segmentation using local appearance-specific atlases and patch-based voxel weighting. In: Proc. MICCAI PROMISE 2012, pp. 12–19 (2012)Google Scholar
  8. 8.
    Ghose, S., et. al.: A random forest based classification approach to prostate segmentation in MRI. In: Proc. MICCAI PROMISE 2012, pp. 20–27 (2012)Google Scholar
  9. 9.
    Harel, J., Koch, C., Perona, P.: Graph based visual saliency. In: NIPS, pp. 545–552 (2006)Google Scholar
  10. 10.
    Kadir, T., Brady, M.: Saliency, scale and image description. International Journal of Computer Vision 45(2), 85–105 (2001)CrossRefGoogle Scholar
  11. 11.
    Liao, S., Gao, Y., Oto, A., Shen, D.: Representation learning: a unified deep learning framework for automatic prostate MR segmentation. In: Mori, K., Sakuma, I., Sato, Y., Barillot, C., Navab, N. (eds.) MICCAI 2013, Part II. LNCS, vol. 8150, pp. 254–261. Springer, Heidelberg (2013) Google Scholar
  12. 12.
    Litjens, G., et al.: Evaluation of prostate segmentation algorithms for MRI: The PROMISE12 challenge. Med. Imag. Anal. 18(2), 359–373 (2014)CrossRefGoogle Scholar
  13. 13.
    Mahapatra, D., Buhmann, J.: Prostate mri segmentation using learned semantic knowledge and graph cuts. IEEE Trans. Biomed. Engg. 61(3), 756–764 (2014)CrossRefGoogle Scholar
  14. 14.
    Malmberg, F., et. al.: Smart paint a new interactive segmentation method applied to MR prostate segmentation. In: Proc. MICCAI PROMISE 2012, pp. 4–11 (2012)Google Scholar
  15. 15.
    Saffari, A., Leistner, C., Santner, J., Godec, M., Bischof., H.: On-line random forests. In: IEEE ICCV Workshops, pp. 1393–1400 (2009)Google Scholar
  16. 16.
    Settles, B., Craven, M.: An analysis of active learning strategies for sequence labeling tasks. In: Empirical Methods in Natural Language Processing, pp. 1070–1079 (2008)Google Scholar
  17. 17.
    Vincent, G., et. al.: Fully automatic segmentation of the prostate using active appearance models. In: Proc. MICCAI PROMISE 2012, pp. 75–81 (2012)Google Scholar

Copyright information

© Springer International Publishing Switzerland 2015

Open Access This chapter is licensed under the terms of the Creative Commons Attribution-NonCommercial 2.5 International License (, which permits any noncommercial use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license and indicate if changes were made.

The images or other third party material in this chapter are included in the chapter's Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the chapter's Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder.

Authors and Affiliations

  1. 1.Department of Computer ScienceETH ZurichZürichSwitzerland

Personalised recommendations