Multi-atlas Based Segmentation Editing with Interaction-Guided Constraints

  • Sang Hyun Park
  • Yaozong Gao
  • Dinggang Shen
Conference paper
Part of the Lecture Notes in Computer Science book series (LNCS, volume 9351)


We propose a novel multi-atlas based segmentation method to address the editing scenario, when given an incomplete segmentation along with a set of training label images. Unlike previous multi-atlas based methods, which depend solely on appearance features, we incorporate interaction-guided constraints to find appropriate training labels and derive their voting weights. Specifically, we divide user interactions, provided on erroneous parts, into multiple local interaction combinations, and then locally search for the training label patches well-matched with each interaction combination and also the previous segmentation. Then, we estimate the new segmentation through the label fusion of selected label patches that have their weights defined with respect to their respective distances to the interactions. Since the label patches are found to be from different combinations in our method, various shape changes can be considered even with limited training labels and few user interactions. Since our method does not need image information or expensive learning steps, it can be conveniently used for most editing problems. To demonstrate the positive performance, we apply our method to editing the segmentation of three challenging data sets: prostate CT, brainstem CT, and hippocampus MR. The results show that our method outperforms the existing editing methods in all three data sets.


Vote Weight Initial Segmentation Training Label Active Shape Model Patch Model 
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.


Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.


  1. 1.
    Boykov, Y., Funka-Lea, G.: Graph cuts and efficient N-D image segmentation. International Journal of Computer Vision 70(2), 109–131 (2006)CrossRefGoogle Scholar
  2. 2.
    Grady, L.: Random walks for image segmentation. IEEE Transactions on Pattern Analysis and Machine Intelligence 28(11), 1768–1783 (2006)CrossRefGoogle Scholar
  3. 3.
    Schwarz, T., Heimann, T., Tetzlaff, R., Rau, A.M., Wolf, I., Meinzer, H.P.: Interactive surface correction for 3D shape based segmentation. In: Proc. SPIE (2008)Google Scholar
  4. 4.
    Park, S.H., Lee, S., Yun, I.D., Lee, S.U.: Structured patch model for a unified automatic and interactive segmentation framework. Medical Image Analysis (2015)Google Scholar
  5. 5.
    Coupe, P., Manjon, J.V., Fonov, V., Pruessner, J., Robles, M., Collins, D.L.: Patch-based segmentation using expert priors: Application to hippocampus and ventricle segmentation. NeuroImage 54(2), 940–954 (2011)CrossRefGoogle Scholar
  6. 6.
    Wu, G., Wang, Q., Zhang, D., Nie, F., Huang, H., Shen, D.: A generative probability model of joint label fusion for multi-atlas based brain segmentation. Medical Image Analysis 18(6), 881–890 (2014)CrossRefGoogle Scholar
  7. 7.
    Besl, P.J., Makay, N.D.: A method of registration of 3-D shapes. IEEE Transactions on Pattern Analysis and Machine Intelligence 14(2), 239–256 (1992)CrossRefGoogle Scholar

Copyright information

© Springer International Publishing Switzerland 2015

Authors and Affiliations

  • Sang Hyun Park
    • 1
  • Yaozong Gao
    • 1
    • 2
  • Dinggang Shen
    • 1
  1. 1.Department of Radiology and BRICUNC at Chapel HillChapel HillUSA
  2. 2.Department of Computer ScienceUNC at Chapel HillChapel HillUSA

Personalised recommendations