High-Throughput Prostate Cancer Gland Detection, Segmentation, and Classification from Digitized Needle Core Biopsies
- 807 Downloads
We present a high-throughput computer-aided system for the segmentation and classification of glands in high resolution digitized images of needle core biopsy samples of the prostate. It will allow for rapid and accurate identification of suspicious regions on these samples. The system includes the following three modules: 1) a hierarchical frequency weighted mean shift normalized cut (HNCut) for initial detection of glands; 2) a geodesic active contour (GAC) model for gland segmentation; and 3) a diffeomorphic based similarity (DBS) feature extraction for classification of glands as benign or cancerous. HNCut is a minimally supervised color based detection scheme that combines the frequency weighted mean shift and normalized cuts algorithms to detect the lumen region of candidate glands. A GAC model, initialized using the results of HNCut, uses a color gradient based edge detection function for accurate gland segmentation. Lastly, DBS features are a set of morphometric features derived from the nonlinear dimensionality reduction of a dissimilarity metric between shape models. The system integrates these modules to enable the rapid detection, segmentation, and classification of glands on prostate biopsy images. Across 23 H & E stained prostate studies of whole-slides, 105 regions of interests (ROIs) were selected for the evaluation of segmentation and classification. The segmentation results were evaluated on 10 ROIs and compared to manual segmentation in terms of mean distance (2.6 ±0.2 pixels), overlap (62±0.07%), sensitivity (85±0.01%), specificity (94±0.003%) and positive predictive value (68±0.08%). Over 105 ROIs, the classification accuracy for glands automatically segmented was (82.5 ±9.10%) while the accuracy for glands manually segmented was (82.89 ±3.97%); no statistically significant differences were identified between the classification results.
KeywordsHigh-throughput geodesic active contour model morphological feature prostate cancer glands needle biopsy digital pathology
Unable to display preview. Download preview PDF.
- 3.Doyle, S., Feldman, M., Tomaszewski, J., Madabhushi, A.: A boosted bayesian multi-resolution classifier for prostate cancer detection from digitized needle biopsies. IEEE Transactions on Biomedical Engineering (in Press)Google Scholar
- 4.Monaco, J.P., Tomaszewski, J.E., Feldman, M.D., Hagemann, I., Moradi, M., Mousavi, P., Boag, A., Davidson, C., Abolmaesumi, P., Madabhushi, A.: High-throughput detection of prostate cancer in histological sections using probabilistic pairwise markov models. Medical Image Analysis 14, 617–629 (2010)CrossRefGoogle Scholar
- 7.Sparks, R., Madabhushi, A.: Novel morphometric based classification via diffeomorphic based shape representation using manifold learning. In: MICCAI 2010 (2010) (in press)Google Scholar
- 9.Janowczyk, A., Chandran, S., Singh, R., Sasaroli, D., Coukos, G., Feldman, M.D., Madabhushi, A.: Hierarchical normalized cuts: Unsupervised segmentation of vascular biomarkers from ovarian cancer tissue microarrays. In: Yang, G.-Z., Hawkes, D., Rueckert, D., Noble, A., Taylor, C. (eds.) MICCAI 2009. LNCS, vol. 5761, pp. 230–238. Springer, Heidelberg (2009)CrossRefGoogle Scholar
- 12.Agner, S., Soman, S., Libfeld, E., McDonald, M., Thomas, K., Englander, S., Rosen, M., Chin, D., Nosher, J., Madabhushi, A.: Textural kinetics: A novel dynamic contrast enhanced (DCE)- MRI feature for breast lesion classification. Journal of Digital Imaging (in press)Google Scholar
- 14.Li, C., Xu, C., Gui, C., Fox, M.D.: Level set evolution without re-initialization: A new variational formulation. In: CVPR, vol. 1, pp. 430–436 (2005)Google Scholar
- 16.Blum, H.: A transformation for extracting new descriptors of shape. In: Models for the Perception of Speech and Visual Form, pp. 367–380. MIT Press, Cambridge (1967)Google Scholar
- 17.Guo, H., Rangarajan, A., Joshi, S.: Diffeomorphic point matching. In: Handbook of Mathematical Models in Computer Vision, pp. 205–219. Springer, US (2005)Google Scholar