Abstract
Mutual information (MI) was introduced for use in multimodal image registration over a decade ago [1,2,3,4]. The MI between two images is based on their marginal and joint/conditional entropies. The most common versions of entropy used to compute MI are the Shannon and differential entropies; however, many other definitions of entropy have been proposed as competitors. In this article, we show how to construct normalized versions of MI using any of these definitions of entropy. The resulting similarity measures are analogous to normalized mutual information (NMI), entropy correlation coefficient (ECC), and symmetric uncertainty (SU), which have all been shown to be superior to MI in a variety of situations. We use publicly available CT, PET, and MR brain images with known ground truth transformations to evaluate the performance of the normalized measures for rigid multimodal registration. Results show that for a number of different definitions of entropy, the proposed normalized versions of mutual information provide a statistically significant improvement in target registration error (TRE) over the non-normalized versions.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Maes, F., Collignon, A., Vandermeulen, D., Marchal, G., Suetens, P.: Multimodality Image Registration by Maximization of Mutual Information. IEEE Trans. Med. Imag. 16(2), 187–198 (1997)
Collignon, A.: Multi-modality Medical Image Registration by Maximization of Mutual Information. Ph.D. Thesis, Catholic University of Leuven, Leuven, Belgium (1998)
Viola, P.A.: Alignment by Maximization of Mutual Information. Ph.D. Thesis, Artificial Intelligence Laboratory, Massachusetts Institute of Technology (1995)
Viola, P.A., Wells III, W.M.: Alignment by Maximization of Mutual Information. Int. J. Comp. Vis. 24(2), 137–154 (1997)
Pluim, J.P.W., Maintz, J.B.A., Viergever, M.A.: Mutual-Information-Based Registration of Medical Images: A Survey. IEEE Trans. Med. Imag. 22(8), 986–1004 (2003)
Studholme, C., Hill, D.L.G., Hawkes, D.J.: An Overlap Invariant Entropy Measure of 3D Medical Image Alignment. Patt. Rec. 32, 71–86 (1999)
Melbourne, A., Hawkes, D., Atkinson, D.: Image Registration Using Uncertainty Coefficients. In: Proc. ISBI 2009, pp. 951–954 (2009)
Wachowiak, M.P., Smoliková, R., Tourassi, G.D., Elmaghraby, A.S.: Similarity Metrics Based on Nonadditive Entropies for 2D–3D Multimodal Biomedical Image Registration. In: Sonka, M., Fitzpatrick, J.M. (eds.) Proc. SPIE of Medical Imaging 2003: Image Processing, vol. 5032, pp. 1090–1100 (2003)
Wang, F., Vemuri, B.C., Rao, M., Chen, Y.: A New & Robust Information Theoretic Measure and Its Application to Image Alignment. In: Taylor, C.J., Noble, J.A. (eds.) IPMI 2003. LNCS, vol. 2732, pp. 388–400. Springer, Heidelberg (2003)
Wang, F., Vemuri, B.C.: Non-rigid Multi-modal Image Registration Using Cross-cumulative Residual Entropy. Int. J. Comp. Vis. 74(2), 201–205 (2007)
Liao, S., Chung, A.C.S.: Multi-modal Image Registration Using the Generalized Survival Exponential Entropy. In: Larsen, R., Nielsen, M., Sporring, J. (eds.) MICCAI 2006. LNCS, vol. 4191, pp. 964–971. Springer, Heidelberg (2006)
Cahill, N.D.: Constructing and Solving Variational Image Registration Problems. D.Phil. Thesis, Department of Engineering Science, University of Oxford (2009)
Cahill, N.D., Schnabel, J.A., Noble, J.A., Hawkes, D.J.: Overlap Invariance of Cumulative Residual Entropy Measures for Multimodal Image Alignment. In: Pluim, J.P.W., Dawant, B.M. (eds.) Proc. SPIE of Medical Imaging 2009: Image Processing, vol. 7259 (2009)
Shannon, C.E.: A Mathematical Theory of Communication. Bell Sys. Tech. J. 27, 379–423, 623–656 (1948)
Rényi, A.: On measures of entropy and information. In: Proc. 4th Berkeley Symp. Mathematical Statistics and Probability, Berkeley, CA, vol. I, pp. 547–561 (1961)
Tsallis, C.: Possible Generalizations of Boltzmann-Gibbs Statistics. J. Stat. Phys. 52, 479–487 (1988)
Campbell, L.L.: Exponential Entropy as a Measure of Extent of a Distribution. Zeitschr. für Wahrsch. und verw. Geb. 5, 217–255 (1966)
Koski, T., Persson, L.-E.: Some Properties of Generalized Exponential Entropies with Applications to Data Compression. Inf. Sci. 62(1-2), 103–132 (1992)
Rao, M., Chen, Y., Vemuri, B.C., Wang, F.: Cumulative Residual Entropy: A New Measure of Information. IEEE Trans. Inf. Th. 50(6), 1220–1228 (2004)
Zografos, K., Nadarajah, S.: Survival Exponential Entropies. IEEE Trans. Inf. Th. 51(3), 1239–1246 (2005)
Drissi, N., Chonavel, T., Boucher, J.M.: Generalized Cumulative Residual Entropy for Distributions with Unrestricted Supports. Res. Let. Sig. Proc. Article ID 790607, 2008 (2008)
Astola, J., Virtanen, I.: A Measure of Overall Statistical Dependence Based on the Entropy Concept. In: Proc. Univ. Vaasa. No. 91 (1983)
Witten, I.H., Frank, E.: Data Mining: Practical Machine Learning Tools and Techniques. Morgan Kaufmann, Amsterdam (2005)
West, J., et al.: Comparison and Evaluation of Retrospective Intermodality Brain Image Registration Techniques. J. Comp. Assist. Tom. 21(4), 554–566 (1997)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2010 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Cahill, N.D. (2010). Normalized Measures of Mutual Information with General Definitions of Entropy for Multimodal Image Registration. In: Fischer, B., Dawant, B.M., Lorenz, C. (eds) Biomedical Image Registration. WBIR 2010. Lecture Notes in Computer Science, vol 6204. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-14366-3_23
Download citation
DOI: https://doi.org/10.1007/978-3-642-14366-3_23
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-14365-6
Online ISBN: 978-3-642-14366-3
eBook Packages: Computer ScienceComputer Science (R0)