Abstract
We have mentioned the MDL principle on several occasions somewhat loosely as the principle that calls for finding the model and model class with which the data together with the model and model class, respectively, can be encoded with the shortest code length. Actually to apply the principle we must distinguish between two types of models — those for data compression and others for general statistical purposes such as prediction. In data compression, we apply the models to the same data from which the models are determined. Hence these models need not have any predictive power; and, in fact, to get the shortest code length we do not even need to fit models in the class considered, say, \( \mathcal{M}_\gamma \) γ. This is because the universal NML model gives a code length, which we called the stochastic complexity and which we consider to be the shortest for all intents and purposes.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Rights and permissions
Copyright information
© 2007 Springer Science+Business Media, LLC
About this chapter
Cite this chapter
(2007). The MDL Principle. In: Information and Complexity in Statistical Modeling. Information Science and Statistics. Springer, New York, NY. https://doi.org/10.1007/978-0-387-68812-1_8
Download citation
DOI: https://doi.org/10.1007/978-0-387-68812-1_8
Publisher Name: Springer, New York, NY
Print ISBN: 978-0-387-36610-4
Online ISBN: 978-0-387-68812-1
eBook Packages: Computer ScienceComputer Science (R0)