Entropy
Reference work entry
DOI: https://doi.org/10.1007/978-1-4419-9863-7_1554
Synonyms
Definition
Uncertainty (or
Entropy, in the sense of Shannon’s information theory) is a quantitative measure for the uncertainty of the outcome of a random experiment modeled by a probabilistic variable. Given a probabilistic variable
X with outcome set
\( \mathcal{X} \)
This is a preview of subscription content, log in to check access.
References
- Jaynes ET (1957) Information theory and statistical mechanics. Phys Rev 106(4):620–630CrossRefGoogle Scholar
Copyright information
© Springer Science+Business Media, LLC 2013