# Review of Univariate Probability

Chapter
Part of the Springer Texts in Statistics book series (STS)

## Abstract

Probability is a universally accepted tool for expressing degrees of confidence or doubt about some proposition in the presence of incomplete information or uncertainty. By convention, probabilities are calibrated on a scale of 0 to 1; assigning something a zero probability amounts to expressing the belief that we consider it impossible, whereas assigning a probability of one amounts to considering it a certainty. Most propositions fall somewhere in between. Probability statements that we make can be based on our past experience, or on our personal judgments. Whether our probability statements are based on past experience or subjective personal judgments, they obey a common set of rules, which we can use to treat probabilities in a mathematical framework, and also for making decisions on predictions, for understanding complex systems, or as intellectual experiments and for entertainment. Probability theory is one of the most applicable branches of mathematics. It is used as the primary tool for analyzing statistical methodologies; it is used routinely in nearly every branch of science, such as biology, astronomy and physics, medicine, economics, chemistry, sociology, ecology, finance, and many others. A background in the theory, models, and applications of probability is almost a part of basic education. That is how important it is.

## Keywords

Cumulative Distribution Function Central Limit Theorem Normal Approximation Sample Space Discrete Random Variable
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.

## References

1. Alon, N. and Spencer, J. (2000). The Probabilistic Method, Wiley, New York.
2. Ash, R. (1972). Real Analysis and Probability, Academic Press, New York.Google Scholar
3. Barbour, A. and Hall, P. (1984). On the rate of Poisson convergence, Math. Proc. Camb. Phil. Soc., 95, 473–480.
4. Bernstein, S. (1927). Theory of Probability, Nauka, Moscow.Google Scholar
5. Bhattacharya, R.N. and Rao, R.R. (1986). Normal Approximation and Asymptotic Expansions, Robert E. Krieger, Melbourne, FL.
6. Bhattacharya, R.N. and Waymire, E. (2009). A Basic Course in Probability Theory, Springer, New York.Google Scholar
7. Billingsley, P.(1995). Probability and Measure, Third Edition, John Wiley, New York.
8. Breiman, L. (1992). Probability, Addison-Wesley, New York.
9. Brown, L., Cai, T., and DasGupta, A. (2001). Interval estimation for a binomial proportion, Statist. Sci., 16, 101–133.
10. Brown, L., Cai, T., and DasGupta, A. (2002). Confidence intervals for a binomial proportion and asymptotic expansions, Ann. Statist., 30, 160–201.
11. Chernoff, H. (1952). A measure of asymptotic efficiency for tests of a hypothesis based on the sum of observations, Ann. Math. Statist., 23, 493–507.
12. Chernoff, H. (1981). A note on an inequality involving the normal distribution, Ann. Prob., 9, 533–535.
13. Chung, K. L. (1974). A Course in Probability, Academic Press, New York.
14. DasGupta, A. (2008). Asymptotic Theory of Statistics and Probability, Springer, New York.
15. DasGupta, A. (2010). Fundamentals of Probability: A First Course, Springer, New York.
16. Diaconis, P. and Zabell, S. (1991). Closed form summation formulae for classical distributions, Statist. Sci., 6, 284–302.
17. Dudley, R. (2002). Real Analysis and Probability, Cambridge University Press, Cambridge, UK.Google Scholar
18. Everitt, B. (1998). Cambridge Dictionary of Statistics, Cambridge University Press, New York.
19. Feller, W. (1968). Introduction to Probability Theory and its Applications, Vol. I, Wiley, New York.
20. Feller, W. (1971). Introduction to Probability Theory and Its Applications, Vol. II, Wiley, New York.
21. Fisher, R.A. (1929). Moments and product moments of sampling distributions, Proc. London Math. Soc., 2, 199–238.Google Scholar
22. Hall, P. (1992). The Bootstrap and Edgeworth Expansion, Springer-Verlag, New York.Google Scholar
23. Johnson, N., Kotz, S., and Balakrishnan, N. (1994). Continuous Univariate Distributions, Vol. I, Wiley, New York.
24. Kagan, A., Linnik, Y., and Rao, C.R. (1973). Characterization Problems in Mathematical Statistics, Wiley, New York.
25. Kendall, M.G. and Stuart, A. (1976). Advanced Theory of Statistics, Vol. I, Wiley, New York.
26. Le Cam, L. (1960). An approximation theorem for the Poisson binomial distribution, Pacific J. Math., 10, 1181–1197.
27. Le Cam, L. (1986). The central limit theorem around 1935, Statist. Sci., 1, 78–96.
28. Paley, R.E. and Zygmund, A. (1932). A note on analytic functions in the unit circle, Proc. Camb. Philos. Soc., 28, 266–272.
29. Petrov, V. (1975). Limit Theorems of Probability Theory, Oxford University Press, Oxford, UK.Google Scholar
30. Pitman, J. (1992). Probability, Springer-Verlag, New York.Google Scholar
31. Rao, C.R. (1973), Linear Statistical Inference and Applications, Wiley, New York.
32. Ross, S. (1984). A First Course in Probability, Macmillan, New York.
33. Steele, J.M. (1994). Le Cam’s inequality and Poisson approximations, Amer. Math Month., 101, 48–54.
34. Stein, C. (1981). Estimation of the mean of a multivariate normal distribution, Ann. Stat., 9, 1135–1151.
35. Stigler, S. (1986). The History of Statistics, Belknap Press, Cambridge, MA.
36. Stirzaker, D. (1994). Elementary Probability, Cambridge University Press, London.
37. Wasserman, L. (2006). All of Nonparametric Statistics, Springer, New York.
38. Widder, D. (1989). Advanced Calculus, Dover, New York.