Abstract
Chapter 3 used the Central Limit Theorem to determine the number of users that can safely share a common cable or link. We saw that this result is also fundamental to calculate confidence intervals. In this section, we prove this theorem. A key tool is the characteristic function that provides a simple way to study sums of independent random variables.
Section 4.1 introduces the characteristic function and calculates it for a Gaussian random variable. Section 4.2 uses that function to prove the Central Limit Theorem. Section 4.3 uses the characteristic function to calculate the moments of a Gaussian random variable. The sum of squares of Gaussian random variables is a common model of noise in communication links. Section 4.4 proves a remarkable property of such a sum. Section 4.5 shows how to use characteristic functions to approximate binomial and geometric random variables. The error function arises in the calculation of the probability of errors in transmission systems and also in decisions based on random observations. Section 4.6 derives useful approximations of that function. Section 4.7 concludes the chapter with a discussion of an adaptive multiple access protocol similar to one used in WiFi networks.
You have full access to this open access chapter, Download chapter PDF
Topics: Characteristic Functions, Proof of Central Limit Theorem, Adaptive CSMA
4.1 Characteristic Functions
Before we explain the proof of the CLT, we have to describe the use of characteristic functions.
Definition 4.1
Characteristic Function The characteristic function of a random variable X is defined as
In this expression, \(i := \sqrt {-1}\). â—‡
Note that
so that Ï• X(u) is the Fourier transform of f X(x). As such, the characteristic function determines the pdf uniquely.
As an important example, we have the following result.
Theorem 4.1 (Characteristic Function of \(\mathcal {N}(0, 1)\))
Let \(X =_D \mathcal {N}(0, 1)\) . Then,
\({\blacksquare }\)
Proof
One has
so that
(The third equation follows by integration by parts.) Thus,
which implies that
Since ϕ X(0) = E(e i0X) = 1, we see that A = 1, and this proves the result (4.1). □
We are now ready to prove the CLT.
4.2 Proof of CLT (Sketch)
The technique to analyze sums of independent random variables is to calculate the characteristic function. Let then
We have
The third equality holds because the X(m) are i.i.d. and the fourth one follows from the Taylor expansion of the exponential:
Thus, the characteristic function of Y (n) converges to that of a \(\mathcal {N}(0, 1)\) random variable. This suggests that the inverse Fourier transform, i.e., the density of Y (n) converges to that of a \(\mathcal {N}(0, 1)\) random variable. This last step can be shown formally, but we will not do it here. \({\square }\)
4.3 Moments of \(\mathcal {N}(0, 1)\)
We can use the characteristic function of a \(\mathcal {N}(0, 1)\) random variable X to calculate its moments. This is how. First we note that, by using the Taylor expansion of the exponential,
Second, again using the expansion of the exponential,
Third, we match the coefficients of u 2m in these two expressions and we find that
This givesFootnote 1
For instance,
Finally, we note that the coefficients of odd powers of u must be zero, so that
(This should be obvious from the symmetry of f X(x).) In particular,
4.4 Sum of Squares of 2 i.i.d. \(\mathcal {N}(0, 1)\)
Let X, Y  be two i.i.d. \(\mathcal {N}(0, 1)\) random variables. The claim is that
Let θ be the angle of the vector (X, Y ) and R 2 = X 2 + Y 2. Thus (see Fig. 4.1)
Note that E(Z) = E(X 2) + E(Y 2) = 2, so that if Z is exponentially distributed, its rate must be 1∕2. Let us prove that it is exponential. One has
where
Thus, the angle θ of (X, Y ) and the norm \(R = \sqrt {X^2 + Y^2}\) are independent and have the indicated distributions. But then, if V = R 2 =: g(R), we find that, for v ≥ 0,
which shows that the angle θ and V = X 2 + Y 2 are independent, the former being uniformly distributed in [0, 2π] and the latter being exponentially distributed with mean 2.
4.5 Two Applications of Characteristic Functions
We have used characteristic functions to prove the CLT. Here are two other cute applications.
4.5.1 Poisson as a Limit of Binomial
A Poisson random variable X with mean λ can be viewed as a limit of a B(n, λ∕n) random variable X n as n →∞. To see this, note that
where the random variables {Z n(1), …, Z n(n)} are i.i.d. Bernoulli with mean λ∕n. Hence,
For the second identity, we use the fact that if Z =D B(p), then
Also, since
we find that
The result then follows from the fact that
4.5.2 Exponential as Limit of Geometric
An exponential random variable can be viewed as a limit of scaled geometric random variables. Let X =D Exp(λ) and X n = G(λ∕n). Then
To see this, recall that
Also,
if the real part of β is positive.
Hence,
Moreover, since
we find that, with \(p = \frac {\lambda }{n}\),
where o(1∕n) → 0 as n →∞. This proves the result.
4.6 Error Function
In the calculation of confidence intervals, one uses estimates of
The function Q(x) is called the error function. With Python or the appropriate smart phone app, you can get the value of Q(x). Nevertheless, the following bounds (see Fig. 4.2) may be useful.
Theorem 4.2 (Bounds on Error Function)
One has
\({\blacksquare }\)
Proof
Here is a derivation of the upper bound. For x > 0, one has
For the lower bound, one uses the following calculation, again with x > 0:
â–¡
4.7 Adaptive Multiple Access
In Sect. 3.4, we explained a randomized multiple access scheme. In this scheme, there are N active station and each station attempts to transmit with probability 1∕N in each time slot. This scheme results in a success rate of about 1∕e ≈ 36%. However, it requires that each station knows how many other stations are active.
To make the scheme adaptive to the number of active devices, say that the devices adjust the probability p(n) with which they transmit at time n as follows:
In these update rules, a and b are constants with a ∈ (0, 1) and b > 1. The idea is to increase p(n) if no device transmitted and to decrease it after a collision. This scheme is due to Hajek and Van Loon (1982) (Fig. 4.3).
Figure 4.4 shows the evolution over time of the success rate T n. Here,
The figure uses a = 0.8 and b = 1.2. We see that the throughput approaches the optimal value for N = 40 and for N = 100. Thus, the scheme adapts automatically to the number of active devices.
4.8 Summary
-
Characteristics Function;
-
Proof of CLT;
-
Moments of Gaussian;
-
Sum of Squares of Gaussians;
-
Poisson as limit of Binomial;
-
Exponential as limit of Geometric;
-
Adaptive Multiple Access Protocol.
4.8.1 Key Equations and Formulas
4.10 Problems
Problem 4.1
Let X be a N(0, 1) random variable. You will recall that E(X 2) = 1 and E(X 4) = 3.
-
(a)
Use Chebyshev’s inequality to get a bound on P(|X| > 2);
-
(b)
Use the inequality that involves the fourth moment of X to bound P(|X| > 2). Do you get a better bound?
-
(c)
Compare with what you know about the N(0, 1) random variable.
Problem 4.2
Write a Python simulation of Hajek’s random multiple access scheme. There are 20 stations. An arrival occurs at each station with probability λ∕20 at each time slot. The stations update their transmission probability as explained in the text. Plot the total backlog in all the stations as a function of time.
Problem 4.3
Consider a multiple access scheme where the N stations independently transmit short reservation packets with duration equal to one time unit with probability p. If the reservation packets collide or no station transmits a reservation packet, the stations try again. Once a reservation is successful, the succeeding station transmits a packet during K time units. After that transmission, the process repeats. Calculate the maximum fraction of time that the channel can be used for transmitting packets. Note: This scheme is called Reservation Aloha.
Problem 4.4
Let X be a random variable with mean zero and variance 1. Show that E(X 4) ≥ 1.
Hint
Use the fact that E((X 2 − 1)2) ≥ 0.
Problem 4.5
Let X, Y  be two random variables. Show that
This is the Cauchy–Schwarz inequality.
Hint
Use E((λX − Y )2) ≥ 0 with λ = E(XY )∕E(X 2).
Notes
- 1.
We used the fact that
$$\displaystyle \begin{aligned} i^{2m} = (-1)^m. \end{aligned}$$
References
D. Bertsekas, J. Tsitsiklis, Introduction to Probability (Athena, Nashua, 2008)
P. Billingsley, Probability and Measure, Third Edition (Wiley, Hoboken, 2012)
G.R. Grimmett, D.R. Stirzaker, Probability and Random Processes (Oxford University Press, Oxford, 2001)
B. Hajek, T. Van Loon, Decentralized dynamic control of a multiple access broadcast channel. IEEE Trans. Autom. Control, AC-27(3), 559–569 (1982)
Author information
Authors and Affiliations
Rights and permissions
Open Access This chapter is licensed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license and indicate if changes were made.
The images or other third party material in this chapter are included in the chapter's Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the chapter's Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder.
Copyright information
© 2021 The Author(s)
About this chapter
Cite this chapter
Walrand, J. (2021). Multiplexing: B. In: Probability in Electrical Engineering and Computer Science. Springer, Cham. https://doi.org/10.1007/978-3-030-49995-2_4
Download citation
DOI: https://doi.org/10.1007/978-3-030-49995-2_4
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-49994-5
Online ISBN: 978-3-030-49995-2
eBook Packages: Computer ScienceComputer Science (R0)