Abstract
It is known that Shannon’s entropy is nonnegative and its maximum value is reached for equiprobable events. Adding or removing impossible events does not affect Shannon’s entropy. However, if we increase the number of events and consider not necessarily all of them equiprobable, but at least as many of them as the initial number of equiprobable events, how does Shannon’s entropy change? We study the lower bound of the interval where the probability value of the a priori assumed equiprobable states must belong when the entropy increases.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
Shannon’s entropy [9] is defined as
where \(P=(p_{1},\ldots ,p_{n})\) is a finite probability distribution. (Here and elsewhere in this paper, \(\log \) denotes the natural logarithm.) It is nonnegative and its maximum value is \(H(U)=\log n\), where \(U=(1/n,\ldots ,1/n)\). Throughout the paper we use the convention \(0\log 0=0\).
The known recursivity (grouping) property of Shannon’s entropy (see for instance [1, 2]) states that
Apparently the simple question ”how do slight modifications of the probabilities affect the entropy?” does not have many answers in the literature, and we stated in [6] the following open problem.
Open Problem. Find the lower bound (threshold) \(a\left( k\right) \ge 0\) such that, if the probability distribution \(P=\left( p_{1},\ldots ,p_{n}\right) \) has at least k nonzero and equal components \(\ge a\left( k\right) \), then the Shannon entropy \(H\left( P\right) \) attains its minimum when \(n-k\) components of P are zero. In other words, find the best (smallest) \(a\left( k\right) \) such that
for all probability distributions \(P=\left( p_{1},\ldots ,p_{n-k},p,\ldots ,p\right) \in \mathbb {R} _{+}^{n}\) such that \(p>0\) and \(a\left( k\right) \le p\le 1/k\ (k\le n-1)\). Obviously \(a\left( k\right) \le 1/k.\)
2 Main results
Our starting point now is the following answer given in [3], useful for computer assisted analysis of the experimental data.
Proposition 1
Let the probability distribution \(P=\left( p_{1},\ldots ,p_{n}\right) \) be such that it has at least k nonzero and equal components \(p_{n-k+1}= \cdots =p_{n}=p\). The best (smallest) \(a\left( k\right) \ge 0\) such that
for all P for which additionally \(a\left( k\right) \le p\le 1/k\) holds\(,\ \)is the value of the abscisse of the first intersection of the horizontal line \(y=\log (k)\) and the graph of the function
Figure 1 shows these intersections for \(k=1,\ldots ,5.\)
In [3], the proof of this result was reduced to the fact that \(a\left( k\right) \) is given as the smallest solution p of the equation
The maximum of the function \(f_{k}\left( p\right) \) is \(\log \left( k+1\right) .\) Therefore, we are interested in the part of the graph which is in between the horizontal lines \(y=\log (k)\) and \(y=\log (k+1).\) The line \( y=\log (k)\) meets the graph of \(f_{k}\left( p\right) \) twice: one point has as abscisse the required bound \(a\left( k\right) ,\) the other is situated at the right endpoint of the domain of \(f_{k}\left( p\right) ,\) \(p=1/k.\)
In [3] we also provided some particular estimates of interest for \(a\left( k\right) ,\) found with the computer package MATLAB, needed for practical purposes, as in Fig. 1.
In what follows, we look for a nicer formula (however still implicit) of the first solution of the equation (2.2), \(a\left( k\right) .\) As a result, the equation (2.2) is solved also in the case when k is not an integer, and we consider this fact of some theoretical importance.
If \(x:=kp\), equation (2.2) takes the form
Since \(F(k,x)=(1-x)\log k+x\log x+(1-x)\log (1-x)\), (2.3 ) is solvable in k, and the solution is
As a result we obtain \(p=a(k)\) as a function of \(x=kp\):
In Fig. 2 (generated with MATLAB as well) we plot the function \((1-x)x^{\frac{1}{1-x}}\) and the straight lines \(\frac{x}{k}\) for \(k=1,\ldots ,5.\) The intersections correspond to \( a\left( k\right) \) for \(k=1,\ldots ,5.\)
Proposition 2
With the above notation, it holds that
for \(k\ge 2.\)
Proof
It is straightforward to observe, as an immediate consequence of the recursivity of Shannon’s entropy (1.1), that
In the case \(p_{1}+p_{2}=1/(k+1),\) \(p_{3}=\cdots =p_{k+2}=1/(k+1)\) this yields
and we can infer that for all \(n>k\) it holds that
for all positive \(p_{1},\ldots ,p_{n-k}\) such that \(p_{1}+\cdots +p_{n-k}=1/(k+1).\)
Then for \(p=1/(k+1)\) inequality (2.1) holds true, therefore \( a\left( k\right) \le 1/(k+1).\) \(\square \)
Geometrically speaking, this means that the intersection of the graph of the function \((1-x)x^{\frac{1}{1-x}}\) with the straight line \(\frac{x}{k}\) has a lower ordinate than the intersection of the straight line \(\frac{x}{k+1}\) with the vertical line \(x=1.\)
Remark 1
Note that, according to Corollary 2 in [3], one also has
The first equality holds true for \(p_{1}=\cdots =p_{n-k-1}=0,\ p_{n-k}=1-kp,\) the second equality is valid for \(p_{1}=\cdots =p_{n-k}=\frac{1-kp}{n-k}.\) The last equality holds true for \(p=1/n.\) In this paper we studied an alternative way to determine the domain of p such that
Such studies become of practical interest when one uses redistributing algorithms to analyze the time series, as in the papers [4,5,6,7,8].
References
Fadeev, D.K.: On the concept of entropy of a finite probabilistic scheme. Uspekhi Math. Nauk. 11(1), 227–31 (1956)
Khinchin Ya A.: On the concept of entropy in the theory of probabilities. Uspekhi Mat. Nauk VIII 3(55) (1953)
Mitroi-Symeonidis, F.-C., Symeonidis, E.: Redistributing algorithms and Shannon’s Entropy. Aequat. Math. 96, 267–277 (2022)
Mitroi-Symeonidis, F.-C., Anghel, I.: The permutation entropy and the assessment of compartment fire development: growth and decay, (ROMFIN2019). Math. Rep. 23(73), 203–210 (2021)
Mitroi-Symeonidis, F.-C., Anghel, I., Minculete, N.: Parametric Jensen-Shannon statistical complexity and its applications on full-scale compartment fire data. Symmetry (Special Issue: Symmetry in Applied Mathematics) 12(1), 22 (2020)
Mitroi-Symeonidis, F.-C., Anghel, I., Lalu, O., Popa, C.: The permutation entropy and its applications on fire tests data. J. Appl. Comput. Mech. 6(SI), 1380–1393 (2020)
Mitroi-Symeonidis, F.-C., Anghel, I., Furuichi, S.: Encodings for the calculation of the permutation hypoentropy and their applications on full-scale compartment fire data. Acta Technica Napocensis 2(6), 607–616 (2019)
Mitroi-Symeonidis, F.-C., Anghel, I.: The PYR-algorithm for time series modeling of temperature values and its applications on full-scale compartment fire data. In: Acta Technica Napocensis, Series: Applied Mathematics, Mechanics, and Engineering, Vol. 63(IV), pp. 403–410 (2020)
Shannon, C.E.: A mathematical theory of communication. Bell Syst. Tech. J. 27(3), 379–423 (1948)
Author information
Authors and Affiliations
Contributions
All authors have seen and approved the manuscript and have contributed significantly in the paper. There are no conflicts of interest to disclose.
Corresponding author
Ethics declarations
Conflict of interest
The authors declare no Conflict of interest.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Symeonidis, E., Mitroi-Symeonidis, FC. Shannon’s entropy and its bounds for some a priori known equiprobable states. Aequat. Math. (2024). https://doi.org/10.1007/s00010-024-01068-y
Received:
Revised:
Accepted:
Published:
DOI: https://doi.org/10.1007/s00010-024-01068-y