1 Introduction

Distance or divergence measures are of key importance in statistics and information theory. Depending upon the nature of the problem, different divergence measures are suitable. A number of measures of divergence that compare two probability distributions have been proposed (see [15, 16, 23, 24, 31, 37] and the references therein). Csiszár [12] introduced the f-divergence functional as follows.

Definition 1.1

Suppose that \(f: \mathbb{R}^{+}\rightarrow (0, \infty )\) is a convex function. Let \(\tilde{\mathbf{r}} = (r_{1}, \ldots , r_{n})\) and \(\tilde{\mathbf{s}} = (s_{1}, \ldots , s_{n})\) be such that \(\sum_{k=1}^{n}r_{k}=1\) and \(\sum_{k=1}^{n}s_{k}=1\). Then an f-divergence functional is stated as

$$\begin{aligned} I_{f}(\tilde{\mathbf{r}}, \tilde{\mathbf{s}}) := \sum _{k=1}^{n}s_{k} f \biggl(\frac{r_{k}}{s_{k}} \biggr), \end{aligned}$$

where f bears the following requirements:

$$\begin{aligned} f(0) := \lim_{\epsilon \rightarrow 0^{+}}f(\epsilon );\qquad 0 f \biggl(\frac{0}{0} \biggr):=0;\qquad 0 f \biggl(\frac{a}{0} \biggr):= \lim_{\epsilon \rightarrow 0^{+}} \epsilon f \biggl( \frac{a}{\epsilon } \biggr),\quad a>0. \end{aligned}$$

The Csiszár’s f-divergence is a broad class of divergences which consists of various divergence measures used in finding out the difference between two probability densities. A significant property of Csiszár’s f-divergence is that several well-known divergence measures can be deduced from this divergence measure by suitable substitutions to the convex function f. In recent years, several researchers have done a considerable work providing various kinds of bounds on the divergences and distances, see e.g. [13, 14, 25, 33]. Jensen’s inequality has an important role in obtaining inequalities for divergence measures. It helps to compute useful upper bounds for several entropic measures used in information theory. In [18], Jain et al. established an information inequality regarding Csiszár f-divergence by utilizing the convexity condition and Jensen’s inequality. This inequality is applied in comparing some well-known divergences which play a significant role in information theory. In [19], Khan et al. obtained new results for the Shannon and Zipf–Mandelbrot entropies. They also computed different bounds for these entropies by using some refinements of the Jensen inequality. In [21], the authors established various inequalities for convex functions and applied them to Csiszár divergence. They also obtained several results for Zipf–Mandelbrot entropy. In [27], Mehmood et al. obtained a new generalized form of cyclic refinements of Jensen’s inequality from convex to higher order convex functions by utilizing Taylor’s formula. They also computed bounds for various notable inequalities utilized in information theory. In [11], Butt et al. used discrete and continuous cyclic refinements of Jensen’s inequality and extended them from convex to higher order convex function by using new Green functions and Abel–Gontscharoff interpolating polynomial. As an application, they established a connection between new entropic bounds for relative, Shannon, and Mandelbrot entropies. In [22], Khan et al. established an elegant refinement of Jensen’s inequality related to two finite sequences. The obtained inequality used to compute bounds for Csiszár divergence, variational distance, Shannon entropy, and Zipf–Mandelbrot entropy. In [29], Pečarić et al. obtained refinements of the integral version of Jensen’s inequality and the Lah–Ribarič inequality and deduced estimates for the integral form of Csiszár divergence and its important particular cases. In [2], Ahmad et al. utilized some results of Jensen’s inequality for convex functions and obtained various estimates for Shannon and generalized Zipf–Mandelbrot entropies. In [10], Butt et al. proved various Jensen–Grüss type inequalities under certain conditions.

The development of the theory of time scales was initiated by Hilger in 1988. The books of Bohner and Peterson [8, 9] related to time scales are compact and resolve a lot of time scales calculus. In the past years, new developments in the theory and applications of dynamic derivatives on time scales emerged. Many results from the continuous case are carried over to the discrete one very easily, but some seem to be completely different. The study on time scales comes to reveal such discrepancies and to make us understand the difference between the two cases. The Jensen inequality has been extended to time scales by Agarwal et al. (see [1, 8]). Various classical inequalities and their converses for isotonic linear functionals on time scales are established in [5]. In [6], Anwar et al. gave the properties and applications of Jensen functionals on time scales for one variable. Further in [7], the authors obtained the Jensen inequality for several variables and deduced Jensen functionals. They also derived properties of Jensen functionals and applied them to generalized means. In recent years, the study of dynamic inequalities on time scales has been considered by several authors, see [1, 28, 30, 32, 36, 39, 40]. In [3], Ansari et al. obtained Shannon type inequalities on an arbitrary time scale. They also deduced bounds of differential entropy on time scale for various distributions. Further in [4], the authors established several inequalities for Csiszár f-divergence among two probability densities on time scales. They also obtained new results for divergence measures in h-discrete calculus and quantum calculus.

Quantum calculus or q-calculus is usually called calculus without limits. In 1910, Jackson [17] described a q-analogue of derivative and integral operator along with their applications. He was the first to establish q-calculus in an organized form. It is important to note that quantum integral inequalities are more significant and constructive than their classical counterparts. It has been primarily for the reason that quantum integral inequalities can interpret the hereditary properties of the fact and technique under consideration. Recently, there has been a rapid development in q-calculus. Consequently, new generalizations of the classical approach of quantum calculus have been proposed and analyzed in various literature works. The concepts of quantum calculus on finite intervals were given by Tariboon and Ntouyas [34, 35], and they obtained certain q-analogues of classical mathematical objects, which motivated numerous researchers to explore the subject in detail. Subsequently, several new results related to the quantum counterpart of classical mathematical results have been established.

2 Preliminaries

An arbitrary nonempty closed subset of the real line is known as time scale \(\mathbb{T} \subset \mathbb{R}\). The subsequent results and definitions are given in [8].

Definition 2.1

Suppose that \(\mathbb{T}\) is a time scale and \(\zeta \in \mathbb{T}\), then the forward, respectively backward, jump operators \(\sigma , \rho : \mathbb{T} \rightarrow \mathbb{T}\) are defined as follows:

$$ \sigma (\zeta ) = \inf \{\nu \in \mathbb{T}: \nu > \zeta \}\quad \text{and} \quad \rho ( \zeta ) = \sup \{\nu \in \mathbb{T}: \nu < \zeta \}. $$

Definition 2.2

Let \(\mathbb{T}\) be a time scale and \(z : \mathbb{T} \rightarrow \mathbb{R}\) be a function, then z is known as rd-continuous or right-dense continuous if its left-sided limits exist (finite) at left-dense points in \(\mathbb{T}\) and it is continuous at right-dense points in \(\mathbb{T}\). The set of rd-continuous functions \(z : \mathbb{T} \rightarrow \mathbb{R}\) is usually denoted by \(C_{rd}\).

Let us introduce the set \(\mathbb{T}^{k}\) as follows:

$$ \mathbb{T}^{k} = \textstyle\begin{cases} \mathbb{T}\backslash (\rho (\sup \mathbb{T}), \sup \mathbb{T}] & \text{if } \sup \mathbb{T} < \infty, \\ \mathbb{T} & \text{if } \sup \mathbb{T} = \infty . \end{cases}$$

Definition 2.3

Consider a function \(z : \mathbb{T} \rightarrow \mathbb{R}\) and \(\zeta \in \mathbb{T}^{k}\). Then we define \(z^{\Delta }(\zeta )\) to be the number (when it exists) with the property that given any \(\epsilon > 0\), there is a neighborhood U of ζ such that

$$ \bigl\vert z\bigl(\sigma (\zeta )\bigr) - z(\nu ) - z^{\Delta }(\zeta ) \bigl(\sigma (\zeta ) - \nu \bigr) \bigr\vert \leq \epsilon \bigl\vert \sigma (\zeta ) - \nu \bigr\vert \quad \text{for all } \nu \in U. $$

In this case, z is said to be delta differentiable at ζ.

For \(\mathbb{T}= \mathbb{R}\), \(z^{\Delta }\) becomes ordinary derivative \(z^{\prime }\), while if \(\mathbb{T} = \mathbb{Z}\), then \(z^{\Delta }\) turns into the usual forward difference operator \(\Delta z(\zeta ) = z(\zeta +1) - z(\zeta )\). If \(\mathbb{T} = \overline{q^{\mathbb{Z}}} = \{q^{n}: n \in \mathbb{Z} \} \bigcup \{0\}\) is the so-called q-difference operator, with \(q > 1\), then

$$ z^{\Delta }(\zeta ) = \frac{z(q\zeta ) - z(\zeta )}{(q-1)\zeta }, \qquad z^{\Delta }(0) = \lim _{\nu \rightarrow 0} \frac{z(\nu ) - z(0)}{\nu }. $$

Theorem 2.1

(Existence of antiderivatives)

Every rd-continuous function has an antiderivative. If \(x_{0} \in \mathbb{T}\), then F is defined by

$$ F(\zeta ):= \int _{x_{0}}^{x}f(\zeta )\Delta \zeta\quad \textit{for } x \in \mathbb{T}^{k} $$

is an antiderivative of f.

For \(\mathbb{T} = \mathbb{R}\), we have \(\int _{a}^{b}f(\zeta )\Delta \zeta = \int _{a}^{b}f(\zeta )\,d\zeta \), and if \(\mathbb{T} = \mathbb{N}\), then \(\int _{a}^{b}f(\zeta )\Delta \zeta = \sum_{\zeta =a}^{b-1}f( \zeta )\), where \(a, b \in \mathbb{T}\) with \(a\leq b\).

In [38], Wong et al. gave the weighted Jensen inequality on time scales which is stated as follows.

Theorem 2.2

Assume that \(I \subset \mathbb{R}\), and let \(r \in C_{rd}([a, b]_{\mathbb{T}}, \mathbb{R})\) be a positive function with

$$ \int _{a}^{b} r(\zeta ) \Delta \zeta > 0, $$

where \(a,b \in \mathbb{T}\). If \(f \in C(I, \mathbb{R})\) is convex and \(g \in C_{rd}([a, b]_{\mathbb{T}}, I)\), then

$$ f \biggl( \frac{\int _{a}^{b} r(\zeta ) g(\zeta ) \Delta \zeta }{\int _{a}^{b} r(\zeta ) \Delta \zeta } \biggr) \leq \frac{\int _{a}^{b} r(\zeta )f(g(\zeta )) \Delta \zeta }{\int _{a}^{b} r(\zeta ) \Delta \zeta }. $$
(1)

When f is a strictly convex function, the inequality sign in (1) is strict.

3 Divergences on time scales

Consider the set of rd-continuous functions on time scale \(\mathbb{T}\) to be

$$ \Omega := \biggl\{ r \in C_{rd}\bigl([a, b]_{\mathbb{T}}, (0, \infty )\bigr), r( \zeta )> 0, \int _{a}^{b} r(\zeta ) \Delta \zeta > 0\biggr\} . $$

In the sequel, we assume that \(r, s \in \Omega \) and the following integrals exist:

$$ R = \int _{a}^{b} r(\zeta ) \Delta \zeta \quad \text{and}\quad S = \int _{a}^{b} s(\zeta ) \Delta \zeta . $$

3.1 Csiszár f-divergence

Csiszár f-divergence on time scale is defined in [4] as follows:

$$ D_{f}(s, r) := \int _{a}^{b}r(\zeta )f \biggl( \frac{s(\zeta )}{r(\zeta )} \biggr)\Delta \zeta , $$
(2)

where f is convex on \((0,\infty )\).

Theorem 3.1

Assume that \(I \subset \mathbb{R}\), and if \(f \in C(I, \mathbb{R})\) is convex, then

$$ R f \biggl(\frac{S}{R} \biggr) \leq D_{f}(s, r), $$
(3)

where \(D_{f}(s, r)\) is given in (2).

Proof

Put \(g(\zeta ) = \frac{s(\zeta )}{r(\zeta )}\) in (1) to get (3). □

Example 3.1

For \(\mathbb{T} = \mathbb{R}\), Theorem 3.1 becomes [20, Theorem 5.2 on p. 10].

Example 3.2

Choose \(\mathbb{T} = h\mathbb{Z}\), \(h > 0\) in Theorem 3.1 to get a lower bound for Csiszár divergence in h-discrete calculus

$$ \sum_{l=\frac{a}{h}}^{\frac{b}{h} - 1}r(lh)h f \biggl( \frac{\sum_{l=\frac{a}{h}}^{\frac{b}{h} - 1}s(lh)h}{\sum_{l=\frac{a}{h}}^{\frac{b}{h} - 1}r(lh)h} \biggr) \leq \sum_{l= \frac{a}{h}}^{\frac{b}{h} - 1}r(lh)h f \biggl(\frac{s(lh)}{r(lh)} \biggr). $$

Remark 3.1

Choose \(h = 1\) in Example 3.2, and let \(a = 0\), \(b = n\), \(r(l) = r_{j}\), and \(s(l) = s_{j}\) to get the discrete Csiszár divergence

$$ \sum_{j = 1}^{n}r_{j} f \biggl( \frac{\sum_{j = 1}^{n}s_{j}}{\sum_{j = 1}^{n}r_{j}} \biggr) \leq I_{f}(s, r), $$
(4)

where

$$ I_{f}(\tilde{\mathbf{s}}, \tilde{\mathbf{r}}) = \sum _{j = 1}^{n}r_{j} f \biggl( \frac{s_{j}}{r_{j}} \biggr), $$
(5)

\(\tilde{\mathbf{s}} = (s_{1}, \ldots , s_{n})\) and \(\tilde{\mathbf{r}} = (r_{1}, \ldots , r_{n})\).

Example 3.3

Choose \(\mathbb{T} = q^{\mathbb{N}_{0}}\) (\(q > 1\)) in Theorem 3.1 to have a new lower bound of the Csiszár divergence in quantum calculus

$$ \sum_{l = 0}^{n - 1} q^{l+1}r\bigl(q^{l}\bigr) f \biggl( \frac{\sum_{l = 0}^{n - 1} q^{l+1}s(q^{l})}{\sum_{l = 0}^{n - 1} q^{l+1}r(q^{l})} \biggr) \leq \sum_{l = 0}^{n - 1} q^{l+1}r\bigl(q^{l}\bigr)f \biggl(\frac{s(q^{l})}{r(q^{l})} \biggr). $$

3.2 Differential entropy (continuous entropy)

Consider a positive density function r on time scale \(\mathbb{T}\) to a continuous random variable X with \(\int _{a}^{b} r(\zeta )\Delta \zeta = 1\), wherever the integral exists.

In [3], Ansari et al. defined the so-called differential entropy on time scale by

$$ h_{\bar{b}}(X) := \int _{a}^{b} r(\zeta ) \log \frac{1}{r(\zeta )} \Delta \zeta , $$
(6)

where \(\bar{b}>1\) is the base of log. In the sequel, we assume that the base of log is greater than 1.

Theorem 3.2

Suppose that \(r, s \in C_{rd}([a, b]_{\mathbb{T}}, \mathbb{R})\) are Δ-integrable functions and r is a positive probability density function with \(S = \int _{a}^{b} s(\zeta ) \Delta \zeta > 0\). If \(f \in C(I, \mathbb{R})\) is convex and \(\bar{b}>1\), then

$$ h_{\bar{b}}(X) \leq \int _{a}^{b} r(\zeta )\log \frac{1}{s(\zeta )} \Delta \zeta + \log (S), $$
(7)

where \(h_{\bar{b}}(\zeta )\) is defined in (6) and \(a,b \in \mathbb{T}\).

Proof

The function \(f(\zeta ) = - \log \zeta \) is convex. Use \(f(\zeta ) = - \log \zeta \) with \(\int _{a}^{b} r(\zeta ) \Delta \zeta = 1\) in (3) to get

$$\begin{aligned} -\log (S) \leq & \int _{a}^{b} -r(\zeta )\log \biggl( \frac{s(\zeta )}{r(\zeta )} \biggr) \Delta \zeta , \\ =& \int _{a}^{b} \bigl(r(\zeta )\log r(\zeta ) - r( \zeta )\log s(\zeta )\bigr) \Delta \zeta , \\ =& \int _{a}^{b} r(\zeta )\log r(\zeta )\Delta \zeta - \int _{a}^{b}r( \zeta )\log s(\zeta ) \Delta \zeta \\ =&- \int _{a}^{b} r(\zeta )\log \frac{1}{r(\zeta )}\Delta \zeta + \int _{a}^{b}r(\zeta )\log \frac{1}{s(\zeta )}\Delta \zeta , \\ =& -h_{\bar{b}}(X) + \int _{a}^{b}r(\zeta )\log \frac{1}{s(\zeta )} \Delta \zeta , \end{aligned}$$

the stated result. □

Remark 3.2

The inequality in (7) holds in the opposite direction for the base of log less than 1.

Example 3.4

For \(\mathbb{T} = \mathbb{R}\), Theorem 3.2 becomes [26, Theorem 21(a)].

Example 3.5

Choose \(\mathbb{T} = h\mathbb{Z}\), \(h > 0\) in Theorem 3.2 to get an upper bound for entropy in h-discrete calculus

$$ \sum_{l=\frac{a}{h}}^{\frac{b}{h} - 1}r(lh)h \log \biggl( \frac{1}{r(lh)h} \biggr) \leq \sum_{l=\frac{a}{h}}^{\frac{b}{h} - 1}r(lh)h \log \biggl(\frac{1}{s(lh)h} \biggr) + \log \Biggl(\sum _{l=\frac{a}{h}}^{ \frac{b}{h} - 1}s(lh)h \Biggr). $$
(8)

Remark 3.3

Put \(h = 1\) in (8) to get [26, Theorem 8 (i)].

Example 3.6

Choose \(\mathbb{T} = q^{\mathbb{N}_{0}}\) (\(q > 1\)) in Theorem 3.2 to have

$$ \sum_{l = 0}^{n - 1} q^{l+1}r\bigl(q^{l}\bigr)\log \biggl(\frac{1}{r(q^{l})} \biggr) \leq \sum_{l = 0}^{n - 1} q^{l+1}r\bigl(q^{l}\bigr)\log \biggl( \frac{1}{s(q^{l})} \biggr) + \log \Biggl(\sum_{l = 0}^{n - 1} q^{l+1}s\bigl(q^{l}\bigr) \Biggr). $$
(9)

Remark 3.4

(9) contains Shannon entropy which is new in quantum calculus up to the knowledge of authors.

3.3 Karl Pearson \(\chi ^{2}\)-divergence

The \(\chi ^{2}\)-divergence on time scale is defined in [4] as follows:

$$ D_{\chi ^{2}}(s, r) := \int _{a}^{b}r(\zeta ) \biggl[ \biggl( \frac{s(\zeta )}{r(\zeta )} \biggr)^{2} - 1 \biggr]\Delta \zeta . $$
(10)

Theorem 3.3

Assume the conditions of Theorem 3.1to get

$$ \frac{1}{R} \bigl[S^{2} - R^{2} \bigr] \leq D_{\chi ^{2}}(s, r), $$
(11)

where \(D_{\chi ^{2}}(s, r)\) is defined in (10).

Proof

Consider \(f(\zeta ) = \zeta ^{2} - 1\) in (3) to obtain

$$ \biggl(\frac{S}{R} \biggr)^{2} - 1 \leq \frac{1}{R} \int _{a}^{b} r( \zeta ) \biggl[ \biggl( \frac{s(\zeta )}{r(\zeta )} \biggr)^{2} - 1 \biggr] \Delta \zeta , $$

after simplification we get

$$ S^{2} - R^{2} \leq R \int _{a}^{b} r(\zeta ) \biggl[ \biggl( \frac{s(\zeta )}{r(\zeta )} \biggr)^{2} - 1 \biggr] \Delta \zeta , $$

the desired result. □

Example 3.7

If \(\mathbb{T} = \mathbb{R}\), then (11) takes the form

$$ \frac{1}{\int _{a}^{b} r(\zeta )\,d\zeta } \biggl[ \biggl( \int _{a}^{b} s( \zeta )\,d\zeta \biggr)^{2} - \biggl( \int _{a}^{b} r(\zeta )\,d\zeta \biggr)^{2} \biggr] \leq \int _{a}^{b}r(\zeta ) \biggl[ \biggl( \frac{s(\zeta )}{r(\zeta )} \biggr)^{2} - 1 \biggr]\,d\zeta . $$

Example 3.8

Choose \(\mathbb{T} = h\mathbb{Z}\), \(h > 0\) in Theorem 3.3 to get a new lower bound for \(\chi ^{2}\)-divergence in h-discrete calculus

$$ \frac{1}{\sum_{l=\frac{a}{h}}^{\frac{b}{h} - 1}r(lh)h} \Biggl[ \Biggl( \sum _{l=\frac{a}{h}}^{\frac{b}{h} - 1}s(lh)h \Biggr)^{2} - \Biggl( \sum_{l=\frac{a}{h}}^{\frac{b}{h} - 1}r(lh)h \Biggr)^{2} \Biggr] \leq \sum_{l=\frac{a}{h}}^{\frac{b}{h} - 1}r(lh)h \biggl[ \biggl( \frac{s(lh)}{r(lh)} \biggr)^{2}-1 \biggr]. $$
(12)

Remark 3.5

Choose \(h = 1\) in (12), let \(a = 0\), \(b = n\), \(r(l) = r_{j}\), and \(s(l) = s_{j}\) to get \(\chi ^{2}\)-divergence

$$ \frac{1}{\sum_{j = 1}^{n}r_{j}} \Biggl[ \Biggl(\sum_{j = 1}^{n}s_{j} \Biggr)^{2} - \Biggl(\sum_{j = 1}^{n}r_{j} \Biggr)^{2} \Biggr] \leq \chi ^{2}(\tilde{\mathbf{s}}, \tilde{\mathbf{r}}), $$

where

$$ \chi ^{2}(\tilde{\mathbf{s}}, \tilde{\mathbf{r}}) = \sum_{j = 1}^{n}r_{j} \biggl[ \biggl(\frac{s_{j}}{r_{j}} \biggr)^{2}-1 \biggr]. $$
(13)

Example 3.9

Choose \(\mathbb{T} = q^{\mathbb{N}_{0}}\) (\(q > 1\)) in Theorem 3.3 to have a new lower bound for \(\chi ^{2}\)-divergence in quantum calculus

$$\begin{aligned}& \frac{1}{\sum_{l = 0}^{n - 1} q^{l+1}r(q^{l})} \Biggl[ \Biggl(\sum _{l = 0}^{n - 1} q^{l+1}s\bigl(q^{l} \bigr) \Biggr)^{2} - \Biggl(\sum_{l = 0}^{n - 1} q^{l+1}r\bigl(q^{l}\bigr) \Biggr)^{2} \Biggr] \\& \quad \leq \sum_{l = 0}^{n - 1} q^{l+1}r \bigl(q^{l}\bigr) \biggl[ \biggl( \frac{s(q^{l})}{r(q^{l})} \biggr)^{2} - 1 \biggr]. \end{aligned}$$
(14)

3.4 Kullback–Leibler divergence

Kullback–Leibler divergence on time scale is defined in [4] as follows:

$$ D(s, r) = \int _{a}^{b} s(\zeta ) \ln \biggl[ \frac{s(\zeta )}{r(\zeta )} \biggr]\Delta \zeta . $$
(15)

Theorem 3.4

Assume the conditions of Theorem 3.1, then we have

$$ S\ln \biggl(\frac{S}{R} \biggr) \leq D(s, r), $$
(16)

where \(D(s, r) \) is defined in (15).

Proof

Consider \(f(\zeta ) = \zeta \ln \zeta \) in (3) to get

$$ \frac{S}{R}\ln \biggl(\frac{S}{R} \biggr) \leq \frac{1}{R} \int _{a}^{b} s( \zeta )\ln \biggl( \frac{s(\zeta )}{r(\zeta )} \biggr) \Delta \zeta , $$

or we have

$$ S\ln \biggl(\frac{S}{R} \biggr) \leq \int _{a}^{b} s(\zeta )\ln \biggl( \frac{s(\zeta )}{r(\zeta )} \biggr) \Delta \zeta , $$

the desired result. □

Example 3.10

For \(\mathbb{T} = \mathbb{R}\), (16) becomes

$$ \int _{a}^{b} s(\zeta )\,d\zeta \ln \biggl( \frac{\int _{a}^{b} s(\zeta )\,d\zeta }{\int _{a}^{b} r(\zeta )\,d\zeta } \biggr) \leq \int _{a}^{b} s(\zeta )\ln \biggl( \frac{s(\zeta )}{r(\zeta )} \biggr)\,d\zeta . $$

Example 3.11

Choose \(\mathbb{T} = h\mathbb{Z}\), \(h > 0\) in Theorem 3.4 to get a new lower bound in h-discrete calculus

$$ \sum_{l=\frac{a}{h}}^{\frac{b}{h} - 1}s(lh)h \ln \biggl( \frac{\sum_{l=\frac{a}{h}}^{\frac{b}{h} - 1}s(lh)h }{\sum_{l=\frac{a}{h}}^{\frac{b}{h} - 1}r(lh)h} \biggr) \leq \sum_{l= \frac{a}{h}}^{\frac{b}{h} - 1}s(lh)h \ln \biggl(\frac{s(lh)}{r(lh)} \biggr). $$
(17)

Remark 3.6

Choose \(h = 1\) in (17), let \(a = 0\), \(b = n\), \(r(l) = r_{j}\), and \(s(l) = s_{j}\) to get the discrete Kullback–Leibler divergence

$$ \sum_{j = 1}^{n}s_{j} \ln \biggl( \frac{\sum_{j = 1}^{n}s_{j}}{\sum_{j = 1}^{n}r_{j}} \biggr) \leq \operatorname{KL}(\tilde{\mathbf{s}}, \tilde{ \mathbf{r}}), $$

where

$$ \operatorname{KL}(\tilde{\mathbf{s}}, \tilde{\mathbf{r}}) = \sum _{j = 1}^{n}s_{j} \ln \biggl( \frac{s_{j}}{r_{j}} \biggr). $$
(18)

Example 3.12

Choose \(\mathbb{T} = q^{\mathbb{N}_{0}}\) (\(q > 1\)) in Theorem 3.4 to have a new lower bound in quantum calculus

$$ \sum_{l = 0}^{n - 1} q^{l+1}s\bigl(q^{l}\bigr) \ln \biggl( \frac{\sum_{l = 0}^{n - 1} q^{l+1}s(q^{l})}{\sum_{l = 0}^{n - 1} q^{l+1}r(q^{l})} \biggr) \leq \sum_{l = 0}^{n - 1} q^{l+1}s\bigl(q^{l}\bigr) \ln \biggl(\frac{s(q^{l})}{r(q^{l})} \biggr). $$

3.5 Hellinger discrimination

Hellinger discrimination on time scale is defined in [4] as follows:

$$ h^{2}(s, r) = \frac{1}{2} \int _{a}^{b} \bigl[\sqrt{s(\zeta )} - \sqrt{r( \zeta )} \bigr]^{2}\Delta \zeta . $$
(19)

Theorem 3.5

Assume the conditions of Theorem 3.1to obtain

$$ \frac{1}{2}(\sqrt{S} - \sqrt{R})^{2} \leq h^{2}(s, r), $$
(20)

where \(h^{2}(s, r)\) is defined in (19).

Proof

Consider \(f(\zeta ) = \frac{1}{2}(\sqrt{\zeta } - 1)^{2}\) in (3) to get

$$ \frac{1}{2} \biggl(\sqrt{\frac{S}{R}} - 1 \biggr)^{2} \leq \frac{1}{2R} \int _{a}^{b} r(\zeta ) \biggl(\sqrt{ \frac{s(\zeta )}{r(\zeta )}} - 1 \biggr)^{2} \Delta \zeta , $$
(21)

after simplification we obtain

$$ \frac{1}{2}(\sqrt{S} - \sqrt{R})^{2} \leq \frac{1}{2} \int _{a}^{b} \bigl(\sqrt{s(\zeta )} - \sqrt{r(\zeta )} \bigr)^{2} \Delta \zeta , $$

the desired result. □

Example 3.13

For \(\mathbb{T} = \mathbb{R}\), (20) becomes

$$ \frac{1}{2} \biggl( \biggl[ \int _{a}^{b} s(\zeta )\,d\zeta \biggr]^{ \frac{1}{2}} - \biggl[ \int _{a}^{b} r(\zeta )\,d\zeta \biggr]^{ \frac{1}{2}} \biggr)^{2} \leq \frac{1}{2} \int _{a}^{b} \bigl(\sqrt{s( \zeta )} - \sqrt{r( \zeta )} \bigr)^{2}\,d\zeta . $$

Example 3.14

Choose \(\mathbb{T} = h\mathbb{Z}\), \(h > 0\) in Theorem 3.5 to get a new lower for Hellinger discrimination in h-discrete calculus

$$ \frac{1}{2} \Biggl[ \Biggl(\sum _{l=\frac{a}{h}}^{\frac{b}{h} - 1}s(lh)h \Biggr)^{\frac{1}{2}} - \Biggl(\sum _{l=\frac{a}{h}}^{\frac{b}{h} - 1}r(lh)h \Biggr)^{\frac{1}{2}} \Biggr]^{2} \leq \frac{1}{2}\sum_{l=\frac{a}{h}}^{ \frac{b}{h} - 1} \bigl(\sqrt{s(lh)h} - \sqrt{r(lh)h} \bigr)^{2}. $$
(22)

Remark 3.7

Choose \(h = 1\) in (22), let \(a = 0\), \(b = n\), \(r(l) = r_{j}\), and \(s(l) = s_{j}\) to get the Hellinger distance

$$ \frac{1}{2} \Biggl[ \Biggl(\sum_{j = 1}^{n}s_{j} \Biggr)^{\frac{1}{2}} - \Biggl(\sum_{j = 1}^{n}r_{j} \Biggr)^{\frac{1}{2}} \Biggr]^{2} \leq h^{2}( \tilde{ \mathbf{s}}, \tilde{\mathbf{r}}), $$

where

$$ h^{2}(\tilde{\mathbf{s}}, \tilde{\mathbf{r}}) = \frac{1}{2}\sum_{j = 1}^{n} ( \sqrt{s_{j}} - \sqrt{r_{j}} )^{2}. $$
(23)

Example 3.15

Choose \(\mathbb{T} = q^{\mathbb{N}_{0}}\) (\(q > 1\)) in Theorem 3.5 to have a new lower for Hellinger discrimination in quantum calculus

$$\begin{aligned}& \frac{1}{2} \Biggl[ \Biggl(\sum_{l = 0}^{n - 1} q^{l+1}s\bigl(q^{l}\bigr) \Biggr)^{\frac{1}{2}} - \Biggl( \sum_{l = 0}^{n - 1} q^{l+1}r \bigl(q^{l}\bigr) \Biggr)^{\frac{1}{2}} \Biggr]^{2} \\& \quad \leq \frac{1}{2}\sum_{k=0}^{n - 1}q^{l+1} \bigl[\sqrt{s\bigl(q^{l}\bigr)} - \sqrt{r\bigl(q^{l}\bigr)} \bigr]^{2} . \end{aligned}$$
(24)

3.6 Bhattacharyya coefficient

The Bhattacharyya coefficient on time scale is defined in [4] as follows:

$$ D_{B}(s, r) = \int _{a}^{b} \sqrt{r(\zeta )s(\zeta )}\Delta \zeta . $$
(25)

Theorem 3.6

Assume the conditions of Theorem 3.1to get

$$ D_{B}(s, r) \leq \sqrt{RS}, $$
(26)

where \(D_{B}(s, r)\) is defined in (25).

Proof

Consider \(f(\zeta ) = -\sqrt{\zeta }\) in (3) to get

$$ -\sqrt{\frac{S}{R}} \leq \frac{-1}{R} \int _{a}^{b} \sqrt{r(\zeta )s( \zeta )} \Delta \zeta , $$

after simplification we obtain

$$ \int _{a}^{b} \sqrt{r(\zeta )s(\zeta )} \Delta \zeta \leq \sqrt{RS}, $$

the desired result. □

Example 3.16

If \(\mathbb{T} = \mathbb{R}\), then (26) takes the form

$$ \int _{a}^{b} \bigl(r(\zeta )s(\zeta ) \bigr)^{\frac{1}{2}}\,d\zeta \leq \biggl( \int _{a}^{b} r(\zeta )\,d\zeta \int _{a}^{b} s(\zeta )\,d\zeta \biggr)^{\frac{1}{2}}. $$

Example 3.17

Choose \(\mathbb{T} = h\mathbb{Z}\), \(h > 0\) in Theorem 3.6 to get a new upper bound for the Bhattacharyya coefficient in h-discrete calculus

$$ \sum_{l=\frac{a}{h}}^{\frac{b}{h} - 1} \bigl(r(lh)h s(lh)h \bigr)^{ \frac{1}{2}} \leq \Biggl(\sum_{l=\frac{a}{h}}^{\frac{b}{h} - 1}r(lh)h \sum_{l=\frac{a}{h}}^{\frac{b}{h} - 1}s(lh)h \Biggr)^{\frac{1}{2}}. $$
(27)

Remark 3.8

Choose \(h = 1\) in (27), let \(a = 0\), \(b = n\), \(r(l) = r_{j}\), and \(s(l) = s_{j}\) to get the Bhattacharyya coefficient

$$ B(\tilde{\mathbf{s}}, \tilde{\mathbf{r}}) \leq \Biggl(\sum _{j = 1}^{n}r_{j} \sum _{j = 1}^{n}s_{j} \Biggr)^{\frac{1}{2}}, $$

where

$$ B(\tilde{\mathbf{s}}, \tilde{\mathbf{r}}) = \sum _{j = 1}^{n}\sqrt{r_{j} s_{j}}. $$
(28)

Example 3.18

Choose \(\mathbb{T} = q^{\mathbb{N}_{0}}\) (\(q > 1\)) in Theorem 3.6 to have a new upper bound for the Bhattacharyya coefficient in quantum calculus

$$ \sum_{l = 0}^{n - 1} q^{l+1} \bigl[r\bigl(q^{l}\bigr) s\bigl(q^{l} \bigr) \bigr]^{ \frac{1}{2}} \leq \Biggl(\sum_{l = 0}^{n - 1} q^{l+1}r\bigl(q^{l}\bigr) \sum_{l = 0}^{n - 1} q^{l+1}s\bigl(q^{l}\bigr) \Biggr)^{\frac{1}{2}}. $$

3.7 Jeffreys distance

Jeffreys distance on time scale is defined in [4] as follows:

$$ D_{J}(s, r) = \int _{a}^{b} \bigl(s(\zeta ) - r(\zeta )\bigr) \ln \biggl[ \frac{s(\zeta )}{r(\zeta )} \biggr]\Delta \zeta . $$
(29)

Theorem 3.7

Assume the conditions of Theorem 3.1to get

$$ (S - R) \ln \biggl(\frac{S}{R} \biggr) \leq D_{J}(s, r), $$
(30)

where \(D_{J}(s, r)\) is defined in (29).

Proof

Consider \(f(\zeta ) = (\zeta - 1)\ln \zeta \) in (3) to get

$$ R \biggl(\frac{S}{R}-1 \biggr)\ln \biggl(\frac{S}{R} \biggr) \leq \int _{a}^{b} r(\zeta ) \biggl(\frac{s(\zeta )}{r(\zeta )} - 1 \biggr) \ln \biggl( \frac{s(\zeta )}{r(\zeta )} \biggr) \Delta \zeta , $$

or we have

$$ (S - R)\ln \biggl(\frac{S}{R} \biggr) \leq \int _{a}^{b} \bigl(s(\zeta ) - r( \zeta )\bigr) \ln \biggl(\frac{s(\zeta )}{r(\zeta )} \biggr) \Delta \zeta , $$

the desired result. □

Example 3.19

For \(\mathbb{T} = \mathbb{R}\), (30) takes the form

$$ \biggl( \int _{a}^{b} s(\zeta )\,d\zeta - \int _{a}^{b} s(\zeta )\,d\zeta \biggr) \ln \biggl( \frac{\int _{a}^{b} s(\zeta )\,d\zeta }{\int _{a}^{b} r(\zeta )\,d\zeta } \biggr) \leq \int _{a}^{b} \bigl[s(\zeta ) - r(\zeta )\bigr]\ln \biggl( \frac{s(\zeta )}{r(\zeta )} \biggr)\,d\zeta . $$

Example 3.20

Choose \(\mathbb{T} = h\mathbb{Z}\), \(h > 0\) in Theorem 3.7 to get a new lower bound for Jeffreys distance in h-discrete calculus

$$\begin{aligned}& \Biggl(\sum_{l=\frac{a}{h}}^{\frac{b}{h} - 1}s(lh)h - \sum_{l= \frac{a}{h}}^{\frac{b}{h} - 1}r(lh)h \Biggr) \ln \biggl( \frac{\sum_{l=\frac{a}{h}}^{\frac{b}{h} - 1}s(lh)h }{\sum_{l=\frac{a}{h}}^{\frac{b}{h} - 1}r(lh)h} \biggr) \\& \quad \leq \sum_{l=\frac{a}{h}}^{\frac{b}{h} - 1}\bigl(s(lh)h - r(lh)h\bigr) \ln \biggl(\frac{s(lh)}{r(lh)} \biggr). \end{aligned}$$
(31)

Remark 3.9

Choose \(h = 1\) in (31), let \(a = 0\), \(b = n\), \(r(l) = r_{j}\), and \(s(l) = s_{j}\) to get Jeffreys distance

$$ \Biggl(\sum_{j = 1}^{n}s_{j} - \sum_{j = 1}^{n}r_{j} \Biggr) \ln \biggl(\frac{\sum_{j = 1}^{n}s_{j} }{\sum_{j = 1}^{n}r_{j}} \biggr) \leq D_{J}(\tilde{\mathbf{s}}, \tilde{\mathbf{r}}), $$

where

$$ D_{J}(\tilde{\mathbf{s}}, \tilde{\mathbf{r}}) = \sum _{j = 1}^{n}(s_{j} - r_{j}) \ln \biggl(\frac{s_{j}}{r_{j}} \biggr). $$
(32)

Example 3.21

Choose \(\mathbb{T} = q^{\mathbb{N}_{0}}\) (\(q > 1\)) in Theorem 3.7 to have a new lower bound for the Jeffreys distance in quantum calculus

$$\begin{aligned}& \Biggl(\sum_{l = 0}^{n - 1} q^{l+1}s\bigl(q^{l}\bigr) - \sum _{l = 0}^{n - 1} q^{l+1}r\bigl(q^{l} \bigr) \Biggr) \ln \biggl( \frac{\sum_{l = 0}^{n - 1} q^{l+1}s(q^{l})}{\sum_{l = 0}^{n - 1} q^{l+1}r(q^{l})} \biggr) \\& \quad \leq \sum_{l = 0}^{n - 1} q^{l+1} \bigl(s\bigl(q^{l}\bigr) - r\bigl(q^{l}\bigr)\bigr) \ln \biggl( \frac{s(q^{l})}{r(q^{l})} \biggr). \end{aligned}$$

3.8 Triangular discrimination

Triangular discrimination on time scale is defined in [4] as follows:

$$ D_{\Delta }(r, s) = \int _{a}^{b} \frac{[s(\zeta ) - r(\zeta )]^{2}}{s(\zeta ) + r(\zeta )}\Delta \zeta . $$
(33)

Theorem 3.8

Assume the conditions of Theorem 3.1to obtain

$$ \frac{[S - R]^{2}}{S + R} \leq D_{\Delta }(r, s), $$
(34)

where \(D_{\Delta }(r, s)\) is defined in (33).

Proof

Consider \(f(\zeta ) = \frac{(\zeta - 1)^{2}}{\zeta + 1}\) in (3) to get

$$ R \frac{ (\frac{S}{R} - 1 )^{2}}{\frac{S}{R} + 1} \leq \int _{a}^{b} r(\zeta ) \frac{ (\frac{s(\zeta )}{r(\zeta )} - 1 )^{2}}{\frac{s(\zeta )}{r(\zeta )} + 1} \Delta \zeta $$

or

$$ \frac{[S - R]^{2}}{S + R} \leq \int _{a}^{b} \frac{[s(\zeta ) - r(\zeta )]^{2}}{s(\zeta ) + r(\zeta )} \Delta \zeta . $$

 □

Example 3.22

For \(\mathbb{T} = \mathbb{R}\), (34) becomes

$$ \frac{ [\int _{a}^{b} s(\zeta )\,d\zeta - \int _{a}^{b} r(\zeta )\,d\zeta ]^{2}}{\int _{a}^{b} s(\zeta )\,d\zeta + \int _{a}^{b} r(\zeta )\,d\zeta } \leq \int _{a}^{b} \frac{[s(\zeta ) - r(\zeta )]^{2}}{s(\zeta ) + r(\zeta )}\,d\zeta . $$

Example 3.23

Choose \(\mathbb{T} = h\mathbb{Z}\), \(h > 0\) in Theorem 3.8 to get a new lower bound for the triangular discrimination in h-discrete calculus

$$ \frac{ (\sum_{l=\frac{a}{h}}^{\frac{b}{h} - 1}s(lh)h - \sum_{l=\frac{a}{h}}^{\frac{b}{h} - 1}r(lh)h )^{2}}{\sum_{l=\frac{a}{h}}^{\frac{b}{h} - 1}s(lh)h + \sum_{l=\frac{a}{h}}^{\frac{b}{h} - 1}r(lh)h} \leq \sum_{l= \frac{a}{h}}^{\frac{b}{h} - 1} \frac{h(s(lh) - r(lh))^{2}}{s(lh) + r(lh)}. $$
(35)

Remark 3.10

Choose \(h = 1\) in (35), let \(a = 0\), \(b = n\), \(r(l) = r_{j}\), and \(s(l) = s_{j}\) to get the triangular discrimination

$$ \frac{ (\sum_{j = 1}^{n}s_{j} - \sum_{j = 1}^{n}r_{j} )^{2}}{\sum_{j = 1}^{n}s_{j} + \sum_{j = 1}^{n}r_{j}} \leq \Delta ( \tilde{\mathbf{s}}, \tilde{\mathbf{r}}), $$

where

$$ \Delta (\tilde{\mathbf{s}}, \tilde{\mathbf{r}}) = \sum _{j = 1}^{n} \frac{(s_{j} - r_{j})^{2}}{s_{j} + r_{j}}. $$
(36)

Example 3.24

Choose \(\mathbb{T} = q^{\mathbb{N}_{0}}\) (\(q > 1\)) in Theorem 3.8 to have a new lower bound for the triangular discrimination in quantum calculus

$$ \frac{ (\sum_{l = 0}^{n - 1} q^{l+1}s(q^{l}) - \sum_{l = 0}^{n - 1} q^{l+1}r(q^{l}) )^{2}}{\sum_{l = 0}^{n - 1} q^{l+1}s(q^{l}) + \sum_{l = 0}^{n - 1} q^{l+1}r(q^{l})} \leq \sum_{l = 0}^{n - 1} q^{l+1} \frac{[s(q^{l}) - r(q^{l})]^{2}}{s(q^{l}) + r(q^{l})}. $$

4 Zipf–Mandelbrot law

The Zipf–Mandelbrot law is a discrete probability distribution and is defined via a probability mass function which is given as follows:

$$ f(j; N, a, b)=\frac{1}{(j+b)^{a}H_{N, a, b}}, \quad j=1, \ldots , N, $$
(37)

where

$$ H_{N, a, b}=\sum_{i=1}^{N} \frac{1}{(i+b)^{a}} $$
(38)

is a generalization of a harmonic number and \(N \in \{1, 2, \dots \}\), \(a > 0\) and \(b \in [0, \infty )\) are parameters.

If \(b = 0\) and N is finite, then the Zipf–Mandelbrot law is commonly known as the Zipf law. By expression (37), the probability mass function in connection with the Zipf law is

$$ f(j; N, a)=\frac{1}{(j)^{a}H_{N, a}}, \quad j=1, \ldots , N, $$
(39)

where

$$ H_{N, a}=\sum_{i=1}^{N} \frac{1}{(i)^{a}}. $$
(40)

Using \(r_{j}=f(j, N, a, b)\) in (37) as a probability mass function, we observe the obtained results via the Zipf–Mandelbrot law.

For this reason, we give results concerning the Csiszár functional \(\tilde{I}_{f}(\tilde{\mathbf{s}}, \tilde{\mathbf{r}})\) for the Zipf–Mandelbrot law.

  1. Case-1

    Define \(\tilde{\mathbf{r}}\) by (37) as a Zipf–Mandelbrot law N-tuple, Csiszár functional (5) becomes

    $$ \tilde{I}_{f}(j, N, a_{2}, b_{2}, \tilde{\mathbf{s}})=\sum_{j=1}^{N} \frac{1}{(j+b_{2})^{a_{2}}H_{N, a_{2}, b_{2}}}f\bigl(s_{j}(j+b_{2})^{a_{2}}H_{N, a_{2}, b_{2}} \bigr), $$
    (41)

    where \(f:I \rightarrow \mathbb{R}\), \(I \subset \mathbb{R}\), and \(N \in \mathbb{N}\), \(a_{2}>0\), \(b_{2}>0\) are such that \(s_{j}(j+b_{2})^{a_{2}}H_{N, a_{2}, b_{2}} \in I\), \(j=1, \ldots, N\).

  2. Case-2

    When \(\tilde{\mathbf{s}}\) and \(\tilde{\mathbf{r}}\) both are defined via the Zipf–Mandelbrot law for N-tuples:

    $$ \tilde{I}_{f}(j, N, a_{1}, a_{2}, b_{1}, b_{2})=\sum _{j=1}^{N} \frac{1}{(j+b_{2})^{a_{2}}H_{N, a_{2}, b_{2}}} f \biggl( \frac{(j+b_{2})^{a_{2}}H_{N, a_{2}, b_{2}}}{(j+b_{1})^{a_{1}}H_{N, a_{1}, b_{1}}} \biggr), $$
    (42)

    where \(f:I \rightarrow \mathbb{R}\), \(I \subset \mathbb{R}\), and \(N \in \mathbb{N}\), \(a_{1}, a_{2}>0\), \(b_{1},b_{2}>0\) are such that \(\frac{(j+b_{2})^{a_{2}}H_{N, a_{2}, b_{2}}}{(j+b_{1})^{a_{1}}H_{N, a_{1}, b_{1}}} \in I\), \(j=1, \ldots, N\).

  3. Case-3

    If \(\tilde{\mathbf{s}}\) and \(\tilde{\mathbf{r}}\) both are defined as the Zipf law for N-tuples, then Csiszár functional (5) becomes

    $$ \tilde{I}_{f}(j, N, a_{1}, a_{2})=\sum_{j=1}^{N} \frac{1}{j^{a_{2}}H_{N, a_{2}}} f \biggl(j^{a_{2}-a_{1}} \frac{H_{N, a_{2}}}{H_{N, a_{1}}} \biggr). $$
    (43)

Start from case-1 which is for the single Zipf–Mandelbrot law \(r_{j}, j=1, \ldots, N\).

Corollary 4.1

Assume that \(I \subset \mathbb{R}\), and let \(N \in \mathbb{N}\), \(a_{2}>0\), \(b_{2}>0\) be such that \(\sum_{j=1}^{N}s_{j}(j+b_{2})^{a_{2}}H_{N, a_{2}, b_{2}} \in I\) for \(j=1, \ldots , N\). If f is a convex function, then

$$ \sum_{j = 1}^{n} \frac{1}{(j+b_{2})^{a_{2}}H_{N, a_{2}, b_{2}}} f \biggl( \frac{\sum_{j = 1}^{n}s_{j}}{\sum_{j = 1}^{n}\frac{1}{(j+b_{2})^{a_{2}}H_{N, a_{2}, b_{2}}}} \biggr) \leq \tilde{I}_{f}(j, N, a_{2}, b_{2}, \tilde{\mathbf{s}}). $$
(44)

Proof

Put \(r_{j} = \frac{1}{(j+b_{2})^{a_{2}}H_{N, a_{2}, b_{2}}}\) for \(j=1, \ldots, N\) in (4) to get (44), where \(\tilde{I}_{f}(j, N, a_{2}, b_{2}, \tilde{\mathbf{s}})\) is defined in (41). □

Remark 4.1

The inequality sign in (44) holds in reverse direction when f is a concave function.

The next result is for case-2 as both \(s_{j}\) and \(r_{j}\) are defined by the Zipf–Mandelbrot law.

Corollary 4.2

Assume that \(I \subset \mathbb{R}\), and let \(N \in \mathbb{N}\), \(a_{1}, a_{2}>0\), \(b_{1}, b_{2}>0\) be such that \(\frac{(j+b_{2})^{a_{2}}H_{N, a_{2}, b_{2}}}{\sum_{j=1}^{N}(j+b_{1})^{a_{1}}H_{N, a_{1}, b_{1}}} \in I\) for \(j=1, \ldots, N\). If f is a convex function, then

$$ \sum_{j = 1}^{n} \frac{1}{(j+b_{2})^{a_{2}}H_{N, a_{2}, b_{2}}} f \biggl( \frac{\sum_{j = 1}^{n}\frac{1}{(j+b_{1})^{a_{1}}H_{N, a_{1}, b_{1}}}}{\sum_{j = 1}^{n}\frac{1}{(j+b_{2})^{a_{2}}H_{N, a_{2}, b_{2}}}} \biggr) \leq \tilde{I}_{f}(j, N, a_{1}, a_{2}, b_{1}, b_{2}). $$
(45)

Proof

Using \(r_{j} = \frac{1}{(j+b_{2})^{a_{2}}H_{N, a_{2}, b_{2}}}\) and \(s_{j} = \frac{1}{(j+b_{1})^{a_{1}}H_{N, a_{1}, b_{1}}}\) for \(j=1, \ldots, N\), in (4), we get (45), where \(\tilde{I}_{f}(j, N, a_{1}, a_{2}, b_{1}, b_{2})\) is defined in (42). □

Remark 4.2

The inequality sign in (45) holds in reverse direction when f is a concave function.

The next result is for case-3 as both \(s_{j}\) and \(r_{j}\) are defined by the Zipf law.

Corollary 4.3

Assume that \(I \subset \mathbb{R}\), and let \(N \in \mathbb{N}\), \(a_{1}, a_{2}>0\) be such that \(\frac{(j)^{a_{2}}H_{N, a_{2}}}{(j)^{a_{1}}H_{N, a_{1}}} \in I\) for \(j=1, \ldots, N\). If f is a convex function, then

$$ \sum_{j = 1}^{n} \frac{1}{(j)^{a_{2}}H_{N, a_{2}}} f \biggl( \frac{\sum_{j = 1}^{n}\frac{1}{(j)^{a_{1}}H_{N, a_{1}}}}{\sum_{j = 1}^{n}\frac{1}{(j)^{a_{2}}H_{N, a_{2}}}} \biggr) \leq \tilde{I}_{f}(j, N, a_{1}, a_{2}). $$
(46)

Proof

Using \(r_{j} = \frac{1}{(j)^{a_{2}}H_{N, a_{2}}}\) and \(s_{j} = \frac{1}{(j)^{a_{1}}H_{N, a_{1}}}\) for \(j=1, \ldots, N\) in (4), we get (46), where \(\tilde{I}_{f}(j, N, a_{1}, a_{2})\) is defined in (43). □

Remark 4.3

The inequality sign in (46) holds in reverse direction when f is a concave function.

To give certain results related to the particular cases of f-divergences, we begin with the well-known Kullback–Leibler divergence (18).

Corollary 4.4

Let \(N \in \mathbb{N}\) and \(a_{2}>0\), \(b_{2}>0\). Then

$$ \sum_{j = 1}^{n}s_{j} \ln \biggl( \frac{\sum_{j = 1}^{n}s_{j}}{\sum_{j = 1}^{n}\frac{1}{(j+b_{2})^{a_{2}}H_{N, a_{2}, b_{2}}}} \biggr) \leq \tilde{\operatorname{KL}}(j, N, a_{2}, b_{2}, \tilde{\mathbf{s}}). $$
(47)

Proof

The function \(f(\zeta ) = \zeta \ln (\zeta )\) is convex. Use \(f(\zeta )=\zeta \ln (\zeta )\) in (44) to obtain (47), where

$$ \tilde{\operatorname{KL}}(j, N, a_{2}, b_{2}, \tilde{ \mathbf{s}}) =\sum_{j=1}^{N}s_{j} \ln \bigl(s_{j}(j+b_{2})^{a_{2}}H_{N, a_{2}, b_{2}} \bigr). $$

 □

If \(s_{j}\) and \(r_{j}\) are defined by the Zipf–Mandelbrot law.

Corollary 4.5

Suppose that \(N \in \mathbb{N}\) and \(a_{1}, a_{2}>0\), \(b_{1}, b_{2}>0\). Then

$$ \sum_{j = 1}^{n} \frac{1}{(j+b_{1})^{a_{1}}H_{N, a_{1}, b_{1}}} \ln \biggl( \frac{\sum_{j = 1}^{n}\frac{1}{(j+b_{1})^{a_{1}}H_{N, a_{1}, b_{1}}}}{\sum_{j = 1}^{n}\frac{1}{(j+b_{2})^{a_{2}}H_{N, a_{2}, b_{2}}}} \biggr) \leq \tilde{ \operatorname{KL}}(j, N, a_{1}, a_{2}, b_{1}, b_{2}). $$
(48)

Proof

The function \(f(\zeta ) = \zeta \ln (\zeta )\) is convex. Use \(f(\zeta ) = \zeta \ln (\zeta )\) in (45) to get (48), where

$$ \tilde{\operatorname{KL}}(j, N, a_{1}, a_{2}, b_{1}, b_{2})=\sum_{j=1}^{N} \frac{1}{(j+b_{1})^{a_{1}}H_{N, a_{1}, b_{1}}} \ln \biggl( \frac{(j+b_{2})^{a_{2}}H_{N, a_{2}, b_{2}}}{(j+b_{1})^{a_{1}}H_{N, a_{1}, b_{1}}} \biggr). $$

 □

The following result holds as both \(s_{j}\) and \(r_{j}\) are defined by the Zipf law.

Corollary 4.6

Let \(N \in \mathbb{N}\), \(a_{1}, a_{2} > 0\). Then

$$ \sum_{j = 1}^{n} \frac{1}{(j)^{a_{1}}H_{N, a_{1}}} \ln \biggl( \frac{\sum_{j = 1}^{n}\frac{1}{(j)^{a_{1}}H_{N, a_{1}}}}{\sum_{j = 1}^{n}\frac{1}{(j)^{a_{2}}H_{N, a_{2}}}} \biggr) \leq \tilde{ \operatorname{KL}}(j, N, a_{1}, a_{2}). $$
(49)

Proof

The function \(f(\zeta ) = \zeta \ln (\zeta )\) is convex. Use \(f(\zeta ) = \zeta \ln (\zeta )\) in (46) to have (49), where

$$ \tilde{\operatorname{KL}}(j, N, a_{1}, a_{2})=\sum _{j=1}^{N} \frac{1}{j^{a_{1}}H_{N, a_{1}}} \ln \biggl(j^{a_{2}-a_{1}} \frac{H_{N, a_{2}}}{H_{N, a_{1}}} \biggr). $$

 □

Analogous results for the Hellinger distance (23) are given as follows.

Corollary 4.7

Let \(N \in \mathbb{N}\), \(a_{2}>0\), \(b_{2}>0\). Then

$$ \frac{1}{2} \Biggl[ \Biggl(\sum _{j = 1}^{n}s_{j} \Biggr)^{\frac{1}{2}} - \Biggl(\sum_{j = 1}^{n}\frac{1}{(j+b_{2})^{a_{2}}H_{N, a_{2}, b_{2}}} \Biggr)^{\frac{1}{2}} \Biggr]^{2} \leq \tilde{h}^{2}(j, N, a_{2}, b_{2}, \tilde{\mathbf{s}}). $$
(50)

Proof

Since \(f(\zeta )=\frac{1}{2}(\sqrt{\zeta }-1)^{2}\) is a convex function, therefore we use \(f(\zeta )=\frac{1}{2}(\sqrt{\zeta }-1)^{2}\) in (44) to get (50), where

$$ \tilde{h}^{2}(j, N, a_{2}, b_{2}, \tilde{ \mathbf{s}})=\sum_{j=1}^{N} \frac{1}{2} \biggl[ (s_{j} )^{\frac{1}{2}} - \biggl( \frac{1}{(j+b_{2})^{a_{2}}H_{N, a_{2}, b_{2}}} \biggr)^{\frac{1}{2}} \biggr]^{2}. $$

 □

The following result holds as both \(s_{j}\) and \(r_{j}\) are defined by the Zipf–Mandelbrot law.

Corollary 4.8

Let \(N \in \mathbb{N}\), \(a_{1}, a_{2}>0\), \(b_{1}, b_{2}>0\). Then

$$\begin{aligned}& \frac{1}{2} \Biggl[ \Biggl(\sum_{j = 1}^{n} \frac{1}{(j+b_{1})^{a_{1}}H_{N, a_{1}, b_{1}}} \Biggr)^{\frac{1}{2}} - \Biggl(\sum _{j = 1}^{n}\frac{1}{(j+b_{2})^{a_{2}}H_{N, a_{2}, b_{2}}} \Biggr)^{\frac{1}{2}} \Biggr]^{2} \\& \quad \leq \tilde{h}^{2}(j, N, a_{1}, a_{2}, b_{1}, b_{2}) . \end{aligned}$$
(51)

Proof

Since \(f(\zeta )=\frac{1}{2}(\sqrt{\zeta }-1)^{2}\) is a convex function, therefore we use \(f(\zeta )=\frac{1}{2}(\sqrt{\zeta }-1)^{2}\) in (45) to get (51), where

$$ \tilde{h}^{2}(j, N, a_{1}, a_{2}, b_{1}, b_{2})=\frac{1}{2}\sum _{j=1}^{N} \biggl[ \biggl(\frac{1}{(j+b_{1})^{a_{1}}H_{N, a_{1}, b_{1}}} \biggr)^{ \frac{1}{2}} - \biggl(\frac{1}{(j+b_{2})^{a_{2}}H_{N, a_{2}, b_{2}}} \biggr)^{\frac{1}{2}} \biggr]^{2}. $$

 □

The following result holds as both \(s_{j}\) and \(r_{j}\) are defined by the Zipf law.

Corollary 4.9

Let \(N \in \mathbb{N}\), \(a_{1}, a_{2}>0\). Then

$$ \frac{1}{2} \Biggl[ \Biggl(\sum _{j = 1}^{n} \frac{1}{(j)^{a_{1}}H_{N, a_{1}}} \Biggr)^{\frac{1}{2}} - \Biggl(\sum_{j = 1}^{n}\frac{1}{(j)^{a_{2}}H_{N, a_{2}}} \Biggr)^{\frac{1}{2}} \Biggr]^{2} \leq \tilde{h}^{2}(j, N, a_{1}, a_{2}). $$
(52)

Proof

Since \(f(\zeta )=\frac{1}{2}(\sqrt{\zeta }-1)^{2}\) is a convex function, therefore we use \(f(\zeta )=\frac{1}{2}(\sqrt{\zeta }-1)^{2}\) in (46) to get (52), where

$$ \tilde{h}^{2}(j, N, a_{1}, a_{2})=\sum _{j=1}^{N}\frac{1}{2} \biggl[ \biggl( \frac{1}{(j)^{a_{1}}H_{N, a_{1}}} \biggr)^{\frac{1}{2}}- \biggl( \frac{1}{(j)^{a_{2}}H_{N, a_{2}}} \biggr)^{\frac{1}{2}} \biggr]^{2}. $$

 □

Similarly, corresponding results for the Karl Pearson divergence (13) and the Jeffrey distance (32) are given below.

Corollary 4.10

Let \(N \in \mathbb{N}\) and \(a_{2}>0\), \(b_{2}>0\). Then

$$\begin{aligned}& \frac{1}{\sum_{j = 1}^{n}\frac{1}{(j+b_{2})^{a_{2}}H_{N, a_{2}, b_{2}}}} \Biggl[ \Biggl(\sum _{j = 1}^{n}s_{j} \Biggr)^{2} - \Biggl(\sum_{j = 1}^{n} \frac{1}{(j+b_{2})^{a_{2}}H_{N, a_{2}, b_{2}}} \Biggr)^{2} \Biggr] \\& \quad \leq \tilde{\chi ^{2}}(j, N, a_{2}, b_{2}, \tilde{\mathbf{s}}). \end{aligned}$$
(53)

Proof

Since \(f(\zeta )= \zeta ^{2} - 1\) is a convex function, therefore we use \(f(\zeta ) = \zeta ^{2} - 1\) in (44) to obtain (53), where

$$ \tilde{\chi ^{2}}(j, N, a_{2}, b_{2}, \tilde{ \mathbf{s}})=\sum_{j=1}^{N} \frac{1}{(j+b_{2})^{a_{2}}H_{N, a_{2}, b_{2}}} \bigl[\bigl(s_{j}(j+b_{2})^{a_{2}}H_{N, a_{2}, b_{2}} \bigr)^{2} - 1\bigr]. $$

 □

If \(s_{j}\) and \(r_{j}\) are defined via the Zipf–Mandelbrot law.

Corollary 4.11

Suppose that \(N \in \mathbb{N}\) and \(a_{1}, a_{2}>0\), \(b_{1}, b_{2}>0\). Then

$$\begin{aligned}& \frac{1}{\sum_{j = 1}^{n}\frac{1}{(j+b_{2})^{a_{2}}H_{N, a_{2}, b_{2}}}} \Biggl[ \Biggl(\sum_{j = 1}^{n} \frac{1}{(j+b_{1})^{a_{1}}H_{N, a_{1}, b_{1}}} \Biggr)^{2} - \Biggl( \sum _{j = 1}^{n}\frac{1}{(j+b_{2})^{a_{2}}H_{N, a_{2}, b_{2}}} \Biggr)^{2} \Biggr] \\& \quad \leq \tilde{\chi ^{2}}(j, N, a_{1}, a_{2}, b_{1}, b_{2}) . \end{aligned}$$
(54)

Proof

Since \(f(\zeta )= \zeta ^{2} - 1\) is a convex function, therefore we use \(f(\zeta ) = \zeta ^{2} - 1\) in (45) to get (54), where

$$ \tilde{\chi ^{2}}(j, N, a_{1}, a_{2}, b_{1}, b_{2})= \sum_{j=1}^{N} \frac{1}{(j+b_{2})^{a_{2}}H_{N, a_{2}, b_{2}}} \biggl[ \biggl( \frac{(j+b_{2})^{a_{2}}H_{N, a_{2}, b_{2}}}{(j+b_{1})^{a_{1}}H_{N, a_{1}, b_{1}}} \biggr)^{2} - 1 \biggr]. $$

 □

The following result holds as both \(s_{j}\) and \(r_{j}\) are defined by the Zipf law.

Corollary 4.12

Let \(N \in \mathbb{N}\), \(a_{1}, a_{2} > 0\). Then

$$ \frac{1}{\sum_{j = 1}^{n}\frac{1}{(j)^{a_{2}}H_{N, a_{2}}}} \Biggl[ \Biggl(\sum _{j = 1}^{n}\frac{1}{(j)^{a_{1}}H_{N, a_{1}}} \Biggr)^{2} - \Biggl(\sum_{j = 1}^{n}\frac{1}{(j)^{a_{2}}H_{N, a_{2}}} \Biggr)^{2} \Biggr] \leq \tilde{\chi ^{2}}(j, N, a_{1}, a_{2}). $$
(55)

Proof

Since \(f(\zeta )= \zeta ^{2} - 1\) is a convex function, therefore we use \(f(\zeta ) = \zeta ^{2} - 1\) in (46) to have (55), where

$$ \tilde{\chi ^{2}}(j, N, a_{1}, a_{2})=\sum _{j=1}^{N} \frac{1}{j^{a_{2}}H_{N, a_{2}}} \biggl[ \biggl(j^{a_{2}-a_{1}} \frac{H_{N, a_{2}}}{H_{N, a_{1}}} \biggr)^{2} - 1 \biggr]. $$

 □

Corollary 4.13

Let \(N \in \mathbb{N}\) and \(a_{2}>0\), \(b_{2}>0\). Then

$$ \Biggl(\sum_{j = 1}^{n}s_{j} - \sum_{j = 1}^{n} \frac{1}{(j+b_{2})^{a_{2}}H_{N, a_{2}, b_{2}}} \Biggr) \ln \biggl( \frac{\sum_{j = 1}^{n}s_{j}}{\sum_{j = 1}^{n}\frac{1}{(j+b_{2})^{a_{2}}H_{N, a_{2}, b_{2}}}} \biggr) \leq \tilde{D_{J}}(j, N, a_{2}, b_{2}, \tilde{\mathbf{s}}). $$
(56)

Proof

The function \(f(\zeta ) = (\zeta - 1)\ln (\zeta )\) is convex. Use \(f(\zeta )=(\zeta - 1)\ln (\zeta )\) in (44) to obtain (56), where

$$ \tilde{D_{J}}(j, N, a_{2}, b_{2}, \tilde{ \mathbf{s}}) =\sum_{j=1}^{N} \biggl(s_{j} - \frac{1}{(j+b_{2})^{a_{2}}H_{N, a_{2}, b_{2}}} \biggr) \ln \bigl(s_{j}(j+b_{2})^{a_{2}}H_{N, a_{2}, b_{2}} \bigr). $$

 □

If \(s_{j}\) and \(r_{j}\) are defined via the Zipf–Mandelbrot law.

Corollary 4.14

Suppose that \(N \in \mathbb{N}\) and \(a_{1}, a_{2}>0\), \(b_{1}, b_{2}>0\). Then

$$\begin{aligned}& \Biggl(\sum_{j = 1}^{n} \frac{1}{(j+b_{1})^{a_{1}}H_{N, a_{1}, b_{1}}} - \sum_{j = 1}^{n} \frac{1}{(j+b_{2})^{a_{2}}H_{N, a_{2}, b_{2}}} \Biggr) \ln \biggl( \frac{\sum_{j = 1}^{n}\frac{1}{(j+b_{1})^{a_{1}}H_{N, a_{1}, b_{1}}}}{\sum_{j = 1}^{n}\frac{1}{(j+b_{2})^{a_{2}}H_{N, a_{2}, b_{2}}}} \biggr) \\& \quad \leq \tilde{D_{J}}(j, N, a_{1}, a_{2}, b_{1}, b_{2}). \end{aligned}$$
(57)

Proof

The function \(f(\zeta ) = (\zeta - 1)\ln (\zeta )\) is convex. Use \(f(\zeta ) = (\zeta - 1)\ln (\zeta )\) in (45) to get (57), where

$$\begin{aligned} \tilde{D_{J}}(j, N, a_{1}, a_{2}, b_{1}, b_{2}) =& \sum_{j=1}^{N} \biggl(\frac{1}{(j+b_{1})^{a_{1}}H_{N, a_{1}, b_{1}}} - \frac{1}{(j+b_{2})^{a_{2}}H_{N, a_{2}, b_{2}}} \biggr) \\ &{}\times\ln \biggl( \frac{(j+b_{2})^{a_{2}}H_{N, a_{2}, b_{2}}}{(j+b_{1})^{a_{1}}H_{N, a_{1}, b_{1}}} \biggr). \end{aligned}$$

 □

The following result holds as both \(s_{j}\) and \(r_{j}\) are defined by the Zipf law.

Corollary 4.15

Let \(N \in \mathbb{N}\), \(a_{1}, a_{2} > 0\). Then

$$ \Biggl(\sum_{j = 1}^{n} \frac{1}{(j)^{a_{1}}H_{N, a_{1}}} - \sum_{j = 1}^{n} \frac{1}{(j)^{a_{2}}H_{N, a_{2}}} \Biggr) \ln \biggl( \frac{\sum_{j = 1}^{n}\frac{1}{(j)^{a_{1}}H_{N, a_{1}}}}{\sum_{j = 1}^{n}\frac{1}{(j)^{a_{2}}H_{N, a_{2}}}} \biggr) \leq \tilde{D_{J}}(j, N, a_{1}, a_{2}). $$
(58)

Proof

The function \(f(\zeta ) = (\zeta - 1)\ln (\zeta )\) is convex. Use \(f(\zeta ) = (\zeta - 1)\ln (\zeta )\) in (46) to have (58), where

$$ \tilde{D_{J}}(j, N, a_{1}, a_{2}) = \sum _{j=1}^{N} \biggl( \frac{1}{j^{a_{1}}H_{N, a_{1}}} - \frac{1}{j^{a_{2}}H_{N, a_{2}}} \biggr) \ln \biggl(j^{a_{2}-a_{1}}\frac{H_{N, a_{2}}}{H_{N, a_{1}}} \biggr). $$

 □

In addition to all, similar findings for triangular discrimination are given as follows.

Corollary 4.16

Let \(N \in \mathbb{N}\), \(a_{2}>0\), \(b_{2}>0\). Then

$$ \frac{ (\sum_{j = 1}^{n}s_{j} - \sum_{j = 1}^{n}\frac{1}{(j+b_{2})^{a_{2}}H_{N, a_{2}, b_{2}}} )^{2}}{\sum_{j = 1}^{n}s_{j} + \sum_{j = 1}^{n}\frac{1}{(j+b_{2})^{a_{2}}H_{N, a_{2}, b_{2}}}} \leq \tilde{\Delta }(j, N, a_{2}, b_{2}, \tilde{\mathbf{s}}). $$
(59)

Proof

Since \(f(\zeta ) = \frac{(\zeta - 1)^{2}}{\zeta + 1}\) is a convex function, therefore we use \(f(\zeta ) = \frac{(\zeta - 1)^{2}}{\zeta + 1}\) in (44) to obtain (59), where

$$ \tilde{\Delta }(j, N, a_{2}, b_{2}, \tilde{\mathbf{s}})= \sum_{j=1}^{N} \frac{1}{(j+b_{2})^{a_{2}}H_{N, a_{2}, b_{2}}} \frac{[s_{j} (j+b_{2})^{a_{2}}H_{N, a_{2}, b_{2}} - 1]^{2}}{s_{j} (j+b_{2})^{a_{2}}H_{N, a_{2}, b_{2}} + 1}. $$

 □

If \(s_{j}\) and \(r_{j}\) are defined via the Zipf–Mandelbrot law.

Corollary 4.17

Suppose that \(N \in \mathbb{N}\) and \(a_{1}, a_{2}>0\), \(b_{1}, b_{2}>0\). Then

$$ \frac{ (\sum_{j = 1}^{n}\frac{1}{(j+b_{1})^{a_{1}}H_{N, a_{1}, b_{1}}} - \sum_{j = 1}^{n}\frac{1}{(j+b_{2})^{a_{2}}H_{N, a_{2}, b_{2}}} )^{2}}{\sum_{j = 1}^{n}\frac{1}{(j+b_{1})^{a_{1}}H_{N, a_{1}, b_{1}}} + \sum_{j = 1}^{n}\frac{1}{(j+b_{2})^{a_{2}}H_{N, a_{2}, b_{2}}}} \leq \tilde{\Delta }(j, N, a_{1}, a_{2}, b_{1}, b_{2}). $$
(60)

Proof

Since \(f(\zeta )= \frac{(\zeta - 1)^{2}}{\zeta + 1}\) is a convex function, therefore we use \(f(\zeta ) = \frac{(\zeta - 1)^{2}}{\zeta + 1}\) in (45) to get (60), where

$$ \tilde{\Delta }(j, N, a_{1}, a_{2}, b_{1}, b_{2})= \sum_{j=1}^{N} \frac{[(j+b_{2})^{a_{2}}H_{N, a_{2}, b_{2}} - (j+b_{1})^{a_{1}}H_{N, a_{1}, b_{1}}]^{2}}{(j+b_{2})^{a_{2}}H_{N, a_{2}, b_{2}} + (j+b_{1})^{a_{1}}H_{N, a_{1}, b_{1}}}. $$

 □

The following result holds when both \(s_{j}\) and \(r_{j}\) are defined via the Zipf law.

Corollary 4.18

Let \(N \in \mathbb{N}\), \(a_{1}, a_{2} > 0\). Then

$$ \frac{ (\sum_{j = 1}^{n}\frac{1}{(j)^{a_{1}}H_{N, a_{1}}} - \sum_{j = 1}^{n}\frac{1}{(j)^{a_{2}}H_{N, a_{2}}} )^{2}}{\sum_{j = 1}^{n}\frac{1}{(j)^{a_{1}}H_{N, a_{1}}} + \sum_{j = 1}^{n}\frac{1}{(j)^{a_{2}}H_{N, a_{2}}}} \leq \tilde{\Delta }(j, N, a_{1}, a_{2}). $$
(61)

Proof

Since \(f(\zeta )= \frac{(\zeta - 1)^{2}}{\zeta + 1}\) is a convex function, we use \(f(\zeta ) = \frac{(\zeta - 1)^{2}}{\zeta + 1}\) in (46) to have (61), where

$$ \tilde{\Delta }(j, N, a_{1}, a_{2})= \sum _{j=1}^{N} \frac{1}{j^{a_{2}}} \frac{ (j^{a_{2} - a_{1}} H_{N, a_{2}} - H_{N, a_{1}} )^{2}}{j^{a_{2} - a_{1}} H_{N, a_{2}} + H_{N, a_{1}}}. $$

 □