1 Introduction

Let \(\{X_n, n\ge 1\}\) be a sequence of identically distributed random variables defined on a fixed probability space \((\Omega , \mathcal {F}, P)\) with a distribution function F. The p-th quantile of F is defined as

$$\begin{aligned} Q_p = \inf \{x: F(x)\ge p\}, \end{aligned}$$

where \(0<p<1\).

Quantiles play an important role in finance, modeling and statistics. In practical applications, quantile estimators are fundamental, which was noticed quite early, e.g. at the paper Galton (1889).

In the literature, there are numerous quantiles estimators. Some quantile estimators were developed for specific distributions, whereas others were designed to be "distribution-free" (in other words—nonparametric estimators), with no assumption about the population density function.

Let \((X_{(1)},X_{(2)},...,X_{(n)})\) be the ordered sample of \((X_1,...,X_n)\) and \(\lfloor x\rfloor \) denotes an integer part of x. Dielman et al. (1994) presented eight different nonparametric estimators of a quantile:

  • Weighted average at \(X_{(\lfloor np+0.5\rfloor )}\)

    $$\begin{aligned} E_1= & {} (0.5+\lfloor np+0.5\rfloor -np)X_{(\lfloor np+0.5\rfloor )} \\{} & {} \quad +(0.5-\lfloor np+0.5\rfloor +np)X_{(\lfloor np+0.5\rfloor +1)}, \end{aligned}$$

    where \(0.5\le np\le (n-0.5)\).

Remark 1

It follows from the condition \(0.5\le np\le (n-0.5)\) that \(\displaystyle n\ge \max \bigg \{\frac{0.5}{1-p},\frac{1}{2p}\bigg \}\).

  • Weighted average at \(X_{(\lfloor np\rfloor )}\)

    $$\begin{aligned} E_2=(1-(np-\lfloor np\rfloor ))X_{(\lfloor np\rfloor )}+(np-\lfloor np\rfloor )X_{(\lfloor np\rfloor +1)}. \end{aligned}$$
  • Lower empirical cumulative distribution function (CDF) value

    $$\begin{aligned} E_3=X_{(\lfloor np\rfloor )}. \end{aligned}$$
  • Upper empirical cumulative distribution function (CDF) value

    $$\begin{aligned} E_4=X_{(\lfloor np\rfloor +1)}. \end{aligned}$$
  • Observation numbered closest to np

    $$\begin{aligned} E_5=\left\{ \begin{array}{ll} X_{(\lfloor np\rfloor )},&{}\text {if}\ \ np-\lfloor np\rfloor <0.5\\ X_{(\lfloor np\rfloor +1)},&{}\text {if} \ \ np-\lfloor np\rfloor \ge 0.5 \end{array}\right. . \end{aligned}$$
  • Empirical cumulative distribution function (CDF)

    $$\begin{aligned} E_6=\left\{ \begin{array}{ll} X_{(\lfloor np\rfloor )},&{}\text {if}\ \ np-\lfloor np\rfloor =0\\ X_{(\lfloor np\rfloor +1)},&{}\text {if} \ \ np-\lfloor np\rfloor > 0 \end{array}\right. . \end{aligned}$$
  • Weighted average at \(X_{(\lfloor (n+1)p\rfloor )}\)

    $$\begin{aligned} E_7= & {} (1-((n+1)p-\lfloor (n+1)p\rfloor ))X_{(\lfloor (n+1)p\rfloor )}\\{} & {} \quad +((n+1)p-\lfloor (n+1)p\rfloor )X_{(\lfloor (n+1)p\rfloor +1)}. \end{aligned}$$
  • Empirical cumulative distribution function (CDF) with averaging

    $$\begin{aligned} E_8=\left\{ \begin{array}{ll} \frac{X_{(\lfloor np\rfloor )}+X_{(\lfloor np\rfloor +1)}}{2},&{}\text {if}\ \ np-\lfloor np\rfloor =0\\ X_{(\lfloor np\rfloor +1)},&{}\text {if} \ \ np-\lfloor np\rfloor > 0 \end{array}\right. . \end{aligned}$$

The most common estimator studied in many papers is \(E_4\). Bahadur (1966) first established an elegant representation for a sample quantile in terms of the empirical distribution function based on independent and identically distributed samples.

Let \(\displaystyle F_n(x) = {1\over n}\sum _{i=1}^n I[X_i\le x]\), \( x\in {\mathbb {R}} \), \(n\ge 1\) be the empirical distribution function for the sample \((X_1, X_2, \dots ,X_ n)\).

Theorem 1

(Bahadur (1966)) Let \(0< p < 1\) and \(\{X_n, n \ge 1\}\) be a sequence of independent identically distributed random variables with the distribution function F. Assume that F has at least two derivatives at some neighborhood of \(Q_p\) and \(F'(Q_p)~=~f(Q_p) > 0\). Then

$$\begin{aligned} E_4 = Q_p - {F_n(Q_p) - p\over f(Q_p)}+ O\bigl (n^{-{3\over 4}} \log n\bigr ) \quad a.s. \end{aligned}$$
(1)

In next years, many researchers have studied the Bahadur representation for sample quantiles for dependent sequences. This is very important problem for practical applications, because in practice we often deal with samples with different dependency structures. Sen (1972); Babu and Singh (1978); Yoshihara (1995); Yang et al. (2019) and Wu et al. (2021) obtained the Bahadur represenation for \(\varphi \)-mixing sequences, Sun (2006); Wang et al. (2011) and Zhang et al. (2014) got the Bahadur representation for \(\alpha \)-mixing sequences and Xing and Yang (2019) for \(\psi \)-mixing sequences. For negatively dependent structures Ling (2008); Xing and Yang (2011) and Xu et al. (2013) considered this problem for negatively associated (NA) sequences and Li et al. (2011) studied it for negatively orthant dependent (NOD) random variables.

Below, we present the definitions of four types of dependence of random variables that will be considered in this work.

Definition 1

(Joag-Dev and Proschan (1983)) A finite family of random variables \(\{X_i, 1\le i\le n\}\) is said to be negatively associated (NA) if for every pair of disjoint subsets A and B of \(\{1,2,...,n\}\), we have

$$\begin{aligned} \text {Cov}(f_1(X_i, i\in A),f_2(X_j, j\in B))\le 0, \end{aligned}$$

wherever \(f_1\) and \(f_2\) are coordinatewise nondecreasing, provided the covariance exists. An infinite family of random variables is said to be NA if every finite subfamily is NA.

Many authors have investigated NA’s statistical properties. For example, Joag-Dev and Proschan (1983) studied NA’s fundamental properties, Yang (2003) investigated uniformly asymptotic normality of regression weighted estimator for NA samples, Liang and Jing (2005) presented asymptotic properties of the estimation of nonparametric regression model based on NA sequences, Liang et al. (2006) studied asymptotic properties of the estimation of semiparametric regression model based on a linear process with NA innovations.

Remark 2

Increasing functions defined on disjoint subsets of a set of NA random variables are NA random variables.

Example 1

(Joag-Dev and Proschan (1983)) Let \({\textbf{Z}}=(Z_1,...,Z_k)\) be a vector having a multinomial distribution, obtained by taking only one observation. Thus only one \(Z_i\) is 1 while the rest are zero. The NA property for \({\textbf{Z}}\) trivially follows from Definition 1. Since the general multinomial is a convolution of independent copies of \({\textbf{Z}}\), the closure property (Remark 2) establishes NA in this case.

Definition 2

(Rozanov and Volkonski (1959)) A sequence of random variables \(\{X_n, n\ge 1\}\) is said to be \(\varphi \)-mixing if

$$\begin{aligned} \varphi (n)=\sup _{m\ge 1}\sup _{A\in {\mathcal {F}}_1^m,B\in {\mathcal {F}}_{m+n}^{\infty },P(A)>0} \big |P(B|A)-P(B)\big |\rightarrow 0, \end{aligned}$$

as \(n\rightarrow \infty \), where \({\mathcal {F}}_n^m=\sigma (X_i, n\le i\le m)\).

\(\varphi \)-mixing property was studied by many researchers. Utev (1990) studied the central limit theorem, Chen et al. (2009) investigated total convergence of the sequences, Yang et al. (2012) obtained the Berry-Esseen bound. The more information on \(\varphi \)-mixing properties one can find in Billingsley (1968) in chapter 4.

Example 2

(Wu et al. (2021)) Let \(\{\epsilon _n, n\ge 1\}\) be a sequence of independent and identically distributed random variables with zero mean and a finite variance. Define

$$\begin{aligned} X_n=\sum _{k=0}^m a_k\epsilon _{n-k} \end{aligned}$$

for some positive integer m and constants \(a_k\), \(k=0,1,...,m\). Then \(\{X_n, n\ge 1\}\) is known as a moving average process with older m. It can be verified that \(\{X_n, n\ge 1\}\) is a \(\varphi \)-mixing process.

Definition 3

(Bradley (1992)) A sequence of random variables \(\{X_n,n\ge 1\}\) is said to be \(\rho ^*\)-mixing, if

$$\begin{aligned} \rho ^*(n)=\sup _{ S,T\subset {\mathbb {N}}, \text {dist}(S,T)\ge n}\{\rho (S,T)\}\rightarrow 0, \end{aligned}$$

as \(n\rightarrow \infty \), where

$$\begin{aligned} \rho (S,T)=\sup _{X\in L_2(\sigma (S)), Y\in L_2(\sigma (T))} \Bigg \{\frac{|Cov(X,Y)|}{\sqrt{Var(X)Var(Y)}}\Bigg \}, \end{aligned}$$

\(\displaystyle \text {dist}(S,T)=\min _{i\in S, j\in T}|j-i|\) and \(\sigma (S)\) and \(\sigma (T)\) are the \(\sigma \)-fields generated by \(\{X_i, i\in S\}\) and \(\{X_j, j\in T\}\), respectively.

\(\rho ^*\)-mixing random variables are a well-described and repeatedly studied structure. Bradley (1992) obtained the central limit theorem, Bryc and Smolenski (1993); Peligrad and Gut (1999); Utev and Peligrad (2003) presented the moment inequalities and Sung (2010) analysed the complete convergance of weighted sums for \(\rho ^*\)-mixing sequences of random variables.

Remark 3

Note that increasing functions defined on a disjoint subset of a \(\rho ^*\)-mixing field \(\{X_k, k\in N^d\}\) with mixing coefficients \(\rho ^*(s)\) are also \(\rho ^*\)-mixing with coefficients not greater that \(\rho ^*(s)\).

Example 3

(Wang et al. (2019)) Let \(\{X_n, n\ge 1\}\) be a strictly stationary, finite-state, irreducible and aperiodic Markov chain. Then it is a \(\rho ^*\)- mixing process with \(\rho ^*(k)=o(e^{-Ck})\) for some \(C>0\).

Definition 4

(Kolmogorov and Rozanov (1960)) A sequence of random variables \(\{X_n, n\ge 1\}\) is said to be \(\rho \)-mixing if

$$\begin{aligned} \rho (n)=\sup _{k\ge 1, X\in L^2({\mathcal {F}}_1^k), Y\in L^2({\mathcal {F}}_{k+n}^{\infty })} \frac{|\text {Cov}(X,Y)|}{\sqrt{\text {Var}(X)\text {Var}(Y)}}\rightarrow 0, \end{aligned}$$

as \(n\rightarrow \infty \), where \({\mathcal {F}}_n^m=\sigma (X_i, n\le i\le m)\) and \(L^2(\cdot )\) is a set of real-valued square-integrable functions.

The \(\rho \)-mixing condition was introduced by Kolmogorov and Rozanov (1960). Shao (1995) in his paper explored the central limit theorem, the law of large numbers and the complete convergence of \(\rho \)-mixing sequences.

Example 4

(Peligrad (1987)) Suppose \(\{Y_k, k\ge 1\}\) and \(\{Z_k, k\ge 1\}\) are independent random variables with the identical standard normal distribution function F and consider the sequence

$$\begin{aligned} X_k^{(\alpha )}=Y_k-Y_{k-1}+\sqrt{\alpha }Z_k. \end{aligned}$$

Because of \(\rho (2)=0\), sequence \(\{X_k^{(\alpha )}, k\ge 1\}\) is \(\rho \)-mixing with \(\displaystyle \sigma _n^2=\text {Var}\bigg (\sum _{k=1}^nX_k^{(\alpha )}\bigg )=2+n\alpha ,\) \(\inf _n \frac{\sigma _n^2}{n}=\alpha \) and \(E(X_k^{(\alpha )})^2=2+\alpha .\)

The following Rosenthal-type inequality will be important in further considerations:

$$\begin{aligned} E\max _{1\le m\le n}\Bigg |\sum _{k=1}^m X_k\Bigg |^q\le C_q\Bigg \{\sum _{k=1}^nE|X_k|^q+\bigg (\sum _{k=1}^nEX_k^2\bigg )^{\frac{q}{2}}\Bigg \}, \end{aligned}$$
(2)

where \(C_q>0\), \(q\ge 2\), \(EX_n=0\) and \(E|X_n|^q<\infty \) for every \(n\ge 1.\)

Remark 4

Inequality (2) is true for NA random variables (Shao and Su (1999)), for \(\rho ^*\)-mixing random variables (Peligrad and Gut (1999)) and for \(\varphi \)-mixing random variables with some additional condition on mixing coefficients \(\varphi (n), n\ge 1\) (Wu et al. (2021)).

For \(\rho \)-mixing random variables, the following inequality plays the same role as (2).

Lemma 1

(Shao (1995)) Let \(q\ge 2\) and \(\{X_n, n\ge 1\}\) be a sequence of \(\rho \)-mixing random variables. Assume that \(EX_i=0\), \(E|X_n|^q<\infty \) and

$$\begin{aligned} \sum _{n=1}^{\infty }\rho ^{\frac{2}{q}}(2^n)<\infty . \end{aligned}$$

Then there exists a positive constant \(K=K(q,\rho (\cdot ))\) depending only on q and \(\rho (\cdot )\) such that for any \(k\ge 0\), \(n\ge 1\),

$$\begin{aligned} E\max _{1\le n}\Bigg |\sum _{i=k+1}^{k+n} X_i\Bigg |^q\le K\Bigg \{n\max _{k<i\le k+n}E|X_i|^q+\bigg (n\max _{k<i\le k+n}EX_i^2\bigg )^{\frac{q}{2}}\Bigg \}. \end{aligned}$$

The first aim of this paper is a comparative analysis of the effectiveness of each of the estimators, depending on the distribution from which the sample is drawn. For this purpose, in section 2 we will compare the fit of the values of the estimators presented above obtained for samples taken from populations with different distributions to the theoretical value. The second aim is to show the strong consistency of each of the estimators and to study that the Bahadur representation holds for each of the estimators when the sample is taken from a NA, \(\varphi \), \(\rho ^*\) or \(\rho \)-mixing population.

2 Comparative analysis of estimators

In the comparative analysis of the estimators well-known probability distributions were used:

  • Normal distribution with \(\mu =0\), \(\sigma =4\),

  • Student’s t distribution with 3 degrees of freedom,

  • Weibull distribution with scale \(\lambda =1\) and shape \(k=5\),

  • Uniform distribution on the interval [0,2],

  • \(\chi ^2\) distribution with 3 degrees of freedom,

  • Exponential distribution with \(\lambda =1\).

The analysis was carried out for the sample sizes: \(n\in \{50,150,555,1130, 2165\}\) and for the different values of \(p\in \{0.025, 0.5, 0.975\}\). Using the R software, in each case the teoretical value of the quantile was generated. Next, samples of sizes \(n\in \{50,150,555,1130, 2165\}\) were taken and the quantile estimator was calculated. The experiment was repeated 1000 times in each case and a mean square error (MSE) was generated.

In the normal distribution case, the best results were obtained for the estimators \(E_4\), \(E_6\) and \(E_8\). The estimator \(E_1\) was the only one that gave better results for small p. For \(p=0.5\), the results for each estimator are similar. The results for the Student’s t distribution are similar to the normal distribution. Again, the best results were obtained for the estimators \(E_4\), \(E_6\), \(E_8\). The estimators \(E_2\), \(E_3\) and \(E_5\) showed much worse results for small p than the others estimators and showed quite good results for high p. The results for \(p=0.5\) are, again, similar for each estimator. In the case of Weibull distribution, the errors are much smaller than in the normal and Student’s t distribution. Only \(E_1\) gives a better result for small p than for large p. As the sample grows, the differences in the errors for individual estimators are practically non-existent. For the uniform distribution the errors are also smaller than for the normal and Student’s t distribution. In this case, the estimator \(E_7\) gave the best results. Additionally, \(E_1\), \(E_2\), \(E_3\) and \(E_5\) gave a better result for small p than the others. For \(\chi ^2\) distribution the best results gave \(E_2\). Also good results were obtained for the estimators \(E_4\), \(E_5\), \(E_6\) and \(E_8\). Each of the estimators performs worse at high p than at low p. For the exponential distribution the greatest differences in the results are seen for high p. The best results are given by the estimator \(E_2\), but the estimators \(E_4\), \(E_5\), \(E_6\) and \(E_8\) also give quite good results.

The Tables 1 and 2 contain an assessment of the fit of each estimator depending on the distribution for, respectively, \(p=0.025\) and \(p=0.975\). Within a given distribution, estimator with the best fit gets three pluses and estimator with the worst fit gets minus. The errors for a given distribution was analogous for all estimators for \(p=0.5\).

Table 1 Estimator fit quality for \(p=0.025\)
Table 2 Estimator fit quality for \(p=0.975\)

In conclusion, none of the estimators performed equally well for different distributions for low and high p. Different estimators turn out to be the best depending on a given distribution. Only for \(p=0.5\) the fit of each estimator is at a very similar level for a given distribution. Common conclusion for each estimator is that as the sample size increases, the error decreases and the results for each estimator are very similar. However, it should be noted that the financial and life situation does not always allow to obtain a sample of 1000 or 2000 elements. That is why it is so important to choose an appropriate estimator for a given distribution especially when the research is based on a small sample.

3 Strong consistency and Bahadur representation

In this section we will study a strong consistency of a quantile estimator and the Bahadur representation for sample quantiles. Further, detailed considerations will be carried out for the estimator \(E_1\).

Theorem 2

Let \(\{X_n, n\ge 1\}\) be a sequence of random variables with one of the following dependency structures:

  1. 1.

    NA dependence,

  2. 2.

    \(\varphi \)-mixing dependence with coefficients satifying \(\displaystyle \sum _{n=1}^{\infty }\varphi (2^n)<\infty \),

  3. 3.

    \(\rho ^*\)-mixing dependence,

  4. 4.

    \(\rho \)-mixing dependence with coefficients satifying \(\displaystyle \sum _{n=1}^{\infty }\rho ^{\frac{2}{q}}(2^n)<\infty \), for \(\displaystyle q>~\frac{1}{\delta }\) for \(\displaystyle 0<\delta <\frac{1}{2}\).

Let \(\{X_n, n\ge 1\}\) be indentically distributed with a common distribution function F and a quantile \(Q_p\). Assume that F possesses a positive continuous density f in some neighborhood \(\mathfrak {D}_p\) of \(Q_p\) such that \( 0< \sup \{f(x); x \in \mathfrak {D}_p \} <\infty . \) Moreover, we assume that \(f'(x)\) is defined in some neighborhood \(\mathfrak {D}_p\) of \(Q_p\),

$$\begin{aligned} |f'(x)| < M, \quad x\in \mathfrak {D}_p,\quad M\in {\mathbb {R}}. \end{aligned}$$
(3)

Then for any \(0<\delta <\frac{1}{2}\)

$$\begin{aligned} P\Big ( E_1-Q_p=o(n^{-\frac{1}{2}+\delta }), \ \ \text {as} \ \ n\rightarrow \infty \Big )=1. \end{aligned}$$
(4)

Proof

It it easy to see that \(\forall ~\varepsilon >0\)

$$\begin{aligned}{} & {} \bigg \{|E_1-Q_p|\ge \varepsilon n^{-\frac{1}{2}+\delta }\bigg \}\\{} & {} \quad =\bigg \{ E_1\ge Q_p+\varepsilon n^{-\frac{1}{2}+\delta }\bigg \} \cup \bigg \{ E_1\le Q_p-\varepsilon n^{-\frac{1}{2}+\delta }\bigg \} =A_n^1 \cup A_n^2. \end{aligned}$$

One can obtain that

$$\begin{aligned} X_{\lfloor np+0.5\rfloor }\le E_1\le X_{\lfloor np+0.5\rfloor +1}. \end{aligned}$$
(5)

Put \(\displaystyle \xi _{ni}=I(X_i\le Q_p+\varepsilon n^{-\frac{1}{2}+\delta })-F(Q_p+\varepsilon n^{-\frac{1}{2}+\delta })\) for \(1\le i\le n\). Therefore, we get

$$\begin{aligned} A_n^1=\bigg \{ E_1\ge Q_p+\varepsilon n^{-\frac{1}{2}+\delta }\bigg \} \subset \bigg \{ \sum _{i=1}^nI(X_i \le Q_p+\varepsilon n^{-\frac{1}{2}+\delta })<\lfloor np+0.5\rfloor +1\bigg \} \end{aligned}$$
$$\begin{aligned} =\bigg \{\sum _{i=1}^n \xi _{ni}<\lfloor np+0.5\rfloor +1-nF(Q_p+\varepsilon n^{-\frac{1}{2}+\delta })\bigg \}. \end{aligned}$$
(6)

Note that, using Taylor’s expansion:

$$\begin{aligned} F(Q_p+\varepsilon n^{-\frac{1}{2}+\delta })=p+ f(Q_p)\varepsilon n^{-\frac{1}{2}+\delta }+o(n^{-\frac{1}{2}+\delta }) \end{aligned}$$

we can obtain that there exists some constant \(c(\varepsilon )>0 \), depending only on \(\varepsilon >0\), such that for a sufficiently large n

$$\begin{aligned}{} & {} \bigg \{\sum _{i=1}^n \xi _{ni}<\lfloor np+0.5\rfloor +1-np-\varepsilon f(Q_p)n^{\frac{1}{2}+\delta }+o(n^{\frac{1}{2}+\delta })\bigg \} \nonumber \\{} & {} \subset \bigg \{\sum _{i=1}^{n}\xi _{ni}<-c(\varepsilon )n^{\frac{1}{2}+\delta }\bigg \} \end{aligned}$$
(7)

Hence, on the basis of Eqs. (6),(7), Markov’s inequality and Eq. (2) (for NA, \(\varphi \) and \(\rho ^*\)-mixing random variables) or Lemma 1 (for \(\rho \)-mixing random variables), for \(\displaystyle r>\frac{1}{\delta }\) where \(\displaystyle 0<\delta <\frac{1}{2}\), we get the following estimation

$$\begin{aligned} \sum _{n=1}^{\infty } P\bigg (E_1\ge Q_p+\varepsilon n^{-\frac{1}{2}+\delta }\bigg )\le \sum _{n=1}^{\infty } P\bigg (\sum _{i=1}^{n}\xi _{ni}<-c(\varepsilon )n^{\frac{1}{2}+\delta }\bigg ) \\ \le \sum _{n=1}^{\infty } P\bigg (\bigg |\sum _{i=1}^{n}\xi _{ni}\bigg |>c(\varepsilon )n^{\frac{1}{2}+\delta }\bigg )\le C\sum _{n=1}^{\infty }n^{-(\frac{1}{2}+\delta )r}E\bigg |\sum _{i=1}^{n}\xi _{ni}\bigg |^r \\ \le C\sum _{n=1}^{\infty }n^{-(\frac{1}{2}+\delta )r}\Big [(nE\xi ^2_{n1})^{\frac{r}{2}}+nE|\xi _{n1}|^r\Big ]\le C\sum _{n=1}^{\infty }n^{-\delta r}<\infty . \end{aligned}$$

We can carry out analogous considerations for \(A_n^2\) using the lower estimate of \(E_1\) in the inequality Eq. (5). Hence, we get \(\displaystyle \sum _{n=1}^{\infty } P\bigg [ E_1\le Q_p-\varepsilon n^{-\frac{1}{2}+\delta }\bigg ] <\infty .\) By the Borel-Cantelli lemma we get thesis Eq. (4).

\(\square \)

Remark 5

One can obtain analogous results for the estimator:

  • \(E_2\)—by assumption that \(X_{(\lfloor np\rfloor )}\le E_2\le X_{(\lfloor np\rfloor +1)}\),

  • \(E_3\)—calculations are similar to the estimator \(E_4\), which was considered in the paper (Dudek and Kuczmaszewska (2022)),

  • \(E_5\) and \(E_6\)—calculations are analogous to the estimators \(E_3\) and \(E_4\), because the proof does not depend on the value of \(np-\lfloor np\rfloor \),

  • \(E_7\)—by the assumption that \(X_{(\lfloor (n+1)p\rfloor ))}\le E_7\le X_{(\lfloor (n+1)p\rfloor +1))}\),

  • \(E_8\)—it is a combination of the estimator \(E_4\) and \(E_3\).

Now, let us focus on the Bahadur representation of sample quantiles.

Theorem 3

Let \(\{X_n, n\ge 1\}\) be a sequence of random variables with one of the following dependency structures:

  1. 1.

    NA dependence,

  2. 2.

    \(\rho ^*\)-mixing dependence.

Let \(\{X_n, n\ge 1\}\) be indentically distributed with a common distribution function F and quantile \(Q_p\). Assume that F possesses a positive continuous density f in some neighborhood \(\mathfrak {D}_p\) of \(Q_p\) such that

$$\begin{aligned} 0< \sup \{f(x); x \in \mathfrak {D}_p \} <\infty . \end{aligned}$$
(8)

Then for any \( \delta >0\)

$$\begin{aligned} P\bigg (\sup _{x \in \mathfrak {I}_n}|F_n(x)-F(x)-(F_n(Q_p)-p)|=O(n^{-\frac{1}{2}+\delta }), \ \ \ n\rightarrow \infty \bigg )=1, \end{aligned}$$

where \(\mathfrak {I}_n=\big [ Q_p-c_0n^{-\frac{1}{4}+\delta },Q_p+c_0n^{-\frac{1}{4}+\delta }\big ]\) for some \(c_0>0.\)

Proof

Let \(\{a_n, n\ge 1\}\) and \(\{b_n, n\ge 1\}\) be two sequences defined as follows

$$\begin{aligned} a_n = c_0 n^{-{1\over 4}+\delta } \quad \text { for some } \ c_0 >0, \quad b_n = \lfloor n^{1\over 4}\rfloor + 1 \end{aligned}$$

and

$$\begin{aligned} G_n(x) = F_n(x) - F_n(Q_p) - F(x) +p. \end{aligned}$$

For each \(n\in {\mathbb {N}}\) and any integer j we define

$$\begin{aligned} \eta _{j,n} = Q_p + ja_nb_n^{-1}, \quad \alpha _{j,n} = F(\eta _{j+1,n}) -F(\eta _{j,n}) \ \ \text { and } \ \ \mathfrak {J}_{j,n} = [ \eta _{j,n}, \eta _{j+1, n}]. \end{aligned}$$

Since \(F_n\) and F are nondecreasing we get for \(x\in \mathfrak {J}_{j,n}\)

$$\begin{aligned} G_n(x) \le F_n(\eta _{j+1, n}) - F_n(Q_p) - F(\eta _{j,n}) + p \le G_n(\eta _{j+1, n}) + \alpha _{j,n} \end{aligned}$$

and

$$\begin{aligned} G_n(x) \ge F_n(\eta _{j, n}) - F_n(Q_p) - F(\eta _{j+1,n}) + p \ge G_n(\eta _{j, n}) - \alpha _{j,n}. \end{aligned}$$

Hence

$$\begin{aligned} \sup _{x \in \mathfrak {I}_n}|F_n(x)-F(x)-(F_n(Q_p)-p)| \le \max _{-b_n \le j\le b_n} \{ |G_n(\eta _{j, n})|\} + \max _{-b_n \le j\le b_n-1}\{\alpha _{j,n}\}. \end{aligned}$$

It is easy to see that by The Mean Value Theorem and (8) we have

$$\begin{aligned} \alpha _{j,n} = F(\eta _{j+1,n}) -F(\eta _{j,n}) \le C(\eta _{j+1,n}- \eta _{j,n}) = C a_nb_n^{-1} \le Cn^{-\frac{1}{2}+\delta }. \end{aligned}$$

Therefore we obtain

$$\begin{aligned} \sum _{n=1}^\infty P\bigg (\sup _{x \in \mathfrak {I}_n}|F_n(x)-F(x)-(F_n(Q_p)-p)|\ge c_0n^{-\frac{1}{2}+\delta } \bigg ) \\ \le C \sum _{n=1}^\infty P\bigg (\max _{-b_n\le j\le b_n}|G_n(\eta _{j,n})|\ge {c_0\over 2} n^{-\frac{1}{2}+\delta } \bigg ). \end{aligned}$$

Moreover, we note that

$$\begin{aligned} G_n(\eta _{j,n}) = F_n(\eta _{j,n}) - F_n(Q_p) - F(\eta _{j,n}) + p =\frac{1}{n}\sum _{i=1}^n\bigg (Y_i^{Q_p}-Y_i^{(j,n)}\bigg ) \end{aligned}$$

where \(Y_i^{Q_p}=E(I[X_i\le Q_p])-I[X_i\le Q_p]\) and \(Y_i^{(j,n)} = E(I[X_i\le \eta _{j,n}])-I[X_i\le \eta _{j,n}]\), \(-b_n \le j \le b_n\) are respectively NA or \(\rho ^*\)-mixing random variables.

It follows from the Markov’s inequality and Eq. (2) (for NA and \(\rho ^*\)-mixing random variables) that for \(\displaystyle r> \max \bigg \{2,\frac{5}{4\delta }\bigg \}\) where \(\displaystyle \delta >0\) we have

$$\begin{aligned}{} & {} \sum _{n=1}^\infty P\bigg (\sup _{x \in \mathfrak {I}_n}|F_n(x)-F(x)-(F_n(Q_p)-p)|\ge c_0n^{-\frac{1}{2}+\delta } \bigg )\\{} & {} \le C\sum _{n=1}^\infty \sum _{j=-b_n}^{b_n} P\bigg (|G_n(\eta _{j,n})|\ge {c_0\over 2} n^{-\frac{1}{2}+\delta } \bigg )\\{} & {} = C \sum _{n=1}^\infty \sum _{j=-b_n}^{b_n} P\bigg (| \sum _{i=1}^n \bigg (Y_i^{Q_p}-Y_i^{(j,n)}\bigg )| \ge {c_0\over 2}n^{\frac{1}{2}+\delta } \bigg )\\{} & {} \le C \sum _{n=1}^\infty \sum _{j=-b_n}^{b_n}P\bigg (|\sum _{i=1}^nY_i^{Q_p}|+|\sum _{i=1}^nY_i^{(j,n)}|\ge {c_0\over 2}n^{\frac{1}{2}+\delta } \bigg )\\{} & {} \le C \sum _{n=1}^\infty \sum _{j=-b_n}^{b_n}\Bigg (P\bigg (|\sum _{i=1}^nY_i^{Q_p}|\ge {c_0\over 4}n^{\frac{1}{2}+\delta }\bigg )+P\bigg (|\sum _{i=1}^nY_i^{(j,n)}|\ge {c_0\over 4}n^{\frac{1}{2}+\delta } \bigg )\Bigg )\\{} & {} \le C \sum _{n=1}^\infty \sum _{j=-b_n}^{b_n} \Bigg [\frac{E\bigg (|\sum _{i=1}^nY_i^{Q_p}|\bigg )^r}{(n^{\frac{1}{2}+\delta })^r} +\frac{E\bigg (|\sum _{i=1}^nY_i^{(j,n)}|\bigg )^r}{(n^{\frac{1}{2}+\delta })^r}\Bigg ]\\{} & {} \le C\sum _{n=1}^\infty 2b_nn^{-\delta r}\le C\sum _{n=1}^\infty n^{\frac{1}{4}-\delta r} <\infty . \end{aligned}$$

By the Borel-Cantelli lemma we get thesis. \(\square \)

Theorem 4

Let \(\{X_n, n\ge 1\}\) be a sequence of random variables with one of the following dependency structures:

  1. 1.

    \(\varphi \)-mixing dependence with coefficients satifying \(\displaystyle \sum _{n=1}^{\infty }\varphi (2^n)<\infty \),

  2. 2.

    \(\rho \)-mixing dependence with coefficients satifying \(\displaystyle \sum _{n=1}^{\infty }\rho ^{\frac{2}{q}}(2^n)<\infty \), for \(q> \max \bigg \{2,{5\over 2\delta }\bigg \}\) for \(\delta >0\).

Let \(\{X_n, n\ge 1\}\) be indentically distributed with a common distribution function F and quantile \(Q_p\). Assume that F possesses a positive continuous density f in some neighborhood \(\mathfrak {D}_p\) of \(Q_p\) such that

$$\begin{aligned} 0< \sup \{f(x); x \in \mathfrak {D}_p \} <\infty . \end{aligned}$$
(9)

Then for any \( \delta >0\)

$$\begin{aligned} P\bigg (\sup _{x \in \mathfrak {I}_n}|F_n(x)-F(x)-(F_n(Q_p)-p)|=O(n^{-\frac{3}{4}+\delta }), \ \ \ n\rightarrow \infty \bigg )=1, \end{aligned}$$

where \(\mathfrak {I}_n=\big [ Q_p-c_0n^{-\frac{1}{2}+\delta },Q_p+c_0n^{-\frac{1}{2}+\delta }\big ]\) for some \(c_0>0\)

Proof

The proof of Theorem 4 is very similar to the proof of Theorem 3. Analogously, let \(\{a_n, n\ge 1\}\) and \(\{b_n, n\ge 1\}\) be two sequences defined as follows

$$\begin{aligned} a_n = c_0 n^{-{1\over 2}+\delta } \quad \text { for some } \ c_0 >0, \quad b_n = \lfloor n^{1\over 4}\rfloor + 1 \end{aligned}$$

and

$$\begin{aligned} G_n(x) = F_n(x) - F_n(Q_p) - F(x) +p. \end{aligned}$$

Let \(\eta _{j,n}\), \(\alpha _{j,n}\) and \(\mathfrak {J}_{j,n}\) be defined as in Theorem 3.

As it was shown in Theorem 3, it is easy to see that

$$\begin{aligned} \alpha _{j,n} = F(\eta _{j+1,n}) -F(\eta _{j,n}) \le C(\eta _{j+1,n}- \eta _{j,n}) = C a_nb_n^{-1} \le Cn^{-\frac{3}{4}+\delta }. \end{aligned}$$

Therefore we get

$$\begin{aligned} \sum _{n=1}^\infty P\bigg (\sup _{x \in \mathfrak {I}_n}|F_n(x)-F(x)-(F_n(Q_p)-p)|\ge c_0n^{-\frac{3}{4}+\delta } \bigg )\\ \le C \sum _{n=1}^\infty P\bigg (\max _{-b_n\le j\le b_n}|G_n(\eta _{j,n})|\ge {c_0\over 2} n^{-\frac{3}{4}+\delta } \bigg ). \end{aligned}$$

Next, we have

$$\begin{aligned} G_n(\eta _{j,n}) = F_n(\eta _{j,n}) - F_n(Q_p) - F(\eta _{j,n}) + p ={1\over n} \sum _{i=1}^n \bigg ( Y_i^{(j,n)} - EY_i^{(j,n)}\bigg ), \end{aligned}$$

where \(Y_i^{(j,n)} = I[Q_p< X_i\le \eta _{j,n}]\), \(-b_n \le j \le b_n\) are respectively \(\varphi \) or \(\rho \)-mixing random variables.

Moreover, it is easy to obtain by The Mean Value Theorem that for \(r\ge 2\)

$$\begin{aligned} E|Y_i^{(j,n)}|^r\le Cja_nb_n^{-1}. \end{aligned}$$

It follows from the Markov’s inequality and Eq. (2) (for \(\varphi \)-mixing random variables) or Lemma 1 (for \(\rho \)-mixing random variables) that for \(\displaystyle r> \max \bigg \{2,{5\over 2\delta }\bigg \}\) where \(\displaystyle \delta >0\) we have

$$\begin{aligned} \sum _{n=1}^\infty P\bigg (\sup _{x \in \mathfrak {I}_n}|F_n(x)-F(x)-(F_n(Q_p)-p)|\ge c_0n^{-\frac{3}{4}+\delta } \bigg )\\ \le C\sum _{n=1}^\infty \sum _{j=-b_n}^{b_n} P\bigg (|G_n(\eta _{j,n})|\ge \frac{c_0}{2} n^{-\frac{3}{4}+\delta } \bigg )\\ = C \sum _{n=1}^\infty \sum _{j=-b_n}^{b_n} P\bigg (| \sum _{i=1}^n \bigg ( Y_i^{(j,n)} - EY_i^{(j,n)}\bigg )| \ge \frac{c_0}{2} n^{\frac{1}{4}+\delta } \bigg )\\ \le C \sum _{n=1}^\infty \sum _{j=-b_n}^{b_n} n^{-({1\over 4}+\delta )r} E|\sum _{i=1}^n \bigg ( Y_i^{(j,n)} - EY_i^{(j,n)}\bigg )|^r \end{aligned}$$
$$\begin{aligned} \le C \sum _{n=1}^\infty \sum _{j=-b_n}^{b_n} n^{-({1\over 4}+\delta )r} \bigg [ \bigg ( nE( Y_1^{(j,n)})^2\bigg )^{r\over 2} + nE|Y_1^{(j,n)}|^r\bigg ]\\ \le C \sum _{n=1}^\infty \sum _{j=-b_n}^{b_n} n^{-({1\over 4}+\delta )r} \bigg ( n ja_nb_n^{-1} \bigg )^{r\over 2} \le C \sum _{n=1}^\infty n^{-{\delta r\over 2}+ {1\over 4}} < \infty , \end{aligned}$$

By the Borel-Cantelli lemma we get thesis. \(\square \)

Theorem 5

Suppose that assumptions of Theorem 2 hold. Then for any \(\displaystyle 0<\delta <\frac{1}{2}\) we have,

$$\begin{aligned} P\Bigg ( E_1= Q_p-\frac{F_n( Q_p)-p}{f( Q_p)} + O(n^{-\frac{1}{2}+\delta }), \ \ \text {as} \ \ n\rightarrow \infty \Bigg )=1 \end{aligned}$$
(10)

for NA and \(\rho ^*\)-mixing random variables.

Proof

One can note that

$$\begin{aligned} F_n( E_1)\le n^{-1}(\lfloor np+0.5\rfloor +1) = p+O(n^{-1}). \end{aligned}$$

By Taylor’s expansion we obtain for \(0<\theta <1\)

$$\begin{aligned} F(E_1)=p+f(Q_p)(E_1-Q_p)+\frac{1}{2}f'( Q_p+\theta ( E_1- Q_p))( E_1- Q_p)^2. \end{aligned}$$

From Eq. (3) and Theorem 2 for \(\displaystyle 0<\delta <\frac{1}{2}\) it follows that

$$\begin{aligned} |F_n( E_1)-F( E_1)+f(Q_p)( E_1-Q_{p})| \end{aligned}$$
$$\begin{aligned} \le \frac{1}{2}|f'(Q_p+\theta (E_1-Q_p))|(E_1-Q_p)^2+O(n^{-1})=o(n^{-1+2\delta }). \end{aligned}$$
(11)

By Eq. (11) and Theorem 3 for \(\displaystyle 0<\delta <\frac{1}{2}\) we get that with probability 1,

$$\begin{aligned}{} & {} |f( Q_p)(E_1-Q_p)+F_n(Q_{p})-p|\\{} & {} \quad \le |F_n(E_1)-F(E_1)+f(Q_p)(E_1-Q_p)|+|F_n(E_1)-F(E_1)-(F_n(Q_{p})-p)|\\{} & {} \quad \le o(n^{-1+2\delta })+\sup _{x\in \mathfrak {I}_n}|F_n(x)-F(x)-(F_n(Q_{p})-p)|=O(n^{-\frac{1}{2}+\delta }), \end{aligned}$$

which gives that \( f( Q_p)(E_1-Q_p)+F_n(Q_{p})-p = O(n^{-\frac{1}{2}+\delta }),\) when \(n\rightarrow \infty . \) Hence, we get Eq. (10). \(\square \)

Theorem 6

Suppose that assumptions of Theorem 2 hold. Then for any \(\displaystyle 0<\delta <\frac{1}{4}\) we get

$$\begin{aligned} P\Bigg ( E_1= Q_p-\frac{F_n( Q_p)-p}{f( Q_p)} + O(n^{-\frac{3}{4}+\delta }), \ \ \text {as} \ \ n\rightarrow \infty \Bigg )=1 \end{aligned}$$
(12)

for \(\varphi \) and \(\rho \)-mixing random variables.

Proof

Analogously as in the proof of Theorem 5, using Eq. (11) and Theorem 4 for \(\displaystyle 0<~\delta <\frac{1}{4}\) we get that with probability 1,

$$\begin{aligned} |f( Q_p)(E_1-Q_p)+F_n(Q_{p})-p|=O(n^{-\frac{3}{4}+\delta }). \end{aligned}$$

Hence, we get Eq. (12). \(\square \)

Remark 6

The calculations for the estimators \(E_2\), \(E_4, E_5,E_6,E_7\) and \(E_8\) proceed analogously through the condition that \(F_n( E_i) = p+O(n^{-1}),\) where \(i=2,4,5,6,7,8\). Assuming that \(F_n( E_3) \le p\), one can obtain analogous results for the estimator \(E_3\).