Mathematical Sciences

, Volume 13, Issue 4, pp 375–385

# A new class of skew-logistic distribution

• Anis Iranmanesh
Open Access
Original Research

## Abstract

In this study, the researchers introduce a new class of the logistic distribution which can be used to model the unimodal data with some skewness present. The new generalization is carried out using the basic idea of Nadarajah (Statistics 48(4):872–895, 2014), called truncated-exponential skew-logistic (TESL) distribution. The TESL distribution is a member of the exponential family; therefore, the skewness parameter can be derived easier. Meanwhile, some important statistical characteristics are presented; the real data set and simulation studies are applied to evaluate the results. Also, the TESL distribution is compared to at least five other skew-logistic distributions.

## Keywords

Azzalini skew distributions Skewness Skew-logistic distribution Truncated-exponential skew-symmetric distributions Truncated-exponential skew-logistic distributions

62E10 62E15

## Introduction

The necessity for skew distributions appears in all fields of science. Univariate skew-symmetric models have been considered by several authors. Ferreira and Steel [7] defined Y as a new random variable with the probability density function (pdf)
\begin{aligned} f_{Y}(y) = f_{X}(y)w(F_{X}(y)),\quad y\in R, \end{aligned}
(1)
where w(.) is a pdf on (0, 1) and X is a symmetric random variable about zero with pdf $$f_{X}(.)$$ and cumulative distribution function (cdf) $$F_{X}(.)$$. Then, Y is a skew version of the symmetric random variable X. By changing w(.) in (1), many of the known families of skew distributions are obtained. For example if $$w(x)= 2F_{X}(\lambda F_{X}^{-1}(x))$$, (1) changes to
\begin{aligned} f_{Y}(y) =2 f_{X}(y)F_{X}(\lambda y),\quad y\in R, \;\;\; \lambda \in R, \end{aligned}
(2)
as Azzalini skew distributions ([2] and [3]). Some special cases of (2) are the class of skew-normal distributions, $$f_{Y}(y) =2 \varphi (y)\phi (\lambda y)$$, where $$\varphi$$ and $$\phi$$ are the standard normal pdf and cdf, respectively, skew-t and skew-Cauchy distributions.
The logistic distribution (LD) has been the center of attention in several areas of scientific research. It is applied as an alternative to the normal distribution in many practical situations. $$L(\mu , \sigma )$$ denote the logistic distribution with parameters $$\mu$$ and $$\sigma$$, with the following pdf and cdf respectively,
\begin{aligned} f(x) &= \exp (-(x-\mu )/ \sigma )/[1+\exp (-(x-\mu )/\sigma )]^{2},\\&\quad x,\mu \in R,~ \sigma>0,\\ F(x)&= 1/[1+\exp (-(x-\mu )/\sigma )],\quad x, \mu \in R, \;\; \sigma >0. \end{aligned}
Furthermore, L(0, 1) denote the standard logistic distribution with $$\mu =0$$ and $$\sigma =1$$. Some properties and applications of the LD are available in Balakrishnan [4]. Similar to the skew-normal distribution of Azzalini [2], Wahed and Ali [15] introduced skew-logistic distribution (SLD) through the family of Azzalini skew distributions. Nadarajah [11], Gupta and Kundu [8] also derived some of its properties. One of the important problems of family (1) is complicated inference methods about finding estimator for skewness parameter. For example, Pewsey [13] has proved the MLE for the skewness parameter in Azzalini skew-normal families does not always exist. However, SLD was extended by many authors and received widespread attention. Gupta and Kundu [8] also discuss another generalization of the logistic distribution using the idea of the proportional reversed hazard family. They named this distribution as the proportional reversed hazard logistic distribution (PRHLD) (or type-I generalized logistic distribution). They also showed that PRHLD has several advantages over SLD. Chakraborty et al. [5] and Hazarika and Chakraborty [9] considered other skew-logistic distribution (NSLD) and the alpha skew-logistic distribution (ASLD), respectively. ASLD is flexible enough to adequately model both unimodal and bimodal data in the presence of positive or negative skewness. Asgharzadeh et al. [1] proposed a generalized skew-logistic distribution (GSLD) using the type-III generalized logistic distribution. A generalized version of the ASLD is also introduced by Hazarika and Chakraborty [10]. Satheesh Kumar and Manju [14] proposed modified skew-logistic distribution (MSLD). Some pdfs of above skew-logistic distributions are given in Table 1.
Table 1

Different skew-logistic distributions

Distribution

pdf

SLD$$(\lambda )$$

$$f_{1}(x)=\frac{2 \exp (-x)}{[1+\exp (-x)]^{2} [1+\exp (-\lambda x )]} ,\;\; x , \lambda \in R$$

PRHLD$$(\lambda )$$

$$f_{2}(x)=\frac{\lambda \exp (-x)}{[1+\exp (-x)]^{\lambda +1} } ,\;\; x , \lambda \in R$$

NSLD$$(\lambda , \alpha , \beta )$$

$$f_{3}(x)=\frac{[1+(\sin (\frac{\lambda x}{2\beta }))/ \alpha ]\exp (-x/ \beta )}{\beta [1+\exp (-x/ \beta )]^{2}} ,\;\; x , \lambda \in R,\;\; \alpha \ge 1, \;\; \beta > 0$$

GSLD$$(\alpha , \lambda )$$

$$f_{4}(x)=\frac{2 B_{y}(\alpha , \alpha ) \exp (- \alpha x)}{ B^{2}(\alpha , \alpha )[1+\exp (-x)]^{2\alpha } } ,\;\; y=[1+\exp (-\lambda x)]^{-1},\;\; x , \lambda \in R, \;\;\alpha >0$$

ASLD$$(\alpha )$$

$$f_{5}(x)=\frac{3[1+(1-\alpha x)^{2} ] \exp (-x)}{[6+\alpha ^{2} \pi ^{2}] [1+\exp (-x)]^{2}},\;\; x , \alpha \in R$$

MSLD$$(\alpha , \beta )$$

$$f_{6}(x)=\frac{2 \exp (-x)}{(\alpha +2)[1+\exp (-x)]^{2}}\left[1+\frac{\alpha \exp (-\beta x)}{1+\exp (-\beta x)}\right],\;\; x, \beta \in R, \;\;\alpha \ge -1$$

Nadarajah et al. [12] introduce a new family of skew distributions, naming it truncated-exponential skew-symmetric (TESS) distributions. TESS is a member of the exponential family; therefore, the estimate of the skewness parameter can be obtained easier.

In this study which is based on TESS distribution, a new generalized logistic distribution is introduced. It refers to truncated-exponential skew-logistic (TESL) distribution and compares it to other skew-logistic distributions. Also some properties of TESL distribution are obtained. In other skew-logistic distributions which are based on (2), it is generally difficult or almost impossible to estimate skewness parameter. The aim of this study is to introduce the new skew-logistic distributions that is a member of exponential family and bears the good characteristics of the family. It is expected that the estimation of skew parameter to be easier and more precise.

## Truncated-exponential skew-logistic distribution

A random variable Y has the truncated-exponential skew-symmetric distribution, denoted by TESS$$(\lambda )$$, if its pdf is given by
\begin{aligned} f_{Y}(y) =c f_{X} (y) \exp (-\lambda F_{X}(y)),\quad x, \lambda \in R. \end{aligned}
(3)
where $$f_{X}(.)$$ and $$F_{X}(.)$$ are, respectively, the pdf and cdf of a symmetric random variable X about zero, $$\lambda$$ is shape parameter and $$c=\frac{\lambda }{1-\exp (-\lambda )}$$. The cdf of Y is obtained as
\begin{aligned} F_{Y}(y) =\frac{c}{\lambda }\{1-\exp (-\lambda F_{X}(y))\},\quad x, \lambda \in R. \end{aligned}
Note that Eq. (3) is a particular case of Eq. (1) for $$w(x) = \frac{\lambda \exp (\lambda x)}{1-\exp (\lambda )}$$.

### Definition 2.1

A random variable Y has the truncated-exponential skew-logistic distribution with parameters $$\mu \in R, \sigma >0$$ and $$\lambda \in R$$, denoted by TESL$$(\mu , \sigma , \lambda )$$, if its pdf is given by
\begin{aligned} f_{Y}(y)&= \frac{\lambda \exp [-(y-\mu )/\sigma ]}{\sigma [1-\exp (-\lambda )][1+\exp [-(y-\mu )/\sigma ]]^{2}}\nonumber \\&\quad \exp \left( \frac{-\lambda }{1+\exp [-(y-\mu )/\sigma ]}\right) ,\quad y \in R. \end{aligned}
(4)
Further, the cdf of Y is
\begin{aligned} F_{Y}(y) =\left[ \frac{1}{1-\exp (-\lambda )}\right] \left[ {1- \exp \left( \frac{-\lambda }{1+\exp [-(y-\mu )/\sigma ]}\right) }\right] ,\quad y \in R. \end{aligned}
If $$\mu =0$$ and $$\sigma =1$$, we denote $$Y \sim {\rm TESL}(\lambda )$$ .

Note that the TESL pdf in (4) can be written as $$f(y)=h(y)c(\lambda ) \exp \{ w(\lambda )t(y)\}$$. where $$t(y)=F_{X}((y-\mu )/\sigma )=1/(1+\exp -(y-\mu )/\sigma )$$, $$c(\lambda )=\lambda /\sigma (1-\exp (-\lambda ))$$, $$h(y)=f_{X}((y-\mu )/\sigma )=\exp (-\lambda )/\sigma [1+\exp (-(y-\mu )/\sigma ]^{2}$$ and $$w(\lambda )=-\lambda$$. Therefore, (4) belongs to the exponential family and $$D({\underline{y}})=\sum _{i=1}^{n}F_{X}((y_{i}-\mu )/\sigma )=\sum _{i=1}^{n}[1/(1+\exp (-(y_{i}-\mu )/\sigma ))]$$ is a complete sufficient statistic for $$\lambda$$, if $$\mu$$ and $$\sigma$$ are assumed known. Also $$E(D({\underline{Y}}))=(1/\lambda )-[\exp (-\lambda )/(1-\exp (-\lambda ))]$$.

The inverse cdf of TESL$$( \lambda )$$ is
\begin{aligned} F^{-1}_{Y}(y) =\ln \left\{ \frac{-\ln [1-y(1-\exp (-\lambda ))]}{\lambda +\ln [1-y(1-\exp (-\lambda ))]} \right\} ,\quad y \in R. \end{aligned}
(5)
The hazard rate function of TESL$$(\mu , \sigma , \lambda )$$ is
\begin{aligned}h_{Y}(y)=\frac{{ \lambda \exp (-(y-\mu )/\sigma )\exp \left( \frac{-\lambda }{1+\exp (-(y-\mu )/\sigma )}\right) }}{{\sigma [1+\exp (-(y-\mu )/\sigma )]^{2}} \left( \exp \left( \frac{-\lambda }{1+\exp (-(y-\mu )/\sigma )}\right) -\exp (-\lambda )\right) },\quad y \in R. \end{aligned}
Figure 1 shows the pdf and hazard rate function of TESL$$(0, \sigma , \lambda )$$ for different values of $$\sigma$$ and $$\lambda$$.

## Properties

In this section, some mathematical properties of TESL distribution are derived. In the sequel, the following notations and definitions are required.

Let $$X_{i:n}$$ denote the ith order statistic for a random sample of size n from L(0, 1), then the moment generating function (mgf) of $$X_{i:n}$$ is
\begin{aligned} M_{X_{i:n}}(t) = \frac{\Gamma (i+t) \Gamma (n-i+1-t)}{\Gamma (i) \Gamma (n-i+1)},\quad 1~\le i\le n. \end{aligned}
Further, $$E(X_{i:n})=\psi (i)-\psi (n-i+1)$$ and $${\rm Var}(X_{i:n})=\psi ^{\prime }(i)+\psi ^{\prime }(n-i+1)$$, where $$\Gamma (.)$$, $$\psi (.)$$ and $$\psi ^{\prime }(.)$$ are gamma, digamma and trigamma functions, respectively.
Let $$Y \sim {\rm TESL}(\lambda )$$ it follows from Nadarajah et al. [12]; the mgf of $$Y_{i:n}$$ is given by
\begin{aligned} M_{Y_{i:n}}(t)&= c\sum _{k=0}^{\infty }\frac{(-\lambda )^{k}}{(k+1)!} M_{X_{k+1:k+1}}(t),\\&= c \sum _{k=0}^{\infty }\frac{(-\lambda )^{k}}{(k+1)!} \frac{\Gamma (k+1+t) \Gamma (1-t)}{\Gamma (k+1) }. \end{aligned}
where $$c=\frac{\lambda }{1-\exp (-\lambda )}$$.
By theorem (1) in Nadarajah et al. [12], if $$Y \sim {\rm TESS}(\lambda )$$ and $$E|X|^{r}, r > 0$$ , exists, then $$E|Y|^{r}$$ exists, so the rth moment of Y when $$Y \sim {\rm TESL}(\lambda )$$ is
\begin{aligned} E(Y^{r})&= c \sum _{k=0}^{\infty }\frac{(-\lambda )^{k}}{(k+1)!} E(X_{k+1:k+1}^r),\\&= c \sum _{k=0}^{\infty }\frac{(-\lambda )^{k}}{(k+1)!} \frac{{\rm d}^{(r)}}{{\rm d}t^{r}} \left\{ \frac{\Gamma (k+1+t) \Gamma (1-t)}{\Gamma (k+1)} \right\} \Bigg\vert _{t=0}. \end{aligned}
In particular
\begin{aligned} E(Y) &= c \sum _{k=0}^{\infty }\frac{(-\lambda )^{k}}{(k+1)!}\{\psi (k+1)-\psi (1)\} \end{aligned}
(6)
\begin{aligned} E(Y^2) &= c \sum _{k=0}^{\infty }\frac{(-\lambda )^{k}}{(k+1)!} \left\{ \left[ \psi \left( k+1 \right) -\psi \left( 1\right) \right] ^{2}\right. \nonumber \\&\quad\left. +\,\left[ \psi ^{\prime }\left( k+1 \right) +\psi ^{\prime }\left( 1 \right) \right] \right\} \end{aligned}
(7)
\begin{aligned} E(Y^{3})&= c \sum _{k=0}^{\infty }\frac{(-\lambda )^{k}}{(k+1)!} \left\{ [\psi (k+1)-\psi (1)]^{3} \right. \nonumber \\&\quad +\, 3[\psi ^{\prime }(k+1)+\psi ^{\prime }(1)] [\psi (k+1)-\psi (1)]\nonumber \\&\quad \left. +\,[\psi ^{\prime \prime }(k+1)-\psi ^{\prime \prime } (1)] \right\} \end{aligned}
(8)
\begin{aligned} E(Y^{4})&= c \sum _{k=0}^{\infty }\frac{(-\lambda )^{k}}{(k+1)!} \left\{ [\psi (k+1)-\psi (1)]^{4} \nonumber \right. \\&\quad +\, 6[\psi ^{\prime }(k+1)+\psi ^{\prime }(1)] [\psi (k+1)-\psi (1)]^{2} \nonumber \\&\quad +\, 4[\psi ^{\prime \prime }(k+1)-\psi ^{\prime \prime }(1)] [\psi (k+1)-\psi (1)] \nonumber \\&\quad \left. +\, 3[\psi ^{\prime }(k+1)+\psi ^{\prime }(1)]^{2} +[\psi ^{\prime \prime \prime }(k+1)+\psi ^{\prime \prime \prime }(1)]\right\} \end{aligned}
(9)
Let $$Y\sim {\rm TESL}(\lambda )$$, the rth L-moment of Y is given by
$$\lambda _{r} = \sum\limits_{{j = 0}}^{{r - 1}} {\frac{{( - 1)^{{r - 1 - j}} }}{{r + 1}}\left( {\begin{array}{cc} {r - 1} \\ j \\ \end{array} } \right)} {\text{ }}\left( {\begin{array}{cc} {r - 1 + j} \\ j \\ \end{array} } \right)(\psi (r + 1) - \psi (1))$$
Table 2

Moments of X for different values of $$\lambda$$

$$\lambda$$

E(X)

$$E(X^2)$$

$$E(X^3)$$

$$E(X^4)$$

$$\mathrm{Var}(X)$$

Sk(X)

Ku(X)

− 10

2.7669

9.5600

39.2601

188.7899

1.9042

0.8641

4.4875

− 9

2.6470

8.9511

36.2059

172.5034

1.9443

0.8185

4.8080

− 8

2.5107

8.3017

33.0139

155.8221

1.9981

0.7568

4.7711

− 7

2.3531

7.6091

29.6658

138.7656

2.0722

0.6736

4.7405

− 6

2.1674

6.8734

26.1389

121.4844

2.1756

0.5641

4.7173

− 5

1.1949

6.1019

22.4048

103.9792

2.3194

0.4290

4.6924

− 4

1.6744

5.3158

18.4315

86.9098

2.5123

0.2808

4.6428

− 3

1.3468

4.5591

14.1901

71.0491

2.7509

0.1450

4.5427

− 2

0.9492

3.9059

9.6709

57.7378

3.0050

0.0497

4.3963

− 1

0.4932

3.4531

4.9085

48.6887

3.2099

0.0068

4.2576

1

− 0.4932

3.4531

− 4.9085

48.6887

3.2099

− 0.0068

4.2576

2

− 0.9492

3.9059

− 9.6709

57.7378

3.0050

− 0.0497

4.3963

3

− 1.3447

4.5591

− 14.1901

71.0491

2.7509

− 0.1450

4.5427

4

− 1.6744

5.3158

− 18.4315

86.9098

2.5122

− 0.2808

4.6428

5

− 1.9449

6.1019

− 22.4048

103.9792

2.3194

− 0.4290

4.6924

6

− 2.1674

6.8734

− 26.1389

121.4148

2.1756

− 0.5641

4.7173

7

− 2.3531

7.6091

− 29.6658

138.7656

2.0722

− 0.6736

4.7405

8

− 2.5107

8.3017

− 33.0139

155.8221

1.9981

− 0.7568

4.7711

9

− 2.6470

8.9511

− 36.2059

172.5034

1.9443

− 0.8185

4.8080

10

− 2.7669

9.5600

− 39.2601

188.7899

1.9042

− 0.8641

4.8475

Table 3

MLEs, log-likelihood, AIC and BIC

Model

Estimates

LOGL

AIC

BIC

SLD$$(\mu , \sigma , \lambda )$$

(44.55831, 2.239627, − 0.469380)

− 549.5123

1105.025

1705.025

PRHLD$$(\mu , \sigma , \lambda )$$

(38.818003, 2.797149, 3.090291)

− 552.324765

1110.6495

1710.6495

ASLD$$(\mu , \sigma , \alpha )$$

(42.914148, 0.955992, 2.764117)

− 579.47206

1164.94411

1764.94411

MSLD$$(\mu , \sigma , \beta , \alpha )$$

( 40.0000, 2.790501, 0.1000, 0.1000)

− 606.82659

1219.65318

1819.65318

SLND$$(\mu , \sigma , \lambda )$$

( 41.11693, 2.406389, 0.43839)

− 550.3552

1106.710

1706.710

SND$$(\mu , \sigma , \lambda )$$

( 42.837784, 3.675828, 0.086758)

− 548.37808

1102.7562

1702.7562

STD$$(\mu , \sigma , \lambda )$$

( 45.18025, 2.405736, − 0.744907)

− 581.1969

1168.394

1768.394

TESL$$(\mu , \sigma , \lambda )$$

(48.709536, 2.605054, 6.070326)

− 548.815763

1103.632

1703.632

Some moments are given in Table 2. Using the above moments, we can calculate the four measures E(X), Var(X), Skewness(X) and Kurtosis(X). Figure 2 illustrates the behavior of the four measures for different values of $$\lambda$$. Based on Fig. 2 it is clear that
1. 1.

E(X) and Skewness(X) decrease with increasing $$\lambda$$;

2. 2.

Var(X) decreases with increasing $$\mid \lambda \mid$$;

3. 3.

Kurtosis(X) increases with increasing $$\mid \lambda \mid$$.

Let $${\mathcal {R}}_{X}\left( \gamma \right)$$ and $${\mathcal {R}}_{Y}\left( \gamma \right)$$ denote Rényi entropies of X and Y, respectively. To derive $${\mathcal {R}}_{Y}\left( \gamma \right)$$, we have
\begin{aligned}&\int _{-\infty }^{\infty }\left( f_{Y}\left( y\right) \right) ^{\gamma }{\rm d}y\\&\quad = \left\{ c^{\gamma } \int _{-\infty }^{\infty }\left( f_{X}(y)\right) ^{\gamma }\exp \left( -\lambda \gamma F_{X}(y)\right) {\rm d}y \right\} \\&\quad = \left\{ c^{\gamma } \int _{-\infty }^{\infty }\left( f_{X}(y)\right) ^{\gamma }\sum _{k=0}^{\infty }\frac{(-1)^{k}\left( \lambda \gamma \right) ^{k}\left( F_{X}(y)\right) ^{k} }{k!} {\rm d}y \right\} \\&\quad = \left\{ c^{\gamma } \sum _{k=0}^{\infty }\frac{(-1)^{k}\left( \lambda \gamma \right) ^{k} }{k!}\int _{-\infty }^{\infty }\left( f_{X}(y)\right) ^{\gamma }\left( F_{X}(y)\right) ^{k} {\rm d}y \right\} \\&\quad = \left\{ c^{\gamma } \left[ \int _{-\infty }^{\infty }\left( f_{X}(y) \right) ^{\gamma }{\rm d}y + \sum _{k=1}^{\infty }\frac{(-1)^{k}\left( \lambda \gamma \right) ^{k} }{k!}\int _{-\infty }^{\infty }\left( f_{X}(y)\right) ^{\gamma }\left( F_{X}(y)\right) ^{k} {\rm d}y\right] \right\} \\&\quad = \left\{ c^{\gamma } \left[ \exp \{(1-\gamma ){\mathcal {R}}_{X}\left( \gamma \right) \}+ \sum _{k=1}^{\infty }\frac{(-1)^{k}\left( \lambda \gamma \right) ^{k} }{k!}I_{k}(y)\right] \right\} \end{aligned}
where
\begin{aligned} I_{k}(y)= & {} \int _{-\infty }^{\infty }\left( f_{X}(y)\right) ^{\gamma }\left( F_{X}(x)\right) ^{k} {\rm d}y \\= & {} \int _{-\infty }^{ \infty }\left( \frac{\exp (-y)}{\left( 1+\exp (-y)\right) ^{2} } \right) ^{\gamma }\left( \frac{1}{1+\exp (-y)} \right) ^{k}{\rm d}y \\= & {} \int _{-\infty }^{ \infty }\frac{\exp (-\gamma y)}{\left( 1+\exp (-y)\right) ^{2\gamma +k} }{\rm d}y \\= & {} \int _{1}^{\infty }\frac{\left( u-1\right) ^{\gamma -1} }{u^{2\gamma +k} }{\rm d}u \\= & {} \int _{1}^{\infty }\sum _{j=0}^{\infty }\frac{{{\gamma -1} \atopwithdelims (){j}} \left( -1 \right) ^{j}u^{\gamma -1-j} }{u^{2\gamma +k} }{\rm d}u \\= & {} \sum _{j=0}^{\infty }{{\gamma -1} \atopwithdelims (){j}} \left( -1 \right) ^{j}\int _{1}^{\infty }u^{-\gamma -j-k-1}{\rm d}u \\= & {} \sum _{j=0}^{\infty }{{\gamma -1} \atopwithdelims (){j}} \left( -1 \right) ^{j}\frac{1}{\gamma +j+k},\quad \gamma >-\left( j+k\right) . \end{aligned}
Then, the Rényi entropy corresponding to $$f_{Y}(.)$$ is given by
\begin{aligned} {\mathcal {R}}_{Y}\left( \gamma \right)&= \frac{1}{1-\gamma }\ln \left\{ c^{\gamma } \left[ \exp \{(1-\gamma ){\mathcal {R}}_{X}\left( \gamma \right) \}\right. \right. \\&\quad \left. \left. +\, \sum _{k=1}^{\infty }\sum _{j=0}^{\infty }{{\gamma -1} \atopwithdelims (){j}}\frac{(-1)^{k+j}\left( \lambda \gamma \right) ^{k} }{k! \left( \gamma +j+k\right) }\right] \right\} . \end{aligned}

## Estimations

Let $$y_{1}, y_{2},\ldots , y_{n}$$ be a random sample from (4). For estimating $$\varvec{\theta }=(\mu , \sigma , \lambda )^{\prime }$$, we derive their moments and maximum likelihood estimators. Let $$m_{k}=(1/n)\sum _{j=1}^{n}y_{i}^{k}$$, for $$k=1, 2, 3$$. By equating the theoretical moments of Eqs. (69) with the sample moments, we obtain the moments estimators.

The likelihood function of $$\varvec{\theta }=(\mu , \sigma , \lambda )^{\prime }$$ is
\begin{aligned} L(\mu , \sigma , \lambda ) &= \left( \frac{\lambda }{\sigma (1-\exp (-\lambda ))}\right) ^{n} \left\{ \prod _{i=1}^{n}\frac{\exp (-z_{i})}{(1+\exp (-z_{i}))^{2}}\right\} \exp \left\{ -\lambda \sum _{i=1}^{n} \frac{1}{1+\exp (-z_{i})}\right\} \end{aligned}
where $$z_{i}=(y_{i}-\mu )/\sigma$$. The log-likelihood function is given by
\begin{aligned} \log L(\mu , \sigma , \lambda ) &= n\ln (\lambda )-n\ln (\sigma )-n\ln (1-\exp (-\lambda ))\\&\quad +\,\sum _{i=1}^{n}[-z_{i}-2\ln (1+\exp (-z_{i})) ]\\&\quad -\,\lambda \sum _{i=1}^{n} \frac{1}{1+\exp (-z_{i})} \end{aligned}
So, the maximum likelihood (ML) estimators can be obtained by solving the following equations
\begin{aligned} \frac{\partial Log L}{\partial \mu }= & {} \sum _{i=1}^{n}\left\{ \frac{1}{\sigma }-\frac{2\exp (-z_{i})}{\sigma (1+\exp (-z_{i}))}\right\} \nonumber \\&+\,\lambda \sum _{i=1}^{n}\left\{ \frac{\exp (-z_{i})}{\sigma (1+\exp (-z_{i}))^{2}}\right\} =0\nonumber \\ \frac{\partial Log L}{\partial \sigma }= & {} \frac{-n}{\sigma }+\sum _{i=1}^{n}\left\{ \frac{z_{i}}{\sigma } -\frac{2z_{i}\exp (-z_{i})}{\sigma (1+\exp (-z_{i}))}\right\} \nonumber \\&+\,\lambda \sum _{i=1}^{n}\left\{ \frac{z_{i}\exp (-z_{i})}{\sigma (1+\exp (-z_{i}))^{2}}\right\} =0\nonumber \\ \frac{\partial Log L}{\partial \lambda }= & {} \frac{n}{\lambda }-\frac{n\exp (-\lambda )}{1-\exp (-\lambda )}\nonumber \\&-\,\sum _{i=1}^{n}\frac{1}{1+\exp (-z_{i})}=0. \end{aligned}
(10)
These equations can be solved numerically for $$\mu$$, $$\sigma$$ and $$\lambda$$. By Theorem 3 in Nadarajah et al. [12], if $$\mu$$ and $$\sigma$$ are assumed known, then the ML estimator of $$\lambda$$ given by Eq. (4) always exists and is unique.
In the following, expressions for the Fisher information matrix are derived. The elements of $${\varvec{I}}_{n}(\varvec{\theta })=(I_{ij}),\; i=1, 2, 3,\; j=1, 2, 3,$$ are given by
\begin{aligned}I_{33}&=\frac{n}{\lambda ^{2}}-\frac{n \exp \lambda }{(\exp \lambda -1)^{2} } \\I_{31}&=I_{13}=-\frac{n}{\sigma }E\left( \frac{\exp \left( -\frac{x-\mu }{\sigma }\right) }{\left( 1+\exp \left( -\frac{x-\mu }{\sigma }\right) \right) ^{2} }\right) \\I_{32}&=I_{23}=-\frac{n}{\sigma }E\left( \frac{x-\mu }{\sigma } \frac{\exp \left( -\frac{x-\mu }{\sigma }\right) }{\left( 1+\exp \left( -\frac{x-\mu }{\sigma }\right) \right) ^{2} }\right) \\I_{11}&=-\frac{n}{\sigma ^{2}}E\left( \frac{1-4\exp \left( -\frac{x-\mu }{\sigma }\right) +\exp \left( \frac{-2(x-\mu )}{\sigma }\right) }{\left( 1+\exp \left( -\frac{x-\mu }{\sigma }\right) \right) ^{2}} \right) +\,\frac{n}{\sigma ^{2}}E\left( \frac{\exp \left( -\frac{x-\mu }{\sigma }\right) -1}{\exp \left( -\frac{x-\mu }{\sigma }\right) +1}\right) ^{2} +\,\frac{n\lambda }{\sigma ^{2}}E\left( \frac{\exp \left( \frac{-2(x-\mu )}{\sigma } \right) -\exp \left( -\frac{x-\mu }{\sigma }\right) }{\left( \exp \left( -\frac{x-\mu }{\sigma }\right) +1\right) ^{2}}\right) \\I_{12}&=I_{21}=-\frac{n}{\sigma ^{2}}E\left( \frac{\exp \left( -\frac{x-\mu }{\sigma }\right) -1 }{\exp \left( -\frac{x-\mu }{\sigma }\right) +1}\right) -\,\frac{n}{\sigma ^2}E \left( \frac{x-\mu }{\sigma }\frac{1-4\exp \left( -\frac{x-\mu }{\sigma }\right) +\exp \left( \frac{-2(x-\mu )}{\sigma }\right) }{\left( \exp \left( -\frac{x-\mu }{\sigma }\right) +1\right) ^{2} }\right) \\&\quad +\,\frac{n}{\sigma ^{2}}E\left( \frac{x-\mu }{\sigma } \left( \frac{\exp \left( -\frac{x-\mu }{\sigma }\right) -1}{\exp \left( -\frac{x-\mu }{\sigma }\right) +1}\right) ^{2}\right) +\,\frac{n\lambda }{\sigma ^{2}}E\left( \frac{\exp \left( -\frac{x-\mu }{\sigma }\right) }{\exp \left( -\frac{x-\mu }{\sigma }\right) +1}\right) \\I_{22}&=-\frac{n}{\sigma ^2}-\frac{2n}{\sigma ^{2}}E\left( \frac{x-\mu }{\sigma } \frac{\exp \left( -\frac{x-\mu }{\sigma }\right) -1}{\exp \left( -\frac{x-\mu }{\sigma }\right) +1}\right) -\,\frac{n}{\sigma ^{2}}E\left( \left( \frac{x-\mu }{\sigma }\right) ^{2} \frac{1-4\exp \left( -\frac{x-\mu }{\sigma }\right) +\exp \left( \frac{-2(x-\mu )}{\sigma } \right) }{\left( \exp \left( -\frac{x-\mu }{\sigma }\right) +1\right) ^{2} }\right) \\&\quad +\,\frac{n}{\sigma ^2}E\left( \left( \frac{x-\mu }{\sigma }\right) ^{2} \left( \frac{\exp \left( -\frac{x-\mu }{\sigma }\right) -1}{\exp \left( -\frac{x-\mu }{\sigma }\right) +1}\right) ^{2}\right) +\,\frac{2n\lambda }{\sigma ^{2}}E\left( \frac{x-\mu }{\sigma }\frac{\exp \left( -\frac{x-\mu }{\sigma }\right) }{\left( \exp \left( -\frac{x-\mu }{\sigma }\right) +1\right) ^{2} }\right) \\&\quad +\,\frac{n\lambda }{\sigma ^{2}}E\left( \left( \frac{x-\mu }{\sigma }\right) ^{2}\frac{\exp \left( \frac{-2(x-\mu )}{\sigma }\right) -\exp \left( -\frac{x-\mu }{\sigma }\right) }{\left( \exp \left( -\frac{x-\mu }{\sigma }\right) +1\right) ^{3} }\right) \end{aligned}
These expectations must be computed numerically. In the next section, for a real data set these expectations are computed by Monte Carlo method to provide asymptotic confidence interval for $$\varvec{\theta }=(\mu , \sigma , \lambda )^{\prime }$$.

In general, $${\widehat{\varvec{\theta }}}$$ is an asymptotically normal and asymptotically efficient estimator for $$\varvec{\theta }$$, that is, $$\sqrt{n}({\widehat{\varvec{\theta }}}-\varvec{\theta }) \overset{d}{\longrightarrow } N_3(0, {\varvec{I}}_{n}^{-1}(\varvec{\theta }))$$, where $${\varvec{I}}_{n}^{-1}(\varvec{\theta })$$ is the inverse of Fisher information matrix, see Farbod [6].

## Application

In this section, we fit the TESL distribution to a real data set. We compare the fits with other skew-logistic distributions. A real data set with 202 observations about hematocrit percent is considered. These data were collected in a study of how data on various characteristics of the blood vary with sport body size and sex of the athlete (see ais data in R). We use this data set to show that the TESL distribution can be a better model than SLD, PRHLD, ASLD, MSLD, SLND and skew-t distribution (STD). In order to compare the models, we used Akaike information criterion (AIC) and Bayesian information criterion (BIC). Table 3 lists the MLEs of the parameters from the fitted models and the values of the AIC and BIC. Based on the values of these statistics, we consider for this data set the TESL distribution is better than others models. Figure 3 shows again that the TESL distribution gives a good fit for these data.
In the following, we computed the inverse of Fisher information matrix by numerical method. The $${\varvec{I}}^{-1}_{n}$$ is as follows
\begin{aligned} {\varvec{I}}_{n}^{-1} &= \left( \begin{array}{lll} 0.0017 &\quad -\,0.0001 &\quad 0.0003\\ -\,0.0001 &\quad 0.0002 &\quad -\,0.0000\\ 0.0003 &\quad -\,0.0000 &\quad 0.0403\\ \end{array}\right) \end{aligned}
Thus, approximate 95 percent confidence intervals for $$\mu$$, $$\sigma$$ and $$\lambda$$ are $$\left( 48.6256,487935\right)$$, $$\left( 2.5690,2.6411\right)$$ and $$\left( 5.6767,6.6440\right)$$, respectively.
Table 4

Moments estimation for the TESLD

$$\lambda$$

$$\mu$$

$$\sigma$$

AE$$({\tilde{\lambda }})$$

AE$$({\tilde{\mu }})$$

AE$$({\tilde{\sigma }})$$

MSE$$({\tilde{\lambda }})$$

MSE$$({\tilde{\mu }})$$

MSE$$({\tilde{\sigma }})$$

AE$$(b({\tilde{\lambda }}))$$

AE$$(b({\tilde{\mu }}))$$

AE$$(b({\tilde{\sigma }}))$$

− 1

− 1

1

− 1.108

− 1.055

1.002

0.079

0.022

0.001

0.108

0.055

− 0.002

− 1

− 1

2

− 1.126

− 1.118

2.000

0.071

0.068

0.003

0.126

0.118

0.000

− 1

− 1

5

− 1.059

− 1.078

5.004

0.029

0.090

0.018

0.059

0.078

− 0.004

− 1

− 1

10

− 1.054

− 1.014

9.988

0.022

0.079

0.045

0.054

0.014

0.012

− 1

0

1

− 1.162

− 0.078

1.002

0.085

0.022

0.001

0.162

0.078

− 0.002

− 1

0

2

− 1.163

− 0.146

2.010

0.076

0.072

0.003

0.163

0.146

− 0.010

− 1

0

5

− 1.041

− 0.017

4.980

0.028

0.084

0.016

0.041

0.017

0.020

− 1

0

10

− 1.056

− 0.165

9.955

0.026

0.359

0.078

0.056

0.165

0.045

− 1

1

1

− 1.423

0.798

1.008

0.339

0.083

0.001

0.423

0.202

− 0.008

− 1

1

2

− 1.466

0.565

2.024

0.371

0.337

0.004

0.466

0.435

− 0.024

− 1

1

5

− 1.166

0.691

5.009

0.097

0.389

0.018

0.166

0.309

− 0.009

− 1

1

10

− 1.043

0.937

9.976

0.029

0.326

0.080

0.043

0.063

0.024

− 1

5

1

− 1.109

4.950

1.000

0.410

0.092

0.001

0.109

0.050

0.000

− 1

5

2

− 1.529

4.506

2.030

0.418

0.380

0.005

0.529

0.494

− 0.030

− 1

5

5

− 1.138

4.751

4.998

0.080

0.369

0.018

0.138

0.249

0.002

− 1

5

10

− 1.078

4.900

9.992

0.030

0.346

0.086

0.078

0.100

0.008

1

− 1

1

1.415

− 0.794

1.014

0.361

0.095

0.001

− 0.415

− 0.206

− 0.014

1

− 1

2

1.411

− 0.599

2.027

0.336

0.326

0.005

− 0.411

− 0.401

− 0.027

1

− 1

5

1.113

− 0.778

5.003

0.075

0.379

0.020

− 0.113

− 0.222

− 0.003

1

− 1

10

1.059

− 0.799

9.978

0.030

0.402

0.067

− 0.059

− 0.201

0.022

1

0

1

1.506

0.240

1.017

0.413

0.099

0.001

− 0.506

− 0.240

− 0.017

1

0

2

1.481

0.441

2.015

0.363

0.324

0.004

− 0.481

− 0.441

− 0.015

1

0

5

1.145

0.265

4.993

0.068

0.325

0.021

− 0.145

− 0.265

0.007

1

0

10

1.066

0.280

9.931

0.031

0.381

0.066

− 0.066

− 0.280

0.069

1

1

1

1.500

1.241

1.010

0.394

0.095

0.001

− 0.500

− 0.241

− 0.010

1

1

2

1.465

1.423

0.015

0.366

0.317

0.004

− 0.465

− 0.423

− 0.015

1

1

5

1.146

1.310

4.991

0.100

0.430

0.023

− 0.146

− 0.310

0.009

1

1

10

1.037

1.071

10.001

0.027

0.342

0.050

− 0.037

− 0.071

− 0.001

1

5

1

1.488

5.240

1.013

0.405

0.099

0.002

− 0.488

− 0.240

− 0.013

1

5

2

1.417

5.377

2.023

0.369

0.339

0.003

− 0.417

− 0.377

− 0.023

1

5

5

1.145

5.271

4.993

0.078

0.313

0.024

− 0.145

− 0.271

0.007

1

5

10

1.058

5.178

9.957

0.028

0.358

0.077

− 0.058

− 0.178

0.043

2

− 1

1

2.193

− 0.920

1.006

0.334

0.069

0.001

− 0.193

− 0.080

− 0.006

2

− 1

2

2.275

− 0.758

2.020

0.282

0.239

0.005

− 0.275

− 0.242

− 0.020

2

− 1

5

2.041

− 1.000

4.976

0.094

0.361

0.024

− 0.041

0.000

0.024

2

− 1

10

2.052

− 0.891

9.919

0.034

0.359

0.087

− 0.052

− 0.109

0.081

2

0

1

2.147

0.065

1.007

0.313

0.075

0.001

− 0.147

− 0.065

− 0.007

2

0

2

2.253

0.218

2.026

0.298

0.241

0.006

− 0.253

− 0.218

− 0.026

2

0

5

2.108

0.140

4.989

0.097

0.360

0.019

− 0.108

− 0.140

0.011

2

0

10

2.051

0.038

9.959

0.031

0.289

0.073

− 0.051

− 0.038

0.041

2

1

1

2.279

1.121

1.011

0.312

0.064

0.001

− 0.279

− 0.121

− 0.011

2

1

2

2.194

1.191

2.018

0.311

0.278

0.005

− 0.194

− 0.191

− 0.018

2

1

5

2.064

1.058

4.980

0.085

0.297

0.019

− 0.064

− 0.058

0.020

2

1

10

2.038

1.056

10.003

0.029

0.320

0.072

− 0.038

− 0.056

− 0.003

2

5

1

2.236

5.100

1.007

0.329

0.069

0.001

− 0.236

− 0.100

− 0.007

2

5

2

2.218

5.200

2.026

0.298

0.246

0.006

− 0.218

− 0.200

− 0.026

2

5

5

2.109

5.179

5.019

0.086

0.384

0.030

− 0.109

− 0.179

− 0.019

2

5

10

2.025

5.064

9.978

0.030

0.333

0.082

− 0.025

− 0.064

0.022

5

− 1

1

4.882

− 1.049

0.993

0.311

0.037

0.002

0.118

0.049

0.007

5

− 1

2

5.049

− 0.995

1.996

0.326

0.140

0.005

− 0.049

− 0.005

0.004

5

− 1

5

5.046

− 0.970

4.999

0.185

0.264

0.021

− 0.046

− 0.030

0.001

5

− 1

10

5.036

− 1.008

10.010

0.074

0.314

0.059

− 0.036

0.008

− 0.009

5

0

1

4.962

− 0.014

0.997

0.322

0.038

0.002

0.038

0.014

0.003

5

0

2

5.021

− 0.005

1.999

0.285

0.131

0.006

− 0.021

0.005

0.001

5

0

5

4.933

− 0.093

4.988

0.150

0.295

0.020

0.067

0.093

0.012

5

0

10

5.043

0.102

9.980

0.078

0.305

0.060

− 0.043

− 0.102

0.020

5

1

1

5.031

1.005

0.998

0.319

0.033

0.001

− 0.031

− 0.005

0.002

5

1

2

5.003

0.968

1.997

0.324

0.153

0.007

− 0.003

0.032

0.003

5

1

5

5.019

1.007

4.984

0.156

0.292

0.014

− 0.019

− 0.007

0.016

5

1

10

5.050

0.983

9.923

0.073

0.345

0.081

− 0.050

0.017

0.077

5

5

1

4.909

4.957

0.990

0.320

0.036

0.002

0.091

0.043

0.010

5

5

2

4.942

4.938

1.989

0.268

0.131

0.005

0.058

0.062

0.011

5

5

5

5.030

4.959

4.981

0.167

0.305

0.020

− 0.030

0.041

0.019

5

5

10

5.017

4.886

9.969

0.071

0.331

0.076

− 0.017

0.114

0.031

Table 5

Maximum likelihood for the TESLD

$$\lambda$$

$$\mu$$

$$\sigma$$

AE$$({\tilde{\lambda }})$$

AE$$({\tilde{\mu }})$$

AE$$({\tilde{\sigma }})$$

MSE$$({\tilde{\lambda }})$$

MSE$$({\tilde{\mu }})$$

MSE$$({\tilde{\sigma }})$$

AE$$(b({\tilde{\lambda }}))$$

AE$$(b({\tilde{\mu }}))$$

AE$$(b({\tilde{\sigma }}))$$

− 1

− 1

1

− 0.576

− 0.785

0.997

0.666

0.172

0.001

− 0.424

−0.215

0.003

− 1

− 1

2

− 0.627

− 0.644

1.987

0.610

0.587

0.005

− 0.373

−0.356

0.013

− 1

− 1

5

− 0.714

− 0.274

4.953

0.141

0.827

0.023

− 0.286

−0.726

0.047

− 1

− 1

10

− 0.847

− 0.169

9.912

0.046

0.922

0.084

− 0.153

− 0.831

0.088

− 1

0

1

− 0.518

0.235

0.991

0.457

0.110

0.001

− 0.482

− 0.235

0.009

− 1

0

2

− 0.673

0.310

1.987

0.314

0.259

0.003

− 0.327

− 0.310

0.013

− 1

0

5

− 0.977

0.064

4.981

0.020

0.033

0.019

− 0.023

− 0.064

0.019

− 1

0

10

− 0.971

0.080

9.978

0.017

0.058

0.067

− 0.029

− 0.080

0.022

− 1

1

1

− 0.676

1.153

0.997

0.644

0.159

0.001

− 0.324

− 0.153

0.003

− 1

1

2

− 1.040

0.932

2.014

0.557

0.525

0.005

0.040

0.068

− 0.014

− 1

1

5

− 1.296

0.308

5.032

0.157

0.813

0.016

0.296

0.692

− 0.032

− 1

1

10

− 1.166

0.198

10.034

0.051

0.861

0.075

0.166

0.802

− 0.034

− 1

5

1

− 0.536

5.236

0.991

0.690

0.167

0.001

− 0.464

− 0.236

0.009

− 1

5

2

− 0.747

5.225

1.995

0.498

0.419

0.003

− 0.253

− 0.225

0.005

− 1

5

5

− 1.026

4.966

4.985

0.063

0.300

0.021

0.026

0.034

0.015

− 1

5

10

− 1.002

5.057

9.911

0.027

0.364

0.080

0.002

− 0.057

0.089

1

− 1

1

0.533

− 1.231

0.997

0.670

0.167

0.001

0.467

0.231

0.003

1

− 1

2

1.031

− 0.960

2.017

0.587

0.543

0.005

− 0.031

− 0.040

− 0.017

1

− 1

5

1.285

− 0.301

5.058

0.143

0.797

0.024

− 0.285

− 0.699

− 0.058

1

− 1

10

1.193

− 0.153

10.003

0.056

0.872

0.054

− 0.193

− 0.847

− 0.003

1

0

1

1.132

0.062

1.003

0.090

0.019

0.001

− 0.132

− 0.062

− 0.003

1

0

2

1.208

0.174

2.016

0.148

0.121

0.004

− 0.208

− 0.174

− 0.016

1

0

5

1.047

0.109

4.997

0.028

0.078

0.015

− 0.047

− 0.109

0.003

1

0

10

1.019

0.045

10.012

0.016

0.032

0.083

− 0.019

− 0.045

− 0.012

1

1

1

0.631

0.810

0.996

0.606

0.155

0.001

0.369

0.190

0.004

1

1

2

0.663

0.649

2.005

0.598

0.600

0.004

0.337

0.351

− 0.005

1

1

5

0.704

0.257

4.965

0.148

0.831

0.016

0.296

0.743

0.035

1

1

10

0.817

0.134

9.911

0.053

0.927

0.071

0.183

0.866

0.089

1

5

1

0.632

4.817

0.997

0.659

0.159

0.001

0.368

0.183

0.003

1

5

2

0.695

4.725

1.996

0.559

0.480

0.004

0.305

0.275

0.004

1

5

5

1.018

5.049

4.992

0.067

0.341

0.018

− 0.018

− 0.049

0.008

1

5

10

1.035

5.133

9.955

0.025

0.298

0.072

− 0.035

− 0.133

0.045

2

− 1

1

2.082

− 0.966

1.003

0.328

0.074

0.001

− 0.082

− 0.034

− 0.003

2

− 1

2

2.097

− 0.916

2.010

0.316

0.288

0.006

− 0.097

− 0.084

− 0.010

2

− 1

5

2.263

− 0.444

5.064

0.150

0.652

0.024

− 0.263

− 0.556

− 0.064

2

− 1

10

2.172

− 0.216

10.042

0.051

0.824

0.069

− 0.172

− 0.784

− 0.042

2

0

1

2.212

0.095

1.013

0.146

0.028

0.001

− 0.212

− 0.095

− 0.013

2

0

2

2.168

0.149

2.013

0.119

0.082

0.003

− 0.168

− 0.149

− 0.013

2

0

5

2.023

0.105

4.986

0.026

0.066

0.012

− 0.023

− 0.105

0.014

2

0

10

2.049

0.051

9.975

0.021

0.040

0.077

− 0.049

− 0.051

0.025

2

1

1

2.124

1.058

1.012

0.220

0.048

0.001

− 0.124

− 0.058

− 0.012

2

1

2

1.879

0.877

2.000

0.361

0.339

0.005

0.121

0.123

0.000

2

1

5

1.767

0.439

4.959

0.150

0.724

0.023

0.233

0.561

0.041

2

1

10

1.830

0.235

9.925

0.053

0.844

0.070

0.170

0.765

0.075

2

5

1

2.095

5.035

1.005

0.328

0.066

0.001

− 0.095

− 0.035

− 0.005

2

5

2

2.113

5.097

2.019

0.259

0.227

0.006

− 0.113

− 0.097

− 0.019

2

5

5

2.065

5.087

4.979

0.072

0.321

0.014

− 0.065

−0.087

0.021

2

5

10

2.018

5.006

9.996

0.037

0.384

0.054

− 0.018

− 0.006

0.004

5

− 1

1

5.011

− 1.005

0.998

0.259

0.031

0.001

− 0.011

0.005

0.002

5

− 1

2

4.822

− 1.130

1.986

0.280

0.125

0.004

0.178

0.130

0.014

5

− 1

5

5.130

− 0.760

5.045

0.243

0.465

0.019

− 0.130

− 0.240

− 0.045

5

− 1

10

5.136

− 0.556

10.093

0.095

0.618

0.069

− 0.136

− 0.444

− 0.093

5

0

1

5.200

0.066

1.009

0.143

0.012

0.001

− 0.200

− 0.066

− 0.009

5

0

2

5.149

0.072

2.004

0.104

0.022

0.003

− 0.149

− 0.072

− 0.004

5

0

5

5.118

0.188

5.030

0.086

0.115

0.011

− 0.118

− 0.188

− 0.030

5

0

10

5.026

0.117

10.029

0.045

0.084

0.056

− 0.026

− 0.117

− 0.029

5

1

1

4.978

0.987

0.997

0.294

0.034

0.001

0.022

0.013

0.003

5

1

2

4.956

0.944

1.990

0.218

0.093

0.004

0.044

0.056

0.010

5

1

5

4.893

0.771

4.964

0.209

0.481

0.027

0.107

0.229

0.036

5

1

10

4.805

0.464

9.952

0.105

0.677

0.066

0.195

0.536

0.048

5

5

1

4.967

4.979

0.999

0.234

0.027

0.001

0.033

0.021

0.001

5

5

2

4.996

4.986

2.000

0.268

0.109

0.005

0.004

0.014

0.000

5

5

5

5.008

4.989

4.998

0.145

0.274

0.014

− 0.008

0.011

0.002

5

5

10

5.027

5.052

10.016

0.083

0.309

0.064

− 0.027

− 0.052

− 0.016

## Simulation

In this section, we evaluate the performances of the methods of moments and maximum likelihood proposed in “Estimations” section. For this purpose, by using (5) samples of size n = 100 are generated, for $$\lambda = -1, 1, 2, 5$$, $$\mu = -1, 0, 1, 5$$ and $$\sigma =1, 2, 5, 10$$. For each sample, the moments and maximum likelihood estimates are computed following the procedures described in “Estimations” section. We repeated this process 1000 times and computed the average of the estimates (AE), the average of bias (Ab) and the mean squared error (MSE). The results are reported in Tables 4 and 5.

## Conclusion

In the present paper, we introduced a new class of the skew-logistic distribution (TESL) which will be useful for analysis and modeling of unimodal data with some skewness. Some various mathematical properties of TESL distribution like moments, Rényi entropy and moment generating function are derived. We considered the TESL distribution belonging to the exponential family has closed form expressions for pdf, cdf and quantiles. In continuation, we fit the TESL distribution to a real data set and compare it with other skew-logistic distributions. Finally, we compare the performances of the methods of moments and maximum likelihood, presented in “Simulation” section. Throughout the paper, some advantages of the TESL over Azzalini’s distributions are established.

## References

1. 1.
Asgharzadeh, A., Esmaeili, L., Nadarajah, S., Shih, S.H.: A generalized skew logistic distribution. REVSTAT 11, 317–338 (2013)
2. 2.
Azzalini, A.: A class of distributions include the normal ones. Scand. J. Stat. 12, 171–178 (1985)
3. 3.
Azzalini, A.: Further results on a class of distributions which includes the normal ones. Statistica XLVI, 199–208 (1986)
4. 4.
Balakrishnan, N.: Handbook of Logistic Distribution. Dekker, New York (1992)
5. 5.
Chakraborty, S., Hazarika, P., Ali, M.M.: A new skew-logistic distribution and its properties. Pak. J. Stat. 28(4), 513–524 (2012)
6. 6.
Farbod, D.: Asymptotic properties of maximum likelihood estimators for a generalized Pareto-type distribution. J. Contemp. Math. Anal. 50(1), 44–51 (2015)
7. 7.
Ferreira, J.T.A.S., Steel, M.F.J.: A constructive representation of univariate skewed distributions. J. Am. Stat. Assoc. 101, 823–829 (2006)
8. 8.
Gupta, R.D., Kundu, D.: Generalized logistic distributions. J. Appl. Stat. Sci. 18(1), 51–66 (2010)
9. 9.
Hazarika, P., Chakraborty, S.: Alpha-skew-logistic-distribution. IOSR J. Math. 10(4), 36–46 (2014)
10. 10.
Hazarika, P., Chakraborty, S.: A note on alpha-skew-generalized logistic distribution. IOSR J. Math. 11(1), 85–88 (2015)Google Scholar
11. 11.
Nadarajah, S.: The skew logistic distribution. AStA Adv. Stat. Anal. 93, 187–203 (2009)
12. 12.
Nadarajah, S., Nassiri, V., Mohammadpour, A.: Truncated-exponential skew-symmetric distributions. Statistics 48(4), 872–895 (2014)
13. 13.
Pewsey, A.: Problem of inference for Azzalini's skew-normal distribution. J. Appl. Statist. 27, 351–360 (2000)
14. 14.
Satheesh Kumar, C., Manju, L.: A flexible class of skew logistic distribution. JIRSS 14(2), 71–92 (2015)
15. 15.
Wahed, A.S., Ali, M.M.: The skew-logistic distribution. J. Stat. Res. 35(2), 71–80 (2001)