Abstract
In this paper, Mathai–Haubold past entropy measure is proposed and its properties are studied. Also some generalized inequalities related to Mathai–Haubold entropy measure are discussed. A Kernel based non-parametric estimator for the proposed measure is provided when the underlying sample follows \(\rho\)-mixing dependence condition. The consistency property and asymptotic normality of the proposed estimator are established. A simulation study is conducted to assess the performance of the estimator. A data set is analyzed for illustrative purposes.
Similar content being viewed by others
References
Asadi M, Ebrahimi N (2000) Residual entropy and its characterizations in terms of hazard function and mean residual life function. Stat. Prob Lett 49(3):263–269
Belzunce F, Navarro J, Ruiz JM, Aguila YD (2004) Some results on residual entropy function. Metrika 59(2):147–161
Cai Z, Roussas GG (1992) Uniform strong estimation under \(\alpha\)-mixing, with rates. Stat. Prob Lett 15(1):47–55
Chakraborty S, Pradhan B (2024) Some properties of weighted survival extropy and its extended measures. Commun Stat Theory Methods 53(1):66–89
Chakraborty S (2023) On cumulative information measures: properties, inference and applications. Doctoral dissertation, Indian Statistical Institute, Kolkata
Dar JG, Al-Zahrani B (2013) On some characterization results of life time distributions using Mathai–Haubold residual entropy. IOSR J Math 5(4):56–60
Di Crescenzo A, Longobardi M (2002) Entropy-based measure of uncertainty in past lifetime distributions. J Appl Probab 39(2):434–440
Ebrahimi N (1996) How to measure uncertainty in the residual life time distribution. Sankhyā Indian J Stat Series A 58:48–56
Ebrahimi N, Kirmani SNUA (1996) Some results on ordering of survival functions through uncertainty. Stat Prob Lett 29(2):167–176
Gross AJ, Clark V (1975) Survival distributions: reliability applications in the biomedical sciences. Wiley, New York
Irshad MR, Maya R (2021) Nonparametric estimation of past extropy under \(\alpha\)-mixing dependence condition. Ricerche di Matematica, pp 1–12
Mathai AM, Haubold HJ (2007) Pathway model, superstatistics, Tsallis statistics, and a generalized measure of entropy. Phys A 375(1):110–122
Masry E (1986) Recursive probability density estimation for weakly dependent stationary processes. IEEE Trans Inf Theory 32(2):254–267
Maya R (2013) Kernel estimation of the past entropy function with dependent data. J Ker Stat Assoc 24:12–36
Maya R, Abdul-Sathar EI, Rajesh G, Nair KM (2014) Estimation of the Renyi’s residual entropy of order \(\alpha\) with dependent data. Stat Pap 55(3):585–602
Maya R, Irshad MR (2019) Kernel estimation of residual extropy function under \(\alpha\)-mixing dependence condition. S Afr Stat J 53(2):65–72
Maya R, Irshad MR (2021) Kernel estimation of Mathai–Haubold entropy and residual Mathai–Haubold entropy functions under \(\alpha\)-mixing dependence condition. Am J Math Manag Sci 41(2):148–159
Nair KM, Rajesh G (1998) Characterization of probability distributions using the residual entropy function. J Indian Statist Assoc 36:157–166
Parzen E (1962) On estimation of a probability density function and mode. Ann Math Stat 33(3):1065–1076
Rajesh G, Abdul-Sathar EI, Maya R, Nair KM (2015) Nonparametric estimation of the residual entropy function with censored dependent data. Braz J Prob Stat 29(4):866–877
Renyi A (1961) On measures of entropy and information. In: Proceedings of the fourth berkeley symposium on mathematical statistics and probability, vol 1 Contributions to the Theory of Statistics (Vol. 4). University of California Press, pp 547-562
Rosenblatt M (1956) A central limit theorem and a strong mixing condition. Proc Natl Acad Sci USA 42(1):43
Rosenblatt M (1970) Density estimates and markov sequences, Nonparametric techniques in statistical inference. Cambridge Univ. Press, London, pp 199–210
Shannon CE (1948) A mathematical theory of communication. Bell Syst Tech J 27(3):379–423
Tsallis C (1988) Possible generalization of Boltzmann–Gibbs statistics. J Stat Phys 52(1):479–487
Varma RS (1966) Generalizations of Renyi’s entropy of order \(\alpha\). J Math Sci 1(7):34–48
Funding
No funding was received for conducting this study.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
The authors declare no Conflict of interest.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Das, O., Chakraborty, S. & Pradhan, B. On Mathai–Haubold Past Entropy Measure. J Indian Soc Probab Stat (2024). https://doi.org/10.1007/s41096-024-00183-y
Accepted:
Published:
DOI: https://doi.org/10.1007/s41096-024-00183-y
Keywords
- Mathai–Haubold entropy
- Log-sum inequality
- \(\rho\)-mixing dependence condition
- Kernel function
- Asymptotic normality