Skip to main content
Log in

Characterizations using past entropy measures

  • Published:
METRON Aims and scope Submit manuscript

Abstract

It is reasonable to presume that in many realistic situations uncertainty is not necessarily related to the future but can also refer to the past. A measure of uncertainty in this context is the cumulative entropy defined for a non-negative random variable. In this paper we extend this definition to the case of a distributions with support in \(\mathbb {R}\). Conditions for the existence of this measure and its properties are also considered. Apart from this, certain characterization results based on past entropy measures are also discussed.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Similar content being viewed by others

References

  1. Ahmed, I.A., Kayid, M.: Characterizations of the RHR and MIT orderings and the DRHR and IMIT classes of life distributions. Probab. Eng. Inform. Sci. 19, 447–461 (2005)

    Google Scholar 

  2. Asha, G., Rejeesh, C.J.: Models characterized by the reversed lack of memory property. Calcutta Stat. Assoc. Bull. 65(233–234), 1–14 (2007)

    Google Scholar 

  3. Asha, G., Rejeesh, C.J.: Characterizations using the generalized reversed lack of memory property. Stat. Probab. Lett. 79, 1480–1487 (2009)

    Article  MATH  MathSciNet  Google Scholar 

  4. Belzunce, F., Navarro, J., Ruiz, J.M., Del Aguila, Y.: Some results on residual entropy function. Metrika 59, 147–161 (2004)

    Article  MATH  MathSciNet  Google Scholar 

  5. Block, H.W., Savits, T.H., Singh, H.: The reversed hazard rate function. Probab. Eng. Inform. Sci. 12, 69–90 (1998)

    Article  MATH  MathSciNet  Google Scholar 

  6. Di Crescenzo, A., Longobardi, M.: Entropy based measure of uncertainty in part life time distributions. J. Appl. Probab. 39, 434–440 (2002)

    Article  MATH  MathSciNet  Google Scholar 

  7. Di Crescenzo, A., Longobardi, M.: On cumulative entropies. J. Stat. Plan. Inference 139, 4072–4087 (2009)

    Article  MATH  Google Scholar 

  8. Drissi, N., Chonavel, T., Boucher, J.M.: Generalized cumulative residual entropy for distributions with unrestricted supports. Res. Lett. Signal Process. Article ID 790607, 5 p (2008)

  9. Gupta, R.D., Nanda, A.K.: \(\alpha \)- and \(\beta \)-entropies and relative entropies of distributions. J. Stat. Theory Appl. 1(3), 177–190 (2002)

  10. Johnson, D.H., Glantz, R.M.: When does interval coding occur? Neurocomputing 59–60, 13–18 (2004)

    Article  MathSciNet  Google Scholar 

  11. Khinchin, A.J.: Mathematical Foundation of Information Theory. Dover, New York (1957)

    Google Scholar 

  12. Kundu, C., Nanda, A.K., Maiti, S.S.: Some distributional results through past entropy. J. Stat Plann. Inference 140, 1280–1291 (2010)

    Article  MATH  MathSciNet  Google Scholar 

  13. Lawless, J.F.: Statistical Models and Methods for Lifetime Data. Wiley, New York (2003)

    MATH  Google Scholar 

  14. Li, X., Lu, J.: Stochastic comparison on residual life and inactivity time of series and parallel systems. Probab. Eng. Inform. Sci. 17, 267–275 (2003)

    Article  MATH  Google Scholar 

  15. Loève, M.: Probability Theory 1, 4th edn. Springer, Berlin (1977)

  16. Nair, N.U., Asha, G.: Some characterizations based on bivariate reversed mean residual life. Prob. Stat. Forum 1, 1–14 (2008)

  17. Nanda, A.K., Paul, P.: Some results on generalized past entropy. J. Stat. Plan. Inference 136, 3659–3674 (2006)

    Article  MATH  MathSciNet  Google Scholar 

  18. Nanda, A.K., Paul, P.: Some properties of past entropy and their applications. Metrika 64, 47–61 (2006)

  19. Nanda, A.K., Singh, H., Misra, N., Paul, P.: Reliability properties of reversed residual lifetime. Commun. Stat. Theory Methods 32, 2031–2042 (2003)

    Article  MATH  MathSciNet  Google Scholar 

  20. Rao, M., Chen, Y., Vemuri, B.C., Wang, F.: Cumulative residual entropy: a new measure of information. IEEE Trans. Inform. Theory 50(6), 120–128 (2004)

    MathSciNet  Google Scholar 

  21. Schroeder, M.J.: An alternative to entropy in the measurement of information. Entropy 6, 388–412 (2004)

    Article  MATH  MathSciNet  Google Scholar 

  22. Shannon, C.E.: A mathematical theory of communications. Bell Syst. Tech. J. 27, 379–423 (1948)

    Article  MATH  MathSciNet  Google Scholar 

  23. Wiener, N.: Cybernetics, 2nd edn. 1961. MIT Press and Wiley, New York (1948)

  24. Zografos, K., Nadarajah, S.: Survival exponential entropies. IEEE Transactions on Information Theory 51, 1239–1246 (2005)

    Article  MATH  MathSciNet  Google Scholar 

Download references

Acknowledgments

We are thankful to Prof. Asok Nanda, IISER, Kolkata, India, for the fruitful discussion during the preparation of the manuscript. We are also thankful to the Associate Editor for his constructive suggestions apart from finding a mistake in Theorem (2.1) in the prerevised version and the anonymous referee for his remarks, both which substantially improved the paper.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to G. Asha.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Asha, G., Rejeesh, C.J. Characterizations using past entropy measures. METRON 73, 119–134 (2015). https://doi.org/10.1007/s40300-014-0053-0

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s40300-014-0053-0

Keywords

Navigation