Hostname: page-component-78857b5c4d-76nq5 Total loading time: 0 Render date: 2023-08-04T00:21:02.039Z Has data issue: false Feature Flags: { "corePageComponentGetUserInfoFromSharedSession": true, "coreDisableEcommerce": false, "coreDisableSocialShare": false, "coreDisableEcommerceForArticlePurchase": false, "coreDisableEcommerceForBookPurchase": false, "coreDisableEcommerceForElementPurchase": false, "coreUseNewShare": false, "useRatesEcommerce": true } hasContentIssue false

ON GENERALIZED CUMULATIVE ENTROPIES

Published online by Cambridge University Press:  13 June 2016

Suchandan Kayal*
Affiliation:
Department of Mathematics, National Institute of Technology Rourkela, Rourkela-769008, India E-mail: suchandan.kayal@gmail.com, kayals@nitrkl.ac.in

Abstract

In the present paper, we introduce a generalization of the cumulative entropy proposed by Di Crescenzo and Longobardi [8]. This new notion is related to the lower records and the reversed relevation transform. Dynamic version of the newly proposed measure is considered. Several properties including the effect of linear transformations, a two-dimensional version, a normalized version, bounds, stochastic ordering, etc. are studied for the generalized cumulative entropy (GCE). Similar results are obtained for the dynamic GCE. Various relationships with other functions are derived. A class of distributions is introduced and several properties are studied. Finally, empirical GCE is proposed to estimate the newly proposed information measure.

Type
Research Article
Copyright
Copyright © Cambridge University Press 2016 

Access options

Get access to the full version of this content by using one of the access options below. (Log in options will check for institutional or personal access. Content may require purchase if you do not have access.)

References

1. Arnold, B.C., Balakrishnan, N., & Nagaraja, H.N. (1998). Records. New York: Wiley.CrossRefGoogle ScholarPubMed
2. Asadi, M. & Zohrevand, Y. (2007). On the dynamic cumulative residual entropy. Journal of Statistical Planning and Inference 137: 19311941.CrossRefGoogle Scholar
3. Belzunce, F., Navarro, J., Ruiz, J.M., & Aguila, Y.D. (2004). Some results on residual entropy function. Metrika 59: 147161.CrossRefGoogle Scholar
4. Block, H.W., Savits, T.H., & Singh, H. (1998). The reversed hazard rate function. Probability in Engineering and Informational Sciences 12: 6990.CrossRefGoogle Scholar
5. Cover, T.M. & Thomas, J.A. (2006). Elements of information theory. New York: Wiley.Google Scholar
6. Di Crescenzo, A. (2000). Some results on the proportional reversed hazards model. Statistics and Probability Letters 50: 313321.CrossRefGoogle Scholar
7. Di Crescenzo, A. & Longobardi, M. (2002). Entropy-based measure of uncertainty relies on the past. Journal of Applied Probability 39: 434440.CrossRefGoogle Scholar
8. Di Crescenzo, A. & Longobardi, M. (2009). On cumulative entropies. Journal of Statistical Planning and Inference 139: 40724087.CrossRefGoogle Scholar
9. Di Crescenzo, A. & Longobardi, M. (2012). Neuronal data analysis based on the empirical cumulative entropy. In Computer aided systems theory EUROCAST 2011. Berlin, Heidelberg: Springer, pp. 7279.CrossRefGoogle Scholar
10. Di Crescenzo, A. & Toomaj, A. (2015). Extensions of the past lifetime and its connections to the cumulative entropy. Journal of Applied Probability 52: 11561174.CrossRefGoogle Scholar
11. Ebrahimi, N. (1996). How to measure uncertainty about residual life time. Sankhya 58(A): 4857.Google Scholar
12. Ebrahimi, N. & Kirmani, S.N.U.A. (1996). Some results on ordering of survival functions through uncertainty. Statistics and Probability Letters 29: 167176.CrossRefGoogle Scholar
13. Ebrahimi, N. & Pellerey, F. (1995). New partial ordering of survival functions based on notion of uncertainty. Journal of Applied Probability 32: 202211.CrossRefGoogle Scholar
14. Gupta, R.C. & Gupta, R.D. (2007). Proportional reversed hazard rate model and its applications. Journal of Statistical Planning and Inference 137: 35253536.CrossRefGoogle Scholar
15. Kapur, J.N. (1967). Generalized entropy of order α and type β. The Mathematics Seminar 4: 7896.Google Scholar
16. Kass, R.E., Ventura, V., & Cai, C. (2003). Statistical smoothing of neural data. Network-Computation in Neural Systems 14: 515.CrossRefGoogle Scholar
17. Krakowski, M. (1973). The relevation transform and a generalization of the Gamma distribution function. Revue Francaise d'Automatique, Informatique et Recherche Opkrationnelle 7(2): 107120.CrossRefGoogle Scholar
18. Kundu, C., Nanda A.K., & Maiti, S.S. (2010). Some distributional results through past entropy. Journal of Statistical Planning and Inference 140: 12801291.CrossRefGoogle Scholar
19.Nanda A.K. & Paul, P. (2006). Some results on generalized past entropy. Journal of Statistical Planning and Inference 136: 36593674.Google Scholar
20. Navarro, J., Del Aguila, Y., & Asadi, M. (2010). Some new results on the cumulative residual entropy. Journal of Statistical Planning and Inference 140: 310322.CrossRefGoogle Scholar
21. Psarrakos, G. & Navarro, J. (2013). Generalized cumulative residual entropy and record values. Metrika 27: 623640.CrossRefGoogle Scholar
22. Pyke, R. (1965). Spacings. Journal of Royal Statistical Society, Series B 27: 395449.Google Scholar
23. Rao, M. (2005). More on a new concept of entropy and information. Journal of Theoretical Probability 18: 967981.CrossRefGoogle Scholar
24. Rao, M., Chen, Y., Vernuri, B.C., & Wang, F. (2004). Cumulative residual entropy: a new measure of information. IEEE Transactions in Information Theory 50: 12201228.CrossRefGoogle Scholar
25. Raqab, M.Z. & Asadi, M. (2010). Some results on the mean residual waiting time of records. Statistics 44: 493504.CrossRefGoogle Scholar
26. Renyi, A. (1961). On measures of entropy and information. In Proceedings of Fourth Berkeley Symposium on Mathematics, Statistics and Probability 1960. Berkeley: University of California Press, vol. 1, 547–461.Google Scholar
27. Sachlas, A. & Papaioannou, T. (2014). Residual and past entropy in actuarial science and survival models. Methodology and Computing in Applied Probability 16: 7999.CrossRefGoogle Scholar
28. Sengupta, D. & Nanda, A.K. (1999). Log-concave and concave distributions in reliability. Naval Research Logistics 46: 419433.3.0.CO;2-B>CrossRefGoogle Scholar
29. Shaked, M. & Shanthikumar, J.G. (2007). Stochastic orders. New York: Springer-Verlag.CrossRefGoogle Scholar
30. Shannon, C. (1948). The mathematical theory of communication. Bell System Technical Journal 27: 379423.CrossRefGoogle Scholar
31. Tsallis, C. (1988). Possible generalization of Boltzmann–Gibbs statistics. Journal of Statistical Physics 52: 479487.CrossRefGoogle Scholar
32. Varma, R.S. (1966). Generalization of Renyi's entropy of order α. Journal of Mathematical Sciences 1: 3448.Google Scholar
33. Wang, F. & Vemuri, B.C. (2007). Non-rigid multi-modal image registration using cross-cumulative residual entropy. International Journal of Computer Vision 74: 201215.CrossRefGoogle ScholarPubMed
34. Wang, F., Vemuri, B.C., Rao, M., & Chen, Y. (2003a). A new and robust information theoretic measure and its application to image alignment. In Proceedings of the 18th International Conference on Information Processing in Medical Imaging (IPMI ’03), vol. 2732 of Lecture Notes in Computer Science. Ambleside, UK: Springer, July 2003, pp. 388400.Google Scholar
35. Wang, F., Vemuri, B.C., Rao, M., & Chen, Y. (2003b). Cumulative residual entropy, a new measure of information & its application to image alignment. In Proceedings on the Ninth IEEE International Conference on Computer Vision (ICCV’03). IEEE Computer Society, vol. 1, pp. 548553.Google Scholar