No CrossRef data available.
Article contents
On an entropy conservation principle
Published online by Cambridge University Press: 14 July 2016
Abstract
We present an entropy conservation principle applicable to either discrete or continuous variables which provides a useful tool for aggregating observations. The associated method of modality grouping transforms a variable Z1 into a new variable Z2 such that the mutual information I(Z2,Y) between Y, a variable of interest, and Z2 is equal to I(Z1,Y).
Keywords
MSC classification
- Type
- Short Communications
- Information
- Copyright
- Copyright © Applied Probability Trust 1999
References
Ebrahimi, N., and Pellerey, F. (1995). New partial ordering of survival functions based on the notion of uncertainty. J. Appl. Prob.
32, 202–211.CrossRefGoogle Scholar
El Hasnaoui, A. (1993). Le concept du gain d'information: une nouvelle approche en épidémiologie quantitative. , Université de Montpellier.Google Scholar
Manuceau, J., Troupé, M., and Vaillant, J. (1999). Information and prognostic value of some variables in the breast cancer. To appear in European Series in Applied and Industrial Mathematics.Google Scholar
Rao, C.R. (1982). Diversity: Its measurement, decomposition, apportionment and analysis. Sankhya Ser. A.
44, 1–22.Google Scholar
Rao, C. R. (1986). Generalization of ANOVA through entropy and cross entropy functions. In Probability Theory and Mathematical Statistics, Vol. 2. Sciences Press, Utrecht, pp. 477–494.Google Scholar
Rényi, A. (1961). On measures of entropy and information. In Proc. 4th Berkeley Symposium on Mathematical Statistics and Probability, Vol. 1. University of California Press, Berkeley, CA, pp. 547–561.Google Scholar
Robert, C. (1990). An entropy concentration theorem: applications in artificial intelligence and descriptive statistics. J. Appl. Prob.
27, 303–313.Google Scholar
Shannon, C. E. (1948). A mathematical theory of communication. Bell System Technical Journal
27, 379–423 & 623–656.CrossRefGoogle Scholar