Research Repository
Some Upper Bounds for Relative Entropy and Applications
Export Tools
Dragomir, Sever S, Scholz, M. L and Sunde, J (1998) Some Upper Bounds for Relative Entropy and Applications. RGMIA research report collection, 2 (2).
![]() |
Text
SSJS33.pdf Download (169kB) |
Abstract
In this paper we derive some upper bounds for the relative entropy D(p || q) of two probability distribution and apply them to mutual information and entropy mapping. To achieve this we use an inequality for the logarithm function, (2.3) below, and some classical inequalities such as the Kantorovič Inequality and Diaz-Metcalf Inequality.
Item Type: | Article |
---|---|
Uncontrolled Keywords: | relative entropy, mutual information, log-mapping, Kantorovič inequality, Diaz-Metcalf inequality |
Subjects: | FOR Classification > 0102 Applied Mathematics FOR Classification > 0103 Numerical and Computational Mathematics Collections > Research Group in Mathematical Inequalities and Applications (RGMIA) |
Depositing User: | Research Group in Mathematical Inequalities and Applications |
Date Deposited: | 19 Jul 2012 00:41 |
Last Modified: | 12 Dec 2014 04:26 |
URI: | http://vuir.vu.edu.au/id/eprint/17202 |
ePrint Statistics: | View download statistics for this item |
Repository staff only
![]() |
View Item |