Research Repository

Some Upper Bounds for Relative Entropy and Applications

Dragomir, Sever S and Scholz, M. L and Sunde, J (1998) Some Upper Bounds for Relative Entropy and Applications. RGMIA research report collection, 2 (2).

[img] Text
SSJS33.pdf

Download (169kB)

Abstract

In this paper we derive some upper bounds for the relative entropy D(p || q) of two probability distribution and apply them to mutual information and entropy mapping. To achieve this we use an inequality for the logarithm function, (2.3) below, and some classical inequalities such as the Kantorovič Inequality and Diaz-Metcalf Inequality.

Item Type: Article
Uncontrolled Keywords: relative entropy, mutual information, log-mapping, Kantorovič inequality, Diaz-Metcalf inequality
Subjects: FOR Classification > 0102 Applied Mathematics
FOR Classification > 0103 Numerical and Computational Mathematics
Collections > Research Group in Mathematical Inequalities and Applications (RGMIA)
Depositing User: Research Group in Mathematical Inequalities and Applications
Date Deposited: 19 Jul 2012 00:41
Last Modified: 12 Dec 2014 04:26
URI: http://vuir.vu.edu.au/id/eprint/17202
ePrint Statistics: View download statistics for this item

Repository staff only

View Item View Item

Search Google Scholar