J I P A

advertisement
Journal of Inequalities in Pure and
Applied Mathematics
FURTHER REVERSE RESULTS FOR JENSEN’S DISCRETE
INEQUALITY AND APPLICATIONS IN INFORMATION THEORY
volume 2, issue 1, article 5,
2001.
I. BUDIMIR, S.S. DRAGOMIR AND J.E. PEČARIĆ
Department of Mathematics
Faculty of Textile Technology
University of Zagreb, CROATIA.
EMail: ivanb@zagreb.tekstil.hr
School of Communications and Informatics
Victoria University of Technology
PO Box 14428, Melbourne City MC
8001 Victoria, AUSTRALIA.
EMail: sever.dragomir@vu.edu.au
URL: http://rgmia.vu.edu.au/SSDragomirWeb.html
Department of Mathematics
Faculty of Textile Technology
University of Zagreb, CROATIA.
EMail: pecaric@mahazu.hazu.hr
URL: http://mahazu.hazu.hr/DepMPCS/indexJP.html
Received 12 April, 2000;
accepted 06 October 2000.
Communicated by: L.-E. Persson
Abstract
Contents
JJ
J
II
I
Home Page
Go Back
Close
c
2000
School of Communications and Informatics,Victoria University of Technology
ISSN (electronic): 1443-5756
007-00
Quit
Abstract
Some new inequalities which counterpart Jensen’s discrete inequality and improve the recent results from [4] and [5] are given. A related result for generalized means is established. Applications in Information Theory are also provided.
2000 Mathematics Subject Classification: 26D15, 94Xxx.
Key words: Convex functions, Jensen’s Inequality, Entropy Mappings.
Contents
1
Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3
2
Some New Counterparts for Jensen’s Discrete Inequality . . . 5
3
A Converse Inequality for Convex Mappings Defined on Rn 12
4
Some Related Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18
5
Applications in Information Theory . . . . . . . . . . . . . . . . . . . . . 23
References
Further Reverse Results for
Jensen’s Discrete Inequality
and Applications in Information
Theory
I. Budimir, S.S. Dragomir and
J. Pečarić
Title Page
Contents
JJ
J
II
I
Go Back
Close
Quit
Page 2 of 31
J. Ineq. Pure and Appl. Math. 2(1) Art. 5, 2001
http://jipam.vu.edu.au
1.
Introduction
Let f : X → R be a convex mapping
Pmdefined on the linear space X and xi ∈ X,
pi ≥ 0 (i = 1, ..., m) with Pm := i=1 pi > 0.
The following inequality is well known in the literature as Jensen’s inequality
!
m
m
1 X
1 X
(1.1)
f
pi xi ≤
pi f (xi ).
Pm i=1
Pm i=1
There are many well known inequalities which are particular cases of Jensen’s
inequality, such as the weighted arithmetic mean-geometric mean-harmonic
mean inequality, the Ky-Fan inequality, the Hölder inequality, etc. For a comprehensive list of recent results on Jensen’s inequality, see the book [25] and the
papers [9]-[15] where further references are given.
In 1994, Dragomir and Ionescu [18] proved the following inequality which
counterparts (1.1) for real mappings of a real variable.
◦
◦
Theorem 1.1. Let f : I ⊆ R → R be a differentiable convex mapping on I (I
◦
P
is the interior of I), xi ∈ I, pi ≥ 0 (i = 1, ..., n) and ni=1 pi = 1. Then we
have the inequality
!
n
n
X
X
(1.2)
0 ≤
pi f (xi ) − f
pi xi
≤
i=1
n
X
i=1
0
pi xi f (xi ) −
i=1
n
X
n
X
i=1
i=1
pi xi
Further Reverse Results for
Jensen’s Discrete Inequality
and Applications in Information
Theory
I. Budimir, S.S. Dragomir and
J. Pečarić
Title Page
Contents
JJ
J
II
I
Go Back
Close
Quit
Page 3 of 31
0
pi f (xi ),
J. Ineq. Pure and Appl. Math. 2(1) Art. 5, 2001
http://jipam.vu.edu.au
◦
where f 0 is the derivative of f on I.
Using this result and the discrete version of the Grüss inequality for weighted
sums, S.S. Dragomir obtained the following simple counterpart of Jensen’s inequality [5]:
◦
Theorem 1.2. With the above assumptions for f and if m, M ∈I and m ≤ xi ≤
M (i = 1, ..., n), then we have
!
n
n
X
X
1
(1.3) 0 ≤
pi f (xi ) − f
pi xi ≤ (M − m) (f 0 (M ) − f 0 (m)) .
4
i=1
i=1
Further Reverse Results for
Jensen’s Discrete Inequality
and Applications in Information
Theory
This was subsequently applied in Information Theory for Shannon’s and Rényi’s
entropy.
In this paper we point out some other counterparts of Jensen’s inequality that
are similar to (1.3), some of which are better than the above inequalities.
Title Page
I. Budimir, S.S. Dragomir and
J. Pečarić
Contents
JJ
J
II
I
Go Back
Close
Quit
Page 4 of 31
J. Ineq. Pure and Appl. Math. 2(1) Art. 5, 2001
http://jipam.vu.edu.au
2.
Some New Counterparts for Jensen’s Discrete
Inequality
The following result holds.
◦
Theorem 2.1. Let f : I ⊆ R → R be a differentiable convex mapping on I and
◦
P
xi ∈I with x1 ≤ x2 ≤ · · · ≤ xn and pi ≥ 0 (i = 1, ..., n) with ni=1 pi = 1.
Then we have
!
n
n
X
X
(2.1)
0 ≤
pi f (xi ) − f
pi xi
i=1
≤ (xn − x1 ) (f 0 (xn ) − f 0 (x1 )) max
1≤k≤n−1
≤
where Pk :=
Pk
i=1
I. Budimir, S.S. Dragomir and
J. Pečarić
i=1
Pk P̄k+1
1
(xn − x1 ) (f 0 (xn ) − f 0 (x1 )) ,
4
Title Page
Contents
pi and P̄k+1 := 1 − Pk .
Proof. We use the following Grüss type inequality due to J. E. Pečarić (see for
example [25]):
(2.2)
Further Reverse Results for
Jensen’s Discrete Inequality
and Applications in Information
Theory
n
n
n
1 X
X
X
1
1
q i ai b i −
q i ai ·
qi b i Qn
Q
Q
n i=1
n i=1
i=1
JJ
J
II
I
Go Back
Close
Quit
≤ |an − a1 | |bn − b1 | max
1≤k≤n−1
Qk Q̄k+1
,
Q2n
Page 5 of 31
J. Ineq. Pure and Appl. Math. 2(1) Art. 5, 2001
http://jipam.vu.edu.au
provided that a, b are
monotonic n−tuples, q is a positive one, Qn :=
P
Ptwo
n
k
i=1 qi > 0, Qk :=
i=1 qi and Q̄k+1 = Qn − Qk+1 .
If in (2.2) we choose qi = pi , ai = xi , bi = f 0 (xi ) (and ai , bi will be monotonic
nondecreasing), then we may state that
(2.3)
n
X
i=1
pi xi f 0 (xi ) −
n
X
pi xi
n
X
i=1
pi f 0 (xi )
i=1
≤ (xn − x1 ) (f 0 (xn ) − f 0 (x1 )) max
1≤k≤n−1
Pk P̄k+1 .
Now, using (1.2) and (2.3) we obtain the first inequality in (2.1).
For the second inequality, we observe that
Pk P̄k+1 = Pk (1 − Pk ) ≤
1
1
(Pk + 1 − Pk )2 =
4
4
max
I. Budimir, S.S. Dragomir and
J. Pečarić
Title Page
Contents
for all k ∈ {1, ..., n − 1} and then
1≤k≤n−1
Further Reverse Results for
Jensen’s Discrete Inequality
and Applications in Information
Theory
Pk P̄k+1
1
≤ ,
4
which proves the last part of (2.1).
Remark 2.1. It is obvious that the inequality (2.1) is an improvement of (1.3) if
we assume that the order for xi is as in the statement of Theorem 2.1.
Another result is embodied in the following theorem.
JJ
J
II
I
Go Back
Close
Quit
Page 6 of 31
J. Ineq. Pure and Appl. Math. 2(1) Art. 5, 2001
http://jipam.vu.edu.au
◦
Theorem 2.2. Let f : I ⊆ R → R be a differentiable convex mapping on I
◦
and
Pn m, M ∈I with m ≤ xi ≤ M (i = 1, ..., n) and pi ≥ 0 (i = 1, ..., n) with
i=1 pi = 1. If S is a subset of the set {1, ..., n} minimizing the expression
X
1 (2.4)
pi − ,
2
i∈S
then we have the inequality
(2.5)
0 ≤
n
X
pi f (xi ) − f
i=1
where
n
X
Further Reverse Results for
Jensen’s Discrete Inequality
and Applications in Information
Theory
!
pi xi
i=1
I. Budimir, S.S. Dragomir and
J. Pečarić
≤ Q (M − m) (f 0 (M ) − f 0 (m))
1
≤
(M − m) (f 0 (M ) − f 0 (m)) ,
4
!
X
X
Q=
pi 1 −
pi .
i∈S
Title Page
Contents
i∈S
Proof. We use the following Grüss type inequality due the Andrica and Badea
[2]:
n
n
n
X
X
X
(2.6) Qn
q i ai b i −
q i ai ·
qi b i i=1
i=1
i=1
!
X
X
≤ (M1 − m1 ) (M2 − m2 )
qi Qn −
qi
i∈S
i∈S
JJ
J
II
I
Go Back
Close
Quit
Page 7 of 31
J. Ineq. Pure and Appl. Math. 2(1) Art. 5, 2001
http://jipam.vu.edu.au
provided that m1 ≤ ai ≤ M1 , m2 ≤ bi ≤ M2 for i = 1, ..., n, and S is the
subset of {1, ..., n} which minimises the expression
X
1 qi − Qn .
2 i∈S
Choosing qi = pi , ai = xi , bi = f 0 (xi ), then we may state that
(2.7)
0 ≤
n
X
pi xi f 0 (xi ) −
i=1
n
X
pi xi
i=1
n
X
Further Reverse Results for
Jensen’s Discrete Inequality
and Applications in Information
Theory
pi f 0 (xi )
i=1
!
0
0
≤ (M − m) (f (M ) − f (m))
X
pi 1 −
X
pi .
i∈S
i∈S
Now, using (1.2) and (2.7), we obtain the first inequality in (2.5). For the last
part, we observe that
1
Q≤
4
!2
X
i∈S
pi + 1 −
X
i∈S
pi
1
=
4
and the theorem is thus proved.
The following inequality is well known in the literature as the arithmetic
mean-geometric mean-harmonic-mean inequality:
(2.8)
I. Budimir, S.S. Dragomir and
J. Pečarić
Title Page
Contents
JJ
J
II
I
Go Back
Close
Quit
Page 8 of 31
An (p, x) ≥ Gn (p, x) ≥ Hn (p, x) ,
J. Ineq. Pure and Appl. Math. 2(1) Art. 5, 2001
http://jipam.vu.edu.au
where
An (p, x) : =
n
X
pi xi - the arithmetic mean,
i=1
Gn (p, x) : =
n
Y
xpi i - the geometric mean,
i=1
1
Hn (p, x) : = P
- the harmonic mean,
n
pi
i=1
xi
P
and ni=1 pi = 1 pi ≥ 0, i = 1, n .
Using the above two theorems, we are able to point out the following reverse
of the AGH - inequality.
P
Proposition 2.3. Let xi > 0 (i = 1, ..., n) and pi ≥ 0 with ni=1 pi = 1.
(i) If x1 ≤ x2 ≤ · · · ≤ xn−1 ≤ xn , then we have
(2.9)
An (p, x)
Gn (p, x)
"
#
2
(xn − x1 )
≤ exp
max Pk P̄k+1
x1 xn 1≤k≤n−1
#
"
2
1 (xn − x1 )
·
.
≤ exp
4
x1 xn
1 ≤
Further Reverse Results for
Jensen’s Discrete Inequality
and Applications in Information
Theory
I. Budimir, S.S. Dragomir and
J. Pečarić
Title Page
Contents
JJ
J
II
I
Go Back
Close
Quit
Page 9 of 31
J. Ineq. Pure and Appl. Math. 2(1) Art. 5, 2001
http://jipam.vu.edu.au
(ii) If the set S ⊆ {1, ..., n} minimizes the expression (2.4), and 0 < m ≤
xi ≤ M < ∞ (i = 1, ..., n), then
(2.10)
1 ≤
An (p, x)
Gn (p, x)
"
(M − m)2
≤ exp Q ·
mM
#
"
#
1 (M − m)2
≤ exp
.
·
4
mM
The proof goes by the inequalities (2.1) and (2.5), choosing f (x) = − ln x.
A similar result can be stated for Gn and Hn .
P
Proposition 2.4. Let p ≥ 1 and xi > 0, pi ≥ 0 (i = 1, ..., n) with ni=1 pi = 1.
(i) If x1 ≤ x2 ≤ · · · ≤ xn−1 ≤ xn , then we have
!p
n
n
X
X
(2.11)
0 ≤
pi xpi −
pi xi
i=1
≤ p (xn − x1 )
i=1
xp−1
n
− x1p−1
Contents
max
1≤k≤n−1
Pk P̄k+1
(ii) If the set S ⊆ {1, ..., n} minimizes the expression (2.4), and 0 < m ≤
xi ≤ M < ∞ (i = 1, ..., n), then
!p
n
n
X
X
p
(2.12)
0≤
pi xi −
pi xi
i=1
I. Budimir, S.S. Dragomir and
J. Pečarić
Title Page
p
(xn − x1 ) xp−1
− xp−1
.
≤
1
n
4
i=1
Further Reverse Results for
Jensen’s Discrete Inequality
and Applications in Information
Theory
JJ
J
II
I
Go Back
Close
Quit
Page 10 of 31
J. Ineq. Pure and Appl. Math. 2(1) Art. 5, 2001
http://jipam.vu.edu.au
≤ pQ (M − m) M p−1 − mp−1
1
≤
p (M − m) M p−1 − mp−1 .
4
Remark 2.2. The above results are improvements of the corresponding inequalities obtained in [5].
Remark 2.3. Similar inequalities can be stated if we choose other convex functions such as: f (x) = x ln x, x > 0 or f (x) = exp (x), x ∈ R. We omit the
details.
Further Reverse Results for
Jensen’s Discrete Inequality
and Applications in Information
Theory
I. Budimir, S.S. Dragomir and
J. Pečarić
Title Page
Contents
JJ
J
II
I
Go Back
Close
Quit
Page 11 of 31
J. Ineq. Pure and Appl. Math. 2(1) Art. 5, 2001
http://jipam.vu.edu.au
3.
A Converse Inequality for Convex Mappings Defined on Rn
In 1996, Dragomir and Goh [15] proved the following converse of Jensen’s
inequality for convex mappings on Rn .
Theorem 3.1. Let f : Rn → R be a differentiable convex mapping on Rn and
∂f (x)
∂f (x)
(∇f ) (x) :=
, ...,
,
∂x1
∂xn
the vector of the partial derivatives, x = (x1 , ..., xn ) ∈ Rn . P
If xi ∈ Rm (i = 1, ..., m), pi ≥ 0, i = 1, ..., m, with Pm := m
i=1 pi > 0, then
!
m
m
1 X
1 X
(3.1) 0 ≤
pi f (xi ) − f
pi xi
Pm i=1
Pm i=1
*
+
m
m
m
1 X
1 X
1 X
≤
pi h∇f (xi ), xi i −
pi ∇f (xi ),
pi xi .
Pm i=1
Pm i=1
Pm i=1
The result was applied to different problems in Information Theory by providing
different counterpart inequalities for Shannon’s entropy, conditional entropy,
mutual information, conditional mutual information, etc.
For generalizations of (3.1) in Normed Spaces and other applications in Information Theory, see Matić’s Ph.D dissertation [23].
Recently, Dragomir [4] provided an upper bound for Jensen’s difference
!
m
m
1 X
1 X
(3.2)
∆ (f, p, x) :=
pi f (xi ) − f
pi xi ,
Pm i=1
Pm i=1
Further Reverse Results for
Jensen’s Discrete Inequality
and Applications in Information
Theory
I. Budimir, S.S. Dragomir and
J. Pečarić
Title Page
Contents
JJ
J
II
I
Go Back
Close
Quit
Page 12 of 31
J. Ineq. Pure and Appl. Math. 2(1) Art. 5, 2001
http://jipam.vu.edu.au
which, even though it is not as sharp as (3.1), provides a simpler way, and for
applications, a better way, of estimating the Jensen’s differences ∆. His result
is embodied in the following theorem.
Theorem 3.2. Let f : Rn → R be a differentiable convex mapping and xi ∈ Rn ,
i = 1, ..., m. Suppose that there exists the vectors φ, Φ ∈ Rn such that
(3.3)
φ ≤ xi ≤ Φ (the order is considered on the co-ordinates)
Further Reverse Results for
Jensen’s Discrete Inequality
and Applications in Information
Theory
and m, M ∈ Rn are such that
m ≤ ∇f (xi ) ≤ M
(3.4)
for all i ∈ {1, ..., m}. Then for all pi ≥ 0 (i = 1, ..., m) with Pm > 0, we have
the inequality
!
m
m
1
1 X
1 X
pi f (xi ) − f
pi xi ≤ kΦ − φk kM − mk ,
(3.5) 0 ≤
Pm i=1
Pm i=1
4
n
where k·k is the usual Euclidean norm on R .
He applied this inequality to obtain different upper bounds for Shannon’s
and Rényi’s entropies.
In this section, we point out another counterpart for Jensen’s difference, assuming that the ∇−operator is of Hölder’s type, as follows.
n
n
Theorem 3.3. Let f : R → R be a differentiable convex mapping and xi ∈ R ,
pi ≥ 0 (i = 1, ..., m) with Pm > 0. Suppose that the ∇−operator satisfies a
condition of r − H−Hölder type, i.e.,
(3.6)
k∇f (x) − ∇f (y)k ≤ H kx − ykr , for all x, y ∈ Rn ,
I. Budimir, S.S. Dragomir and
J. Pečarić
Title Page
Contents
JJ
J
II
I
Go Back
Close
Quit
Page 13 of 31
J. Ineq. Pure and Appl. Math. 2(1) Art. 5, 2001
http://jipam.vu.edu.au
where H > 0, r ∈ (0, 1] and k·k is the Euclidean norm.
Then we have the inequality:
(3.7)
m
m
1 X
1 X
0 ≤
pi f (xi ) − f
pi xi
Pm i=1
Pm i=1
H X
≤
pi pj kxi − xj kr+1 .
2
Pm 1≤i<j≤m
!
Further Reverse Results for
Jensen’s Discrete Inequality
and Applications in Information
Theory
Proof. We recall Korkine’s identity,
m
1 X
pi hyi , xi i −
Pm i=1
*
m
m
1 X
1 X
pi yi ,
pi xi
Pm i=1
Pm i=1
+
I. Budimir, S.S. Dragomir and
J. Pečarić
n
1 X
=
pi pj hyi − yj , xi − xj i , x, y ∈ Rn ,
2
2Pm i,j=1
and simply write
m
1 X
pi h∇f (xi ), xi i −
Pm i=1
*
m
m
1 X
1 X
pi ∇f (xi ),
pi xi
Pm i=1
Pm i=1
+
n
1 X
=
pi pj h∇f (xi ) − ∇f (xj ), xi − xj i .
2Pm2 i,j=1
Title Page
Contents
JJ
J
II
I
Go Back
Close
Quit
Page 14 of 31
J. Ineq. Pure and Appl. Math. 2(1) Art. 5, 2001
http://jipam.vu.edu.au
Using (3.1) and the properties of the modulus, we have
!
m
m
1 X
1 X
0 ≤
pi f (xi ) − f
pi xi
Pm i=1
Pm i=1
m
1 X
≤
pi pj |h∇f (xi ) − ∇f (xj ), xi − xj i|
2Pm2 i,j=1
m
1 X
≤
pi pj k∇f (xi ) − ∇f (xj )k kxi − xj k
2Pm2 i,j=1
Further Reverse Results for
Jensen’s Discrete Inequality
and Applications in Information
Theory
m
H X
pi pj kxi − xj kr+1
≤
2
Pm i,j=1
I. Budimir, S.S. Dragomir and
J. Pečarić
and the inequality (3.7) is proved.
Title Page
Corollary 3.4. With the assumptions of Theorem 3.3 and if
∆ = max1≤i<j≤m kxi − xj k, then we have the inequality
!
m
m
1 X
1 X
H∆r+1
pi f (xi ) − f
pi xi ≤
(3.8) 0 ≤
Pm i=1
Pm i=1
2Pm2
Contents
1−
m
X
i=1
!
p2i
.
JJ
J
II
I
Go Back
Close
Proof. Indeed, as
Quit
X
1≤i<j≤m
r+1
pi pj kxi − xj k
r+1
≤∆
X
pi pj .
Page 15 of 31
1≤i<j≤m
J. Ineq. Pure and Appl. Math. 2(1) Art. 5, 2001
http://jipam.vu.edu.au
However,
1
pi pj =
2
1≤i<j≤m
X
m
X
!
pi pj −
i,j=1
X
pi pj
i=j
1
=
2
1−
m
X
!
p2i
,
i=1
and the inequality (3.8) is proved.
The case of Lipschitzian mappings is embodied in the following corollary.
Corollary 3.5. Let f : Rn → R be a differentiable convex mapping and
xi ∈ Rn , pi ≥ 0 (i = 1, ..., n) with Pm > 0. Suppose that the ∇−operator
is Lipschitzian with the constant L > 0, i.e.,
Further Reverse Results for
Jensen’s Discrete Inequality
and Applications in Information
Theory
k∇f (x) − ∇f (y)k ≤ L kx − yk , for all x, y ∈ Rn ,
I. Budimir, S.S. Dragomir and
J. Pečarić
(3.9)
where k·k is the Euclidean norm. Then
(3.10)
!
m
m
1 X
1 X
0 ≤
pi f (xi ) − f
pi xi
Pm i=1
Pm i=1

2 
m
m
X
X
1
1
2

pi kxi k − pi xi  .
≤ L
Pm
Pm i=1
i=1
Proof. The argument is obvious by Theorem 3.3, taking into account that for
r = 1,
2
m
m
X
X
X
2
2
pi pj kxi − xj k = Pm
pi kxi k − pi xi ,
1≤i<j≤m
and k·k is the Euclidean norm.
i=1
Title Page
Contents
JJ
J
II
I
Go Back
Close
Quit
Page 16 of 31
i=1
J. Ineq. Pure and Appl. Math. 2(1) Art. 5, 2001
http://jipam.vu.edu.au
Moreover, if we assume more about the vectors (xi )i=1,n , we can obtain a
simpler result that is similar to the one in [4].
Corollary 3.6. Assume that f is as in Corollary 3.5. If
(3.11)
φ ≤ xi ≤ Φ (on the co-ordinates), φ, Φ ∈ Rn (i = 1, .., m) ,
then we have the inequality
(3.12)
0 ≤
≤
1
Pm
m
X
pi f (xi ) − f
i=1
1
Pm
m
X
!
pi xi
Further Reverse Results for
Jensen’s Discrete Inequality
and Applications in Information
Theory
i=1
1
· L · kΦ − φk2 .
4
Proof. It follows by the fact that in Rn , we have the following Grüss type inequality (as proved in [4])
2
m
m
1 X
1 X
1
2
(3.13)
pi kxi k − pi xi ≤ kΦ − φk2 ,
Pm
Pm i=1
4
i=1
provided that (3.11) holds.
Remark 3.1. For some Grüss type inequalities in Inner Product Spaces, see [7].
I. Budimir, S.S. Dragomir and
J. Pečarić
Title Page
Contents
JJ
J
II
I
Go Back
Close
Quit
Page 17 of 31
J. Ineq. Pure and Appl. Math. 2(1) Art. 5, 2001
http://jipam.vu.edu.au
4.
Some Related Results
Start with the following definitions from [3].
Definition 4.1. Let −∞ < a < b < ∞. Then CM [a, b] denotes the set of all
functions with domain [a, b] that are continuous and strictly monotonic there.
Definition 4.2. Let −∞ < a < b < ∞, and let f ∈ CM [a, b]. Then, for
each positive integer n, each n−tuple x = (x1 , ..., xn ) , where a ≤ xj ≤ b
(j =P
1, 2, ..., n), and each n-tuple p = (p1 , p2 , ..., pn ) , where pj > 0 (j = 1, 2, ..., n)
and nj=1 pj = 1, let Mf (x, y) denote the (weighted) mean
( n
)
X
f −1
pj f (xj ) .
Further Reverse Results for
Jensen’s Discrete Inequality
and Applications in Information
Theory
I. Budimir, S.S. Dragomir and
J. Pečarić
j=1
Title Page
We may state now the following result.
Theorem
4.1. Let
P
S be the subset of {1, ..., n} which minimizes the expression
i∈S pi − 1/2 . If f, g ∈ CM [a, b], then
0 00 sup {|Mf (x, p) − Mg (x, p)|} ≤ Q· f −1 · f ◦ g −1 ·|g(b) − g(a)|2 ,
∞
x
∞
provided that the right-hand side of the inequality is finite, where, as above,
!
!
X
X
Q=
pi
1−
pi ,
i∈S
and k·k∞ is the usual sup-norm.
i∈S
Contents
JJ
J
II
I
Go Back
Close
Quit
Page 18 of 31
J. Ineq. Pure and Appl. Math. 2(1) Art. 5, 2001
http://jipam.vu.edu.au
Proof. Let, as in [3], h = f ◦ g −1 , n > 1,
x = (x1 , x2 , ..., xn ) and p = (p1 , p2 , ..., pn )
be as in the Definition 4.2, and yj = g (xj ) (j = 1, 2, ..., n). By the mean-value
theorem, for some α in the open interval joining f (a) to f (b), we have
( n
)
" ( n
)#
X
X
Mf (x, p) − Mg (x, p) = f −1
pj f (xj ) − f −1 h
pj g(xj )
j=1
=
n
X
0
f −1 (α)
pj f (xj ) − h
"
=
0
f −1 (α)
( n
X
j=1
j=1
n
X
( n
X
pj h(yj ) − h
j=1
"
=
0
f −1 (α)
Further Reverse Results for
Jensen’s Discrete Inequality
and Applications in Information
Theory
j=1
"
n
X
j=1
)#
pj g(xj )
I. Budimir, S.S. Dragomir and
J. Pečarić
)#
pj yj
j=1
(
pj
h(yj ) − h
n
X
Title Page
!)#
pk yk
.
k=1
Using the mean-value theorem a second time, we conclude that there exists
points z1 , z2 , ..., zn in the open interval joining g(a) to g(b), such that
Mf (x, p) − Mg (x, p)
0
= f −1 (α) p1 {(1 − p1 ) y1 − p2 y2 − · · · − pn yn } h0 (z1 )
+ p2 {−p1 y1 + (1 − p2 ) y2 − · · · − pn yn } h0 (z2 )
+ ···
+ pn {−p1 y1 − p2 y2 − · · · + (1 − pn ) yn } h0 (zn )
Contents
JJ
J
II
I
Go Back
Close
Quit
Page 19 of 31
J. Ineq. Pure and Appl. Math. 2(1) Art. 5, 2001
http://jipam.vu.edu.au
0
f −1 (α) p1 {p2 (y1 − y2 ) + · · · + pn (y1 − yn )} h0 (z1 )
+p2 {p1 (y2 − y1 ) + · · · + pn (y2 − yn )} h0 (z2 )
+···
+pn {p1 (yn − y1 ) + · · · + pn−1 (yn − yn−1 )} h0 (zn )
X
0
= f −1 (α)
pi pj (yi − yj ) {h0 (zi ) − h0 (zj )} .
=
1≤i<j≤n
Using the mean value theorem a third time, we conclude that there exists points
ωij (1 ≤ i < j ≤ n) in the open interval joining g(a) to g(b), such that
X
0
f −1 (α)
pi pj (yi − yj ) {h0 (zi ) − h0 (zj )}
Further Reverse Results for
Jensen’s Discrete Inequality
and Applications in Information
Theory
I. Budimir, S.S. Dragomir and
J. Pečarić
1≤i<j≤n
X
0
pi pj (yi − yj ) (zi − zj ) h00 (ωij ).
= f −1 (α)
1≤i<j≤n
Contents
Consequently,
|Mf (x, p) − Mg (x, p)|
X
−1 0
≤ f
(α)
pi pj |yi − yj | · |zi − zj | · |h00 (ωij )|
1≤i<j≤n
0 ≤ f −1 · kh00 k∞ ·
∞
Title Page
JJ
J
II
I
Go Back
Close
X
pi pj |yi − yj | · |zi − zj |
Quit
1≤i<j≤n
≤ (by the Cauchy-Buniakowski-Schwartz inequality)
Page 20 of 31
J. Ineq. Pure and Appl. Math. 2(1) Art. 5, 2001
http://jipam.vu.edu.au
s X
−1 0 −1 00 ≤ f
pi pj |yi − yj |2
· f ◦g
·
∞
∞
1≤i<j≤n
s X
pi pj |zi − zj |2
·
1≤i<j≤n
≤ (by the Andrica and Badea result)
v
!
!
u
X
X
u
00
−1 0 ≤ f
pi
1−
pi |g(b) − g(a)|2
· f ◦ g −1 · t
∞
∞
i∈S
i∈S
v
!
!
u
X
u X
·t
1−
pi
pi |g(b) − g(a)|2
i∈S
Further Reverse Results for
Jensen’s Discrete Inequality
and Applications in Information
Theory
I. Budimir, S.S. Dragomir and
J. Pečarić
i∈S
−1 0 2
−1 00 = Q f
· f ◦g
· |g(b) − g(a)| ,
∞
Title Page
∞
Contents
and the theorem is proved.
Corollary 4.2. If f, g ∈ CM [a, b], then
0 0 1 1 f 2
sup {|Mf (x, p) − Mg (x, p)|} ≤ Q · f 0 · g 0 g 0 · |g(b) − g(a)| ,
x
∞
∞
provided that the right hand side of the inequality exists.
Proof. This follows at once from the fact that
f
−1 0
=
f0
1
◦ f −1
JJ
J
II
I
Go Back
Close
Quit
Page 21 of 31
J. Ineq. Pure and Appl. Math. 2(1) Art. 5, 2001
http://jipam.vu.edu.au
and
f ◦g
−1 00
0 0 (g 0 ◦ g −1 ) (f 00 ◦ g −1 ) − (f 0 ◦ g −1 ) (g 00 ◦ g −1 )
1 f
◦g −1 .
=
= 0
3
g g0
(g 0 ◦ g −1 )
Remark 4.1. This establishes Theorem 4.3 from [3] and replaces the multiplicative factor 14 by Q. In Corollary 4.2, we also replaced the multiplicative factor
1
by Q.
4
Further Reverse Results for
Jensen’s Discrete Inequality
and Applications in Information
Theory
I. Budimir, S.S. Dragomir and
J. Pečarić
Title Page
Contents
JJ
J
II
I
Go Back
Close
Quit
Page 22 of 31
J. Ineq. Pure and Appl. Math. 2(1) Art. 5, 2001
http://jipam.vu.edu.au
5.
Applications in Information Theory
We give some new applications for Shannon’s entropy
Hb (X) :=
r
X
i=1
pi logb
1
,
pi
where X is a random variable with the probability distribution (pi )i=1,r .
Theorem 5.1. Let X be as above and assume that p1 ≥ p2 ≥ · · · ≥ pr or
p1 ≤ p2 ≤ · · · ≤ pr . Then we have the inequality
(5.1)
(p1 − pr )2
0 ≤ logb r − Hb (X) ≤
max Pk P̄k+1 .
1≤k≤r
p1 pr
Proof. We choose in Theorem 2.1, f (x) = − logb x, x > 0, xi = p1i (i = 1, ..., r).
Then we have x1 ≤ x2 ≤ · · · ≤ xr and by (2.1) we obtain
!
1
1
1
1
0 ≤ logb r − Hb (X) ≤
−
max Pk P̄k+1 ,
1 + 1
1≤k≤r
pr p1
− pr
p1
which is equivalent to (5.1). The same inequality is obtained if p1 ≤ p2 ≤ · · · ≤
pr .
Theorem 5.2. Let X be as above and suppose that
pM : = max {pi |i = 1, ..., r} ,
pm : = min {pi |i = 1, ..., r} .
Further Reverse Results for
Jensen’s Discrete Inequality
and Applications in Information
Theory
I. Budimir, S.S. Dragomir and
J. Pečarić
Title Page
Contents
JJ
J
II
I
Go Back
Close
Quit
Page 23 of 31
J. Ineq. Pure and Appl. Math. 2(1) Art. 5, 2001
http://jipam.vu.edu.au
P
If S is a subset of the set {1, ..., r} minimizing the expression i∈S pi − 1/2,
then we have the estimation
(pM − pm )2
0 ≤ logb r − Hb (X) ≤ Q ·
.
ln b · pM pm
(5.2)
Proof. We shall choose in Theorem 2.2,
1
pi
f (x) = − logb x, x > 0, xi =
Then m =
1
,
pM
M=
1
,
pm
i = 1, r .
1
f 0 (x) = − x ln
and the inequality (2.3) becomes:
b
0 ≤ logb r −
r
X
≤ Q
1
ln b
1
1
−
pm pM
Title Page
−
1
1
pm
+
2
= Q·
1 (pM − pm )
·
,
ln b
pM pm
1
pM
Contents
JJ
J
II
I
Close
Consider the Shannon entropy
H (X) := He (X) =
1
!
Go Back
hence the estimation (5.2) is proved.
(5.3)
I. Budimir, S.S. Dragomir and
J. Pečarić
pi logb
i=1
1
pi
Further Reverse Results for
Jensen’s Discrete Inequality
and Applications in Information
Theory
Quit
r
X
i=1
pi ln
1
pi
Page 24 of 31
J. Ineq. Pure and Appl. Math. 2(1) Art. 5, 2001
http://jipam.vu.edu.au
and Rényi’s entropy of order α (α ∈ (0, ∞) \ {1})
(5.4)
r
X
1
H[α] (X) :=
ln
pαi
1−α
i=1
!
.
Using the classical Jensen’s discrete inequality for convex mappings, i.e.,
!
r
r
X
X
(5.5)
f
pi xi ≤
pi f (xi ),
i=1
i=1
where f : I ⊆ R → R is a convex mapping on I, xi ∈ I (i = 1, ..., r) and
(pi )i=1,r is a probability distribution, for the convex mapping f (x) = − ln x,
we have
!
r
r
X
X
(5.6)
ln
pi xi ≥
pi ln xi .
i=1
I. Budimir, S.S. Dragomir and
J. Pečarić
Title Page
Contents
i=1
Choose xi = pα−1
(i = 1, ..., r) in (5.6) to obtain
i
!
r
r
X
X
α
ln
pi ≥ (α − 1)
pi ln pi ,
i=1
Further Reverse Results for
Jensen’s Discrete Inequality
and Applications in Information
Theory
i=1
JJ
J
II
I
Go Back
Close
Quit
which is equivalent to
(1 − α) H[α] (X) − H (X) ≥ 0.
Page 25 of 31
J. Ineq. Pure and Appl. Math. 2(1) Art. 5, 2001
http://jipam.vu.edu.au
Now, if α ∈ (0, 1) , then H[α] (X) ≤ H (X) , and if α > 1 then H[α] (X) ≥
H (X) . Equality holds iff (pi )i=1,r is a uniform distribution and this fact follows
by the strict convexity of − ln (·) . This inequality also follows as a special case
of the following well known fact: H[α] (X) is a nondecreasing function of α.
See for example [26] or [22].
Theorem 5.3. Under the above assumptions, given that pm = mini=1,r pi ,
pM = maxi=1,r pi , then we have the inequality
2
α−1
pM
− pα−1
m
(5.7)
0 ≤ (1 − α) H[α] (X) − H (X) ≤ Q ·
,
α−1
pα−1
M pm
Further Reverse Results for
Jensen’s Discrete Inequality
and Applications in Information
Theory
I. Budimir, S.S. Dragomir and
J. Pečarić
for all α ∈ (0, 1) ∪ (1, ∞).
Proof. If α ∈ (0, 1), then
Title Page
α−1
xi := pα−1
∈ pα−1
i
M , pm
Contents
and if α ∈ (1, ∞), then
α−1
xi = pα−1
∈ pα−1
, for i ∈ {1, ..., n} .
m , pM
i
Applying Theorem 2.2 for xi := pα−1
and f (x) = − ln x, and taking into
i
account that f 0 (x) = − x1 , we obtain
(1 − α) H[α] (X) − H (X)

1
α−1
1
α−1

Q
p
−
p
−
+
if α ∈ (0, 1) ,

m
M

pα−1
pα−1
m
M
≤

1

 Q pα−1 − pα−1 − α−1
+
m
M
p
M
1
pα−1
m
JJ
J
II
I
Go Back
Close
Quit
Page 26 of 31
if α ∈ (1, ∞)
J. Ineq. Pure and Appl. Math. 2(1) Art. 5, 2001
http://jipam.vu.edu.au
=

α−1 2
(pα−1
m −pM )


Q
·
if α ∈ (0, 1) ,

α−1
α−1

pm pM

α−1
α−1 2

m )

 Q · (pMα−1−pα−1
if α ∈ (1, ∞)
p
p
M
m
α−1
pα−1
M − pm
=Q·
α−1
pα−1
M pm
2
for all α ∈ (0, 1) ∪ (1, ∞) and the theorem is proved.
Further Reverse Results for
Jensen’s Discrete Inequality
and Applications in Information
Theory
Using a similar argument to the one in Theorem 5.3, we can state the following direct application of Theorem 2.2.
I. Budimir, S.S. Dragomir and
J. Pečarić
Theorem 5.4. Let (pi )i=1,r be as in Theorem 5.3. Then we have the inequality
Title Page
(5.8)
α−1 2
pα−1
M − pm
0 ≤ (1 − α) H[α] (X) − ln r − α ln Gr (p) ≤ Q ·
,
α−1 α−1
PM
pm
for all α ∈ (0, 1) ∪ (1, ∞).
Contents
JJ
J
II
I
Go Back
Remark 5.1. The above results improve the corresponding results from [5] and
[4] with the constant Q which is less than 41 .
Acknowledgement 1. The authors would like to thank the anonymous referee
for valuable comments and for the references [26] and [22].
Close
Quit
Page 27 of 31
J. Ineq. Pure and Appl. Math. 2(1) Art. 5, 2001
http://jipam.vu.edu.au
References
[1] A. RÉNYI, On measures of entropy and information, Proc. Fourth Berkley
Symp. Math. Statist. Prob., 1 (1961), 547–561, Univ. of California Press,
Berkley.
[2] D. ANDRICA AND C. BADEA, Grüss’ inequality for positive linear functionals, Periodica Math. Hung., 19(2) (1988), 155–167.
[3] G.T. CARGO AND O. SHISHA, A metric space connected with general
means, J. Approx. Th., 2 (1969), 207–222.
Further Reverse Results for
Jensen’s Discrete Inequality
and Applications in Information
Theory
[4] S.S. DRAGOMIR, A converse of the Jensen inequality for convex mappings of several variables and applications. (Electronic Preprint:
http://matilda.vu.edu.au/~rgmia/InfTheory/
ConverseJensen.dvi)
I. Budimir, S.S. Dragomir and
J. Pečarić
[5] S.S. DRAGOMIR, A converse result for Jensen’s discrete inequality via
Grüss’ inequality and applications in information theory, Analele Univ.
Oradea, Fasc. Math., 7 (1999-2000), 178–189.
Contents
[6] S.S. DRAGOMIR, A further improvement of Jensen’s inequality, Tamkang
J. Math., 25(1) (1994), 29–36.
[7] S.S. DRAGOMIR, A generalisation of Grüss’s inequality in inner product
spaces and applications, J. Math. Anal. Appl., 237 (1999), 74–82.
[8] S.S. DRAGOMIR, A new improvement of Jensen’s inequality, Indian J.
Pure Appl. Math., 26(10) (1995), 959–968.
Title Page
JJ
J
II
I
Go Back
Close
Quit
Page 28 of 31
J. Ineq. Pure and Appl. Math. 2(1) Art. 5, 2001
http://jipam.vu.edu.au
[9] S.S. DRAGOMIR, An improvement of Jensen’s inequality, Bull. Math.
Soc. Sci. Math. Romania., 34 (1990), 291–296.
[10] S.S. DRAGOMIR, On some refinements of Jensen’s inequality and applications, Utilitas Math., 43 (1993), 235–243.
[11] S.S. DRAGOMIR, Some refinements of Jensen’s inequality, J. Math. Anal.
Appl., 168 (1992), 518–522.
[12] S.S. DRAGOMIR, Some refinements of Ky Fan’s inequality, J. Math.
Anal. Appl., 163 (1992), 317–321.
Further Reverse Results for
Jensen’s Discrete Inequality
and Applications in Information
Theory
[13] S.S. DRAGOMIR, Two mappings in connection with Jensen’s inequality,
Extracta Math., 8(2-3) (1993), 102–105.
I. Budimir, S.S. Dragomir and
J. Pečarić
[14] S.S. DRAGOMIR AND S. FITZPARTICK, s - Orlicz convex functions in
linear spaces and Jensen’s discrete inequality, J. Math. Anal. Appl., 210
(1997), 419–439.
[15] S.S. DRAGOMIR AND C.J. GOH, A counterpart of Jensen’s discrete inequality for differentiable convex mappings and applications in Information Theory, Math. Comput. Modelling, 24(2) (1996), 1–11.
Title Page
Contents
JJ
J
II
I
Go Back
[16] S.S. DRAGOMIR AND C.J. GOH, Some bounds on entropy measures in
information theory, Appl. Math. Lett., 10(3) (1997), 23–28.
[17] S.S. DRAGOMIR AND C.J. GOH, Some counterpart inequalities for a
functional associated with Jensen’s inequality, J. Inequal. Appl., 1 (1997),
311–325.
Close
Quit
Page 29 of 31
J. Ineq. Pure and Appl. Math. 2(1) Art. 5, 2001
http://jipam.vu.edu.au
[18] S.S. DRAGOMIR AND N.M. IONESCU, Some Converse of Jensen’s inequality and applications, Anal. Num. Theor. Approx. (Cluj-Napoca), 23
(1994), 71–78.
[19] S.S. DRAGOMIR, C.E.M. PEARCE AND J. E. PEČARIĆ, New inequalities for logarithmic map and their applications for entropy and mutual
information, Kyungpook Math. J., in press.
[20] S.S. DRAGOMIR, C.E.M. PEARCE AND J. E. PEČARIĆ, On Jensen’s
and related inequalities for isotonic sublinear functionals, Acta Sci. Math.,
(Szeged), 61 (1995), 373–382.
Further Reverse Results for
Jensen’s Discrete Inequality
and Applications in Information
Theory
[21] S.S. DRAGOMIR, J. E. PEČARIĆ AND L.E. PERSSON, Properties of
some functionals related to Jensen’s inequality, Acta Math. Hungarica,
20(1-2) (1998), 129–143.
I. Budimir, S.S. Dragomir and
J. Pečarić
Title Page
[22] T. KOSKI AND L.E. PERSSON, Some properties of generalized entropy
with applications to compression of data, J. of Int. Sciences, 62 (1992),
103–132.
[23] M. MATIĆ, Jensen’s Inequality and Applications in Information Theory,
(in Croatian), Ph.D. Dissertation, Split, Croatia, 1998.
Contents
JJ
J
II
I
Go Back
[24] R.J. McELIECE, The Theory of Information and Coding, Addison Wesley
Publishing Company, Reading, 1977.
Close
[25] J. E. PEČARIĆ, F. PROSCHAN AND Y.L. TONG, Convex Functions, Partial Orderings and Statistical Applications, Academic Press, 1992.
Page 30 of 31
Quit
J. Ineq. Pure and Appl. Math. 2(1) Art. 5, 2001
http://jipam.vu.edu.au
[26] J. PEETRE AND L.E. PERSSON, A general Beckenbach’s inequality with
applications, In: Function Spaces, Differential Operators and Nonlinear
Analysis, Pitman Research Notes in Math., 211 (1989), 125–139.
Further Reverse Results for
Jensen’s Discrete Inequality
and Applications in Information
Theory
I. Budimir, S.S. Dragomir and
J. Pečarić
Title Page
Contents
JJ
J
II
I
Go Back
Close
Quit
Page 31 of 31
J. Ineq. Pure and Appl. Math. 2(1) Art. 5, 2001
http://jipam.vu.edu.au
Download