2,117
14
Essay, 34 pages (9000 words)

New estimates for the jensen gap using s -convexity with applications

1. Introduction

Convex functions and their generalizations play a significant role in scientific observation and calculation of various parameters in modern analysis, especially in the theory of optimization. Moreover, convex functions have some nice properties, such as differentiability, monotonicity, and continuity, which are useful in applications [ 1 5 ]. Interest in mathematical inequalities for convex and generalized convex functions has been growing exponentially, and research in this respect has had a significant impact on modern analysis [ 6 20 ]. Several mathematical inequalities have been established for s -convex functions in particular [ 21 28 ], one of the most important being the Jensen inequality. In this paper, we study the Jensen inequality in a more standard framework for s -convex functions.

Definition 1. 1 ( s -convexity [ 29 ]). For s > 0 and a convex subset B of a real linear space S , a function Γ : B → ℝ is said to be s -convex if the inequality

Γ ( κ 1 ε 1 + κ 2 ε 2 ) κ 1 s Γ ( ε 1 )+ κ 2 s Γ ( ε 2 )( 1. 1 )

holds for all ε 1 , ε 2 B and κ 1 , κ 2 ≥ 0 with κ 1 + κ 2 = 1.

The function Γ is said to be s -concave if the inequality (1. 1) holds in the reverse direction. Obviously, for s = 1 an s -convex function becomes a convex function, which shows that s -convexity of a function is a generalization of ordinary convexity of that function.

Lemma 1. 2 ([ 29 ]). Let B be a convex subset of a real linear space S and let Γ : B → ℝ be a convex function. Then the following two statements hold :

(a) Γ is s-convex for 0 < s ≤ 1 if Γ is non-negative ;

(b) Γ is s-convex for 1 ≤ s < ∞ if Γ is non-positive .

The Green function [ 30 ]

G 1 ( t , x )={ α 1 x , α 1 x t , α 1 t , t x α 2 ( 1. 2 )

defined on [α 1 , α 2 ] × [α 1 , α 2 ] and the integral identity

Γ ( t )= Γ ( α 1 )+( t α 1 ) Γ ( α 2 )+ α 1 α 2 G 1 ( t , x ) Γ ( x ) d x ( 1. 3 )

for the function Γ C 2 [ α 1 , α 2 ] will be used to obtain the main results. Note that G 1 is convex and continuous with respect to both variables.

This paper is organized as follows. In section 2 we give a bound for the Jensen gap in discrete form, which pertains to functions for which the absolute value of the second derivative is s -convex. We also derive a bound for the integral version of the Jensen gap. Then we conduct two numerical experiments that provide evidence for the tightness of the bound in the main result. We deduce a converse of the Hölder inequality from the discrete result and a bound for the Hermite-Hadamard gap from the integral result. Moreover, as a consequence of the integral result we obtain a converse of the Hölder inequality in its corresponding integral version. At the beginning of section 3 we present bounds for the Csiszár and Rényi divergences in the discrete case. Finally, we give estimates for the Shannon entropy, Kullback-Leibler divergence, χ 2 divergence, Bhattacharyya coefficient, Hellinger distance, and triangular discrimination as applications of the bound obtained for the Csiszár divergence. Conclusions are presented in the final section.

2. Main Results

Using the concept of s -convexity, we derive a bound for the Jensen gap in discrete form, which is presented in the following theorem.

Theorem 2. 1. Suppose | Γ|″ is s convex for a function Γ C 2 [ α 1 , α 2 ] and that z i ∈ [α 1 , α 2 ] and κ i ∈ [0, ∞) for i = 1, …, n with i = 1 n κ i = K > 0 . Then the following inequality holds :

| 1 K i = 1 n κ i Γ ( z i ) Γ ( 1 K i = 1 n κ i z i )|| Γ ( α 1 )|( s + 1 )( s + 2 )( α 2 α 1 ) s ( 1 K i = 1 n κ i ( α 2 z i ) s + 2 ( α 2 1 K i = 1 n κ i z i ) s + 2 )+| Γ ( α 2 )|( s + 1 )( s + 2 )( α 2 α 1 ) s ( 1 K i = 1 n κ i ( z i α 1 ) s + 2 ( 1 K i = 1 n κ i z i α 1 ) s + 2 ).( 2. 4 )

Proof: Using (1. 3), we get

1 K i = 1 n κ i Γ ( z i )= 1 K i = 1 n κ i ( Γ ( α 1 )+( z i α 1 ) Γ ( α 2 )+ α 1 α 2 G 1 ( z i , x ) Γ ( x ) d x )( 2. 5 )

and

Γ ( 1 K i = 1 n κ i z i )= Γ ( α 1 )+( 1 K i = 1 n κ i z i α 1 ) Γ ( α 2 )+ α 1 α 2 G 1 ( 1 K i = 1 n κ i z i , x ) Γ ( x ) d x .( 2. 6 )

Equations (2. 5) and (2. 6) give

1 K i = 1 n κ i Γ ( z i ) Γ ( 1 K i = 1 n κ i z i )= α 1 α 2 ( 1 K i = 1 n κ i G 1 ( z i , x ) G 1 ( 1 K i = 1 n κ i z i , x )) Γ ( x ) d x .( 2. 7 )

Taking the absolute value of (2. 7), we get

| 1 K i = 1 n κ i Γ ( z i ) Γ ( 1 K i = 1 n κ i z i )|=| α 1 α 2 ( 1 K i = 1 n κ i G 1 ( z i , x ) G 1 ( 1 K i = 1 n κ i z i , x )) Γ ( x ) d x | α 1 α 2 | 1 K i = 1 n κ i G 1 ( z i , x )( G 1 1 K i = 1 n κ i z i , x )|| Γ ( x )| d x .( 2. 8 )

By applying a change of variable x = 1 + (1 − t 2 for t ∈ [0, 1] and using the convexity of G 1 ( t, x ), the inequality (2. 8) is transformed to

| 1 K i = 1 n κ i Γ ( z i ) Γ ( z ¯)|( α 2 α 1 ) 0 1 ( 1 K i = 1 n κ i G 1 ( z i , t α 1 +( 1 t ) α 2 ) G 1 ( z ¯, t α 1 +( 1 t ) α 2 ))×| Γ ( t α 1 +( 1 t ) α 2 )| d t ,( 2. 9 )

where z = 1 K i = 1 n κ i z i . The inequality (2. 9) leads to the following by using s -convexity of the function | Γ|″:

| 1 K i = 1 n κ i Γ ( z i ) Γ ( z ¯)|( α 2 α 1 ) 0 1 ( 1 K i = 1 n κ i G 1 ( z i , t α 1 +( 1 t ) α 2 ) G 1 ( z ¯, t α 1 +( 1 t ) α 2 ))×( t s | Γ ( α 1 )|+( 1 t ) s | Γ ( α 2 )|) d t .=( α 2 α 1 ) 0 1 ( 1 K i = 1 n κ i G 1 ( z i , t α 1 +( 1 t ) α 2 ) t s | Γ ( α 1 )|+ 1 K i = 1 n κ i G 1 ( z i , t α 1 +( 1 t ) α 2 )( 1 t ) s | Γ ( α 2 )| G 1 ( z ¯, t α 1 +( 1 t ) α 2 ) t s | Γ ( α 1 )| G 1 ( z ¯, t α 1 +( 1 t ) α 2 )( 1 t ) s | Γ ( α 2 )|) d t =( α 2 α 1 )(| Γ ( α 1 )| 1 K i = 1 n κ i 0 1 t s G 1 ( z i , t α 1 +( 1 t ) α 2 ) d t +| Γ ( α 2 )| 1 K i = 1 n κ i 0 1 ( 1 t ) s G 1 ( z i , t α 1 +( 1 t ) α 2 ) d t | Γ ( α 1 )| 0 1 t s G 1 ( z ¯, t α 1 +( 1 t ) α 2 ) d t | Γ ( α 2 )| 0 1 ( 1 t ) s G 1 ( z ¯, t α 1 +( 1 t ) α 2 ) d t ).( 2. 10 )

Now, by using the change of variable x = 1 + (1 − t 2 for t ∈ [0, 1], we obtain

0 1 t s G 1 ( z i , t α 1 +( 1 t ) α 2 ) d t = 1 ( α 2 α 1 ) s + 1 (( α 2 z i ) s + 2 ( s + 1 )( s + 2 )( α 2 α 1 ) s + 2 ( s + 1 )( s + 2 )).( 2. 11 )

Upon replacing z i by z in (2. 11), we get

0 1 t s G 1 ( z ¯, t α 1 +( 1 t ) α 2 ) d t = 1 ( α 2 α 1 ) s + 1 (( α 2 z ¯) s + 2 ( s + 1 )( s + 2 )( α 2 α 1 ) s + 2 ( s + 1 )( s + 2 )).( 2. 12 )

Also,

0 1 ( 1 t ) s G 1 ( z i , t α 1 +( 1 t ) α 2 ) d t = 1 ( α 2 α 1 ) s + 1 (( z i α 1 ) s + 2 ( s + 1 )( s + 2 )( z i α 1 )( α 2 α 1 ) s + 1 ( s + 1 )).( 2. 13 )

Upon replacing z i by z in (2. 13), we get

0 1 ( 1 t ) s G 1 ( z ¯, t α 1 +( 1 t ) α 2 ) d t = 1 ( α 2 α 1 ) s + 1 (( z ¯ α 1 ) s + 2 ( s + 1 )( s + 2 )( z ¯ α 1 )( α 2 α 1 ) s + 1 ( s + 1 )).( 2. 14 )

The result (2. 4) is then obtained by substituting the values from (2. 11)–(2. 14) into (2. 10).

Remark 2. 2. If we use the Green function G 2 , G 3 , or G 4 instead of G 1 in Theorem 2. 1, where G 2 , G 3 , and G 4 are given in [ 30 ], we obtain the same result (2. 4).

In the following theorem, we give a bound for the Jensen gap in integral form.

Theorem 2. 3. Suppose | Γ″| is an s convex function for Γ C 2 [ α 1 , α 2 ], and let ξ 1 and ξ 2 be real-valued functions defined on [ c 1 , c 2 ] with ξ 1 ( y ) ∈ [α 1 , α 2 ] for all y ∈ [ c 1 , c 2 ] and such that ξ 2 , ξ 1 ξ 2 , and (Γ ◦ ξ 1 ) ξ 2 are all integrable functions on [ c 1 , c 2 ]. Then the inequality

| c 1 c 2 ( Γ ξ 1 )( y ) ξ 2 ( y ) d y ξ Γ ( c 1 c 2 ξ 1 ( y ) ξ 2 ( y ) d y ξ )|| Γ ( α 1 )|( s + 1 )( s + 2 )( α 2 α 1 ) s { c 2 c 2 ξ 2 ( y )( α 2 ξ 1 ( y )) s + 2 d y ξ α 2 ( c 1 c 2 ξ 2 ( y ) ξ 1 ( y ) d y ξ ) s + 2 }+| Γ ( α 2 )|( s + 1 )( s + 2 )( α 2 α 1 ) s { c 1 c 2 ξ 2 ( y )( ξ 1 ( y ) α 1 ) s + 2 d y ξ ( c 1 c 2 ξ 2 ( y ) ξ 1 ( y ) d y ξ α 1 ) s + 2 }( 2. 15 )

holds provided that c 1 c 2 ξ 2 ( y ) d y := ξ > 0 when ξ 2 ( y ) ∈ [0, ∞) for all y ∈ [ c 1 , c 2 ].

Proof: Using the same procedure as in the proof of Theorem 2. 1, (2. 15) can be obtained.

Example 1. Let Γ ( y )= 4 15 y 5 2 , ξ 1 ( y )= y 2 , and ξ 2 ( y ) = 1 for all y ∈ [0, 1]. Then Γ ( y )= y 1 2 > 0 for all y ∈ [0, 1]. This shows that Γ is a convex function while | Γ″| is 1 2 -convex. Also , ξ 1 ( y ) ∈ [0, 1] for all y ∈ [0, 1] and we have 1 , α 2 ] = [ c 1 , c 2 ] = [0, 1]. Now, the left-hand side of inequality (2. 15) gives 0 1 Γ ( ξ 1 ( y )) d y Γ ( 0 1 ξ 1 ( y ) d y )= 0 . 0444 0 . 0171 = 0 . 0273 = E 1 , which shows how sharp the Jensen inequality is. The right-hand side of (2. 15) gives 0. 0274, which is very close to the true discrepancy E 1 . That is, from inequality (2. 15) we have

0 . 0273 < 0 . 0274 .( 2. 16 )

The difference 0. 0274 − 0. 0273 = 0. 0001 between the two sides of (2. 16) shows that the bound for the Jensen gap given by inequality (2. 15) is very close to the true value .

Example 2. Let Γ ( y )= 100 231 y 21 10 , ξ 1 ( y ) = y , and ξ 2 ( y ) = 1 for all y ∈ [0, 1]. Then Γ ( y )= y 1 10 > 0 for all y ∈ [0, 1], which shows that Γ is a convex function while | Γ″| is s -convex with s = 1 10 . Also , ξ 1 ( y ) ∈ [0, 1] for all y ∈ [0, 1] and we have 1 , α 2 ] = [ c 1 , c 2 ] = [0, 1]. Therefore, from the left-hand side of inequality (2. 15) we obtain 0 1 Γ ( ξ 1 ( y )) d y Γ ( 0 1 ξ 1 ( y ) d y )= 0 . 1396 0 . 1010 = 0 . 0386 = E 2 , which shows that the Jensen inequality is quite sharp. The right-hand side of (2. 15) gives 0. 0387 , a value very close to the true discrepancy E 2 . Finally, from inequality (2. 15) we have

0 . 0386 < 0 . 0387 .( 2. 17 )

The difference 0. 0387 − 0. 0386 = 0. 0001 between the two sides of (2. 17) provides further evidence of the tightness of the bound for the Jensen gap given by inequality (2. 15).

As an application of Theorem 2. 1, we derive a converse of the Hölder inequality, stated in the following proposition.

Proposition 2. 4. Let q 2 > 1 and q 1 ∉ (2, 3) be such that 1 q 1 + 1 q 2 = 1 , and let s ∈ (0, 1]. Also, let 1 , α 2 ] be a positive interval and let ( d 1 , …, d n ) and ( b 1 , …, b n ) be two positive n tuples such that i = 1 n d i b i i = 1 n b i q 2 , with d i b i q 2 q 1 [ α 1 , α 2 ] for i = 1, …, n . Then

( i = 1 n d i q 1 ) 1 q 1 ( i = 1 n b i q 2 ) 1 q 2 i = 1 n d i b i [ q 1 ( q 1 1 )( s + 1 )( s + 2 )( α 2 α 1 ) s { α 1 q 1 2 ( n = 1 n b i q 2 ( α 2 d i b i q 2 q 1 ) s + 2 n = 1 n b i q 2 ( α 2 n = 1 n d i b i n = 1 n b i q 2 ) s + 2 )+ α 2 q 1 2 ( n = 1 n b i q 2 ( d i b i q 2 q 1 α 1 ) s + 2 n = 1 n b i q 2 ( n = 1 n d i b i n = 1 n b i q 2 α 1 ) s + 2 )}] 1 q 1 i = 1 n b i q 2 .( 2. 18 )

Proof: Let Γ ( x )= x q 1 for x ∈ [α 1 , α 2 ]; then Γ ( x )= q 1 ( q 1 1 ) x q 1 2 > 0 and | Γ |( x )= q 1 ( q 1 1 )( q 1 2 )( q 1 3 ) x q 1 4 > 0 , which shows that Γ and | Γ″| are convex functions. The function | Γ″| is also non-negative, so by Lemma 1. 2 it is also an s -convex function for s ∈ (0, 1]. Thus, using (2. 4) with Γ ( x )= x q 1 , κ i = b i q 2 , and z i = d i b i q 2 q 1 , we derive

(( i = 1 n d i q 1 )( i = 1 n b i q 2 ) q 1 1 ( i = 1 n d i b i ) q 1 ) 1 q 1 [ q 1 ( q 1 1 )( s + 1 )( s + 2 )( α 2 α 1 ) s { α 1 q 1 2 ( i = 1 n b i q 2 ( α 2 d i b i q 2 q 1 ) s + 2 i = 1 n b i q 2 ( α 2 i = 1 n d i b i i = 1 n b i q 2 ) s + 2 )+ α 2 q 1 2 ( i = 1 n b i q 2 ( d i b i q 2 q 1 α 1 ) s + 2 i = 1 n b i q 2 ( i = 1 n d i b i i = 1 n b i q 2 α 1 ) s + 2 )}] 1 q 1 i = 1 n b i q 2 .( 2. 19 )

By using the inequality x γ y γ ≤ ( x y ) γ for 0 ≤ y x and γ ∈ [0, 1] with x =( i = 1 n d i q 1 )×( i = 1 n b i q 2 ) q 1 1 , y =( i = 1 n d i b i ) q 1 , and γ = 1 q 1 , we obtain

( i = 1 n d i q 1 ) 1 q 1 ( i = 1 n b i q 2 ) 1 q 2 i = 1 n d i b i (( i = 1 n d i q 1 )( i = 1 n b i q 2 ) q 1 1 ( i = 1 n d i b i ) q 1 ) 1 q 1 .( 2. 20 )

The inequality (2. 18) follows from (2. 19) and (2. 20).

In the following proposition, we provide a converse of the Hölder inequality in integral form as an application of Theorem 2. 3.

Proposition 2. 5. Let q 2 > 1 and q 1 ∉ (2, 3) be such that 1 q 1 + 1 q 2 = 1 . Also, let ζ 1 , ζ 2 :[ c 1 , c 2 ]+ be two functions such that ζ 1 q 1 ( y ), ζ 2 q 2 ( y ), and ζ 1 ( y 2 ( y ) are integrable on [ c 1 , c 2 ] with ζ 1 ( y ) ζ 2 q 2 / q 1 ( y )[ α 1 , α 2 ] when 1 , α 2 ] ⊂ ℝ. Then the inequality

( c 1 c 2 ζ 1 q 1 ( y ) d y ) 1 q 1 ( c 1 c 2 ζ 2 q 2 ( y ) d y ) 1 q 2 c 1 c 2 ζ 1 ( y ) ζ 2 ( y ) d y [ q 1 ( q 1 1 )( s + 1 )( s + 2 )( α 2 α 1 ) s { α 1 q 1 2 ( 1 c 1 c 2 ζ 2 q 2 ( y ) d y c 1 c 2 ζ 2 q 2 ( y )( α 2 ζ 1 ( y ) ζ 2 q 2 q 1 ( y )) s + 2 d y ( α 2 1 c 1 c 2 ζ 2 q 2 ( y ) d y c 1 c 2 ζ 1 ( y ) ζ 2 ( y ) d y ) s + 2 )+ α 2 q 1 2 ( 1 c 1 c 2 ζ 2 q 2 ( y ) d y c 1 c 2 ζ 2 q 2 ( y )( ζ 1 ( y ) ζ 2 q 2 q 1 ( y ) α 1 ) s + 2 d y ( 1 c 1 c 2 ζ 2 q 2 ( y ) d y c 1 c 2 ζ 1 ( y ) ζ 2 ( y ) d y α 1 ) s + 2 )}] 1 q 1 c 1 c 2 ζ 2 q 2 ( y ) d y ( 2. 21 )

holds for s ∈ (0, 1].

Proof: Using (2. 15) with Γ ( x )= x q 1 for x [ α 1 , α 2 ], ξ 2 ( y )= ζ 2 q 2 ( y ), and ξ 1 ( y )= ζ 1 ( y ) ζ 2 q 2 q 1 ( y ) and following the procedure of Proposition 2. 4, we deduce (2. 21).

As an application of Theorem 2. 3, in the following corollary we establish a bound for the Hermite-Hadamard gap.

Corollary 2. 6. Let ψ C 2 [ c 1 , c 2 ] be a function such that | ψ″| is s-convex; then

| 1 c 2 c 1 c 2 c 1 ψ ( y ) d y ψ ( c 1 + c 2 2 )|( c 2 c 1 ) 2 ( s + 1 )( s + 2 )(| ψ ( c 1 )|+| ψ ( c 2 )|)( 1 s + 3 1 2 s + 2 ).( 2. 22 )

Proof: The inequality (2. 22) can be obtained by using (2. 15) with ψ = Γ, [α 1 , α 2 ] = [ c 1 , c 2 ], ξ 2 ( y ) = 1, and ξ 1 ( y ) = y for y ∈ [ c 1 , c 2 ].

3. Applications to Information Theory

Definition 3. 1 (Csiszár f -divergence [ 31 ]). Let t =( t 1 ,, t n ) n and r =( r 1 ,, r n )+ n with t i r i [ α 1 , α 2 ]( i = 1 ,, n ) for 1 , α 2 ] ⊂ ℝ. For a function f :[α 1 , α 2 ] → ℝ, the Csiszár f -divergence functional is defined as

D c ( t , r )= i = 1 n r i f ( t i r i ).

Theorem 3. 2. Let f C 2 [ α 1 , α 2 ] be a function such that | f ″| is s convex. Then for t =( t 1 ,, t n ) n and r =( r 1 ,, r n )+ n the inequality

| 1 i = 1 n r i D ¯ c ( t , r ) f ( i = 1 n t i i = 1 n r i )|| f ( α 1 )|( s + 1 )( s + 2 )( α 2 α 1 ) s { 1 i = 1 n r i i = 1 n r i ( α 2 t i r i ) s + 2 ( α 2 i = 1 n t i i = 1 n r i ) s + 2 }+| f ( α 2 )|( s + 1 )( s + 2 )( α 2 α 1 ) s { 1 i = 1 n r i i = 1 n r i ( t i r i α 1 ) s + 2 ( i = 1 n t i i = 1 n r i α 1 ) s + 2 }( 3. 23 )

holds provided that i = 1 n t i i = 1 n r i , t i r i [ α 1 , α 2 ] for i = 1, …, n .

Proof: The inequality (3. 23) can easily be deduced from (2. 4) by taking Γ = f , z i = t i r i , and κ i = r i i = 1 n r i .

Definition 3. 3 (Rényi divergence [ 31 ]). For μ ≥ 0 with μ ≠ 1 and two positive probability distributions t= ( t 1 , …, t n ) and r= ( r 1 , …, r n ), the Rényi divergence is defined as

D re ( t , r )= 1 μ 1 log ( i = 1 n t i μ r i 1 μ ).

Corollary 3. 4. Let 0 < s ≤ 1 and [ α 1 , α 2 ]+. Then for positive probability distributions t= ( t 1 , …, t n ) and r= ( r 1 , …, r n ), the inequality

D re ( t , r ) 1 μ 1 i = 1 n t i log ( t i r i ) μ 1 1 ( μ 1 ) α 1 2 ( α 2 α 1 ) s ( s + 1 )( s + 2 )×{ i = 1 n t i ( α 2 ( t i r i ) μ 1 ) s + 2 ( α 2 i = 1 n v i μ w i 1 μ ) s + 2 }+ 1 ( μ 1 ) α 2 2 ( α 2 α 1 ) s ( s + 1 )( s + 2 )×{ i = 1 n t i (( t i r i ) μ 1 α 1 ) s + 2 ( i = 1 n v i μ w i 1 μ α 1 ) s + 2 }.( 3. 24 )

holds provided that i = 1 n r i ( t i r i ) μ ,( t i r i ) μ 1 [ α 1 , α 2 ] for i = 1, …, n with μ > 1.

Proof: Let Γ ( x )= 1 μ 1 log x for x ∈ [α 1 , α 2 ]. Then Γ ( x )= 1 ( μ 1 ) x 2 > 0 and | Γ |( x )= 6 ( μ 1 ) x 4 > 0 , which shows that Γ and | Γ″| are convex functions with | Γ″| ≥ 0; so by Lemma 1. 2 the function | Γ″| is s -convex for s ∈ (0, 1]. Therefore, using (2. 4) with Γ ( x )= 1 μ 1 log x , κ i = t i , and z i =( t i r i ) μ 1 , we derive (3. 24).

Definition 3. 5 (Shannon entropy [ 31 ]). Let r= ( r 1 , …, r n ) be a positive probability distribution; then the Shannon entropy is defined as

E s ( r )= i = 1 n r i log r i .

Corollary 3. 6. Let [ α 1 , α 2 ]+, and let r= ( r 1 , …, r n ) be a positive probability distribution such that 1 r i [ α 1 , α 2 ] for i = 1, …, n with 0 < s ≤ 1. Then

log n E s ( r ) 1 α 1 2 ( α 2 α 1 ) s ( s + 1 )( s + 2 ){ i = 1 n r i ( α 2 1 r i ) s + 2 ( α 2 n ) s + 2 }+ 1 α 2 2 ( α 2 α 1 ) s ( s + 1 )( s + 2 ){ i = 1 n r i ( 1 r i α 1 ) s + 2 ( n α 1 ) s + 2 }.( 3. 25 )

Proof: Let f ( x ) = −log x for x ∈ [α 1 , α 2 ]. Then f ( x )= 1 x 2 > 0 and | f |( x )= 6 x 4 > 0 , which shows that f and | f ″| are convex functions. Also, | f ″| is non-negative and so by Lemma 1. 2 we conclude that it is s -convex for s ∈ (0, 1]. Therefore, using (3. 23) with f ( x ) = −log x and ( t 1 , …, t n ) = (1, …, 1), we get (3. 25).

Definition 3. 7 (Kullback-Leibler divergence [ 31 ]). For two positive probability distributions t= ( t 1 , …, t n ) and r= ( r 1 , …, r n ), the Kullback-Leibler divergence is defined as

D kl ( t , r )= i = 1 n t i log t i r i .

Corollary 3. 8. Let 0 < s ≤ 1 and 0 < α 1 < α 2 , and let t= ( t 1 , …, t n ) and r= ( r 1 , …, r n ) be positive probability distributions such that t i r i [ α 1 , α 2 ] for i = 1, …, n . Then

D kl ( t , r ) 1 α 1 ( α 2 α 1 ) s ( s + 1 )( s + 2 ){ i = 1 n r i ( α 2 t i r i ) s + 2 ( α 2 1 ) s + 2 }+ 1 α 2 ( α 2 α 1 ) s ( s + 1 )( s + 2 ){ i = 1 n r i ( t i r i α 1 ) s + 2 ( 1 α 1 ) s + 2 }.( 3. 26 )

Proof: Let f ( x ) = x log x for x ∈ [α 1 , α 2 ]. Then f ( x )= 1 x > 0 and | f |( x )= 2 x 3 > 0 , which shows that f and | f ″| are convex functions. Also, | f ″| ≥ 0, and so Lemma 1. 2 guarantees the s -convexity of | f ″| for s ∈ (0, 1]. Therefore, using (3. 23) with f ( x ) = x log x , we get (3. 26).

Definition 3. 9 (χ 2 divergence [ 31 ]). The χ 2 divergence D χ 2 ( t , r ) for two positive probability distributions t= ( t 1 , …, t n ) and r= ( r 1 , …, r n ) is defined as

D χ 2 ( t , r )= i = 1 n ( t i r i ) 2 r i .

Corollary 3. 10. Let 0 < s ≤ 1 and 0 < α 1 < α 2 , and lett= ( t 1 , …, t n ) andr= ( r 1 , …, r n ) be positive probability distributions such that t i r i [ α 1 , α 2 ] for i = 1, …, n . Then

D χ 2 ( t , r ) 2 ( α 2 α 1 ) s ( s + 1 )( s + 2 ){ i = 1 n r i ( α 2 t i r i ) s + 2 ( α 2 1 ) s + 2 }+ 2 ( α 2 α 1 ) s ( s + 1 )( s + 2 ){ i = 1 n r i ( t i r i α 1 ) s + 2 ( 1 α 1 ) s + 2 }.( 3. 27 )

Proof: Let f ( x ) = ( x − 1) 2 for x ∈ [α 1 , α 2 ]. Then f ″( x ) = 2 > 0 and | f ″|″( x ) = 0, which shows that f and | f ″| are convex functions. Also, the function | f ″| is non-negative, and so Lemma 1. 2 confirms its s -convexity for s ∈ (0, 1]. Therefore, using (3. 23) with f ( x ) = ( x − 1) 2 , we obtain (3. 27).

Definition 3. 11 (Bhattacharyya coefficient [ 31 ]). For two positive probability distributions t= ( t 1 , …, t n ) and r= ( r 1 , …, r n ), the Bhattacharyya coefficient is defined as

C b ( t , r )= i = 1 n t i r i .

Corollary 3. 12. Let 0 < s ≤ 1 and [ α 1 , α 2 ]+, and let t= ( t 1 , …, t n ) and r= ( r 1 , …, r n ) be two positive probability distributions such that t i r i [ α 1 , α 2 ] for i = 1, …, n . Then

1 C b ( t , r ) 1 4 α 1 3 2 ( α 2 α 1 ) s ( s + 1 )( s + 2 ){ i = 1 n r i ( α 2 t i r i ) s + 2 ( α 2 1 ) s + 2 }+ 1 4 α 2 3 2 ( α 2 α 1 ) s ( s + 1 )( s + 2 ){ i = 1 n r i ( t i r i α 1 ) s + 2 ( 1 α 1 ) s + 2 }.( 3. 28 )

Proof: Let f ( x )= x for x ∈ [α 1 , α 2 ]. Then f ( x )= 1 4 x 3 2 > 0 and | f |( x )= 15 16 x 7 2 > 0 , which shows that f and | f ″| are convex functions. Also, | f ″| ≥ 0 implies its s -convexity for s ∈ (0, 1] by Lemma 1. 2. Therefore, using (3. 23) with f ( x )= x , we obtain (3. 28).

Definition 3. 13 (Hellinger distance [ 31 ]). The Hellinger distance D h 2 ( t , r ) between two positive probability distributions t= ( t 1 , …, t n ) and r= ( r 1 , …, r n ) is defined as

D h 2 ( t , r )= 1 2 i = 1 n ( t i r i ) 2 .

Corollary 3. 14. Let 0 < α 1 < α 2 and 0 < s ≤ 1, and let t= ( t 1 , …, t n ) and r= ( r 1 , …, r n ) be positive probability distributions such that t i r i [ α 1 , α 2 ] for i = 1, …, n . Then

D h 2 ( t , r ) 1 4 α 1 3 2 ( α 2 α 1 ) s ( s + 1 )( s + 2 ){ i = 1 n r i ( α 2 t i r i ) s + 2 ( α 2 1 ) s + 2 }+ 1 4 α 2 3 2 ( α 2 α 1 ) s ( s + 1 )( s + 2 ){ i = 1 n r i ( t i r i α 1 ) s + 2 ( 1 α 1 ) s + 2 }.( 3. 29 )

Proof: Let f ( x )= 1 2 ( 1 x ) 2 for x ∈ [α 1 , α 2 ]. Then f ( x )= 1 4 x 3 2 > 0 and | f |( x )= 15 16 x 7 2 > 0 , which shows that f and | f ″| are convex functions. Also, | f ″| ≥ 0, and so from Lemma 1. 2 we conclude its s -convexity for s ∈ (0, 1]. Therefore, using (3. 23) with f ( x )= 1 2 ( 1 x ) 2 , we deduce (3. 29).

Definition 3. 15 (Triangular discrimination [ 31 ]). For two positive probability distributions t= ( t 1 , …, t n ) and r= ( r 1 , …, r n ), the triangular discrimination is defined as

D ( t , r )= i = 1 n ( t i r i ) 2 t i + r i .

Corollary 3. 16. Let 0 < s ≤ 1 and 0 < α 1 < α 2 , and let t= ( t 1 , …, t n ) and r= ( r 1 , …, r n ) be positive probability distributions such that t i r i [ α 1 , α 2 ] for i = 1, …, n . Then

D ( t , r ) 8 ( α 1 + 1 ) 3 ( α 2 α 1 ) s ( s + 1 )( s + 2 ){ i = 1 n r i ( α 2 t i r i ) s + 2 ( α 2 1 ) s + 2 }+ 8 ( α 2 + 1 ) 3 ( α 2 α 1 ) s ( s + 1 )( s + 2 ){ i = 1 n r i ( t i r i α 1 ) s + 2 ( 1 α 1 ) s + 2 }.( 3. 30 )

Proof: Let f ( x )=( x 1 ) 2 ( x + 1 ) for x ∈ [α 1 , α 2 ]. Then f ( x )= 8 ( x + 1 ) 3 > 0 and | f |( x )= 96 ( x + 1 ) 5 > 0 , which shows that f and | f ″| are convex functions. Also, | f ″| is non-negative, and thus s -convexity of the function | f ″| for s ∈ (0, 1] follows from Lemma 1. 2. Therefore, using (3. 23) with f ( x )=( x 1 ) 2 ( x + 1 ), we get (3. 30).

Remark 3. 17. Analogously, bounds for various divergences in integral form can be derived as applications of Theorem 2. 3 .

4. Conclusion

The Jensen inequality has numerous applications in engineering, economics, computer science, information theory, and coding; it has been derived for convex and generalized convex functions. This paper presents a novel approach to bounding the Jensen gap. Some bounds are obtained for the Jensen gap via s -convex functions. Numerical experiments not only confirm the sharpness of the Jensen inequality but also provide evidence for the tightness of the bound given in (2. 15) for the Jensen gap. These experiments also show that the bound in (2. 15) gives very close estimates for the Jensen gap even when the functions are not convex. The bounds are used to obtain new estimates for the Hermite-Hadamard and Hölder inequalities. Furthermore, based on the main results, various divergences are estimated. These estimates for divergences can be applied to signal processing, magnetic resonance image analysis, image segmentation, pattern recognition, and other areas. The ideas in this paper can also be used with other inequalities and for some other classes of convex functions.

Data Availability Statement

The original contributions presented in the study are included in the article/supplementary materials, further inquiries can be directed to the corresponding author/s.

Author Contributions

MA gave the main idea. MA and SK worked on Main Results while Y-MC worked on Introduction. All authors checked carefully the whole manuscript and approved.

Conflict of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Acknowledgments

This work was supported by the Natural Science Foundation of China (Grant Nos. 61673169, 11301127, 11701176, 11626101, and 11601485).

References

1. Asplund E. Fréchet differentiability of convex functions. Acta Math.(1968)121: 31–47. doi: 10. 1007/BF02391908

|

2. Phelps RR. Convex Functions, Monotone Operators and Differentiability, Vol. 1364 . Lecture Notes in Mathematics. Berlin: Springer-Verlag. (1989).

3. Udrişte C. Continuity of convex functions on Riemannian manifolds. Bull Math Soc Sci . (1977)21: 215–8.

4. Ger R, Kuczma M. On the boundedness and continuity of convex functions and additive functions. Aequ Math.(1970)4: 157–62. doi: 10. 1007/BF01817756

|

5. Minty GJ. On the monotonicity of the gradient of a convex function. Pac J Math . (1964)14: 243–7.

6. Khan S, Adil Khan M, Chu Y-M. Converses of the Jensen inequality derived from the Green functions with applications in information theory. Math Method Appl Sci.(2020)43: 2577–87. doi: 10. 1002/mma. 6066

|

7. Khan S, Adil Khan M, Chu Y-M. New converses of Jensen inequality via Green functions with applications. RACSAM . (2020)114: 1–14. doi: 10. 1007/s13398-020-00843-1

|

8. Adil Khan M, Pečarić Ð, Pečarić J. New refinement of the Jensen inequality associated to certain functions with applications. J Inequal Appl.(2020)2020: 1–11. doi: 10. 1186/s13660-020-02343-7

|

9. Bakula MK, Özdemir ME, Pečarić J. Hadamard type inequalities for m-convex and (α, m)-convex functions. J Inequal Pure Appl Math.(2008)9: 1–12.

10. Bombardelli M, Varošanec S. Properties of h-convex functions related to the Hermite-Hadamard-Fejér inequalities. Comput Math Appl.(2009)58: 1869–77. doi: 10. 1016/j. camwa. 2009. 07. 073

|

11. Khan J, Adil Khan M, Pečarić J. On Jensen’s type inequalities via generalized majorization inequalities. Filomat . (2018)32: 5719–33. doi: 10. 2298/FIL1816719K

|

12. Dragomir SS, Pearce CEM. Jensen’s inequality for quasi-convex functions. Numer Algebra Control Opt . (2012)2: 279–91. doi: 10. 3934/naco. 2012. 2. 279

|

13. Wang M-K, Zhang W, Chu Y-M. Monotonicity, convexity and inequalities involving the generalized elliptic integrals. Acta Math Sci.(2019)39B: 1440–50. doi: 10. 1007/s10473-019-0520-z

||

14. Wu S-H, Chu Y-M. Schur m -power convexity of generalized geometric Bonferroni mean involving three parameters. J Inequal Appl . (2019)2019: 1–11. doi: 10. 1186/s13660-019-2013-y

|

15. Jain S, Mehrez K, Baleanu D, Agarwal P. Certain Hermite-Hadamard inequalities for logarithmically convex functions with applications. Mathematics.(2019)7: 1–12. doi: 10. 3390/math7020163

|

16. Agarwal P, Jleli M, Tomar M. Certain Hermite-Hadamard type inequalities via generalized k-fractional integrals. J Inequal Appl.(2017)2017: 1–10. doi: 10. 1186/s13660-017-1318-y

||

17. Agarwal P. Some inequalities involving Hadamard-type k-fractional integral operators. Math Method Appl Sci.(2017)40: 3882–91. doi: 10. 1002/mma. 4270

|

18. Liu Z, Yang W, Agarwal P. Certain Chebyshev type inequalities involving the generalized fractional integral operator. J Comput Anal Appl.(2017)22: 999–1014.

19. Choi J, Agarwal P. Certain inequalities involving pathway fractional integral operators. Kyungpook Math J.(2016)56: 1161–8. doi: 10. 5666/KMJ. 2016. 56. 4. 1161

|

20. Mehrez K, Agarwal P. New Hermite-Hadamard type integral inequalities for convex functions and their applications. J Comput Appl Math.(2019)350: 274–85. doi: 10. 2140/pjm. 1964. 14. 243

|

21. Chen X. New convex functions in linear spaces and Jensen’s discrete inequality. J Inequal Appl.(2013)2013: 1–8. doi: 10. 1186/1029-242X-2013-472

|

22. Set E. New inequalities of Ostrowski type for mappings whose derivatives are s-convex in the second sense via fractional integrals. Comput Math Appl.(2012)63: 1147–54. doi: 10. 1016/j. camwa. 2011. 12. 023

|

23. Sarikaya MZ, Set E, Özdemir ME. On new inequalities of Simpson’s type for s-convex functions. Comput Math Appl.(2010)60: 2191–9. doi: 10. 1016/j. camwa. 2010. 07. 033

||

24. Alomari M, Darus M, Dragomir SS, Cerone P. Ostrowski type inequalities for functions whose derivatives are s-convex in the second sense. Appl Math Lett.(2010)23: 1071–6. doi: 10. 1016/j. aml. 2010. 04. 038

|

25. Chen J, Huang X. Some new inequalities of Simpson’s type for s -convex functions via fractional integrals. Filomat . (2017)31: 4989–97. doi: 10. 2298/FIL1715989C

||

26. Özcan S, Işcan I. Some new Hermite-Hadamard type inequalities for s -convex functions and their applications. J Inequal Appl.(2019)2019: 1–11. doi: 10. 1186/s13660-019-2151-2

|

27. Almutairi O, kiliçman A. Integral inequalities for s-convexity via generalized fractional integrals on fractal sets. Mathematics . (2020)8. doi: 10. 3390/math8010053

|

28. Özdemir ME, Yildiz Ç, Akdemir AO, Set E. On some inequalities for s-convex functions and applications. J Inequal Appl.(2013)2013: 1–11. doi: 10. 1186/1029-242X-2013-333

|

29. Adil Khan M, Hanif M, Khan ZAH, Ahmad K, Chu Y-M. Association of Jensen’s inequality for s -convex function with Csiszár divergence. J Inequal Appl . (2019)2019: 1–14. doi: 10. 1186/s13660-019-2112-9

|

30. Butt SI, Mehmood N, Pečarić J. New generalizations of Popoviciu type inequalities via new green functions and Fink’s identity. Trans A Razmadze Math Inst.(2017)171: 293–303. doi: 10. 1016/j. trmi. 2017. 04. 003

||

31. Lovričević N, Pečarić Ð, Pečarić J. Zipf-Mandelbrot law, f −divergences and the Jensen-type interpolating inequalities. J Inequal Appl.(2018)2018: 1–20. doi: 10. 1186/s13660-018-1625-y

||

Thank's for Your Vote!
New estimates for the jensen gap using s -convexity with applications. Page 1
New estimates for the jensen gap using s -convexity with applications. Page 2
New estimates for the jensen gap using s -convexity with applications. Page 3
New estimates for the jensen gap using s -convexity with applications. Page 4
New estimates for the jensen gap using s -convexity with applications. Page 5
New estimates for the jensen gap using s -convexity with applications. Page 6
New estimates for the jensen gap using s -convexity with applications. Page 7
New estimates for the jensen gap using s -convexity with applications. Page 8
New estimates for the jensen gap using s -convexity with applications. Page 9

This work, titled "New estimates for the jensen gap using s -convexity with applications" was written and willingly shared by a fellow student. This sample can be utilized as a research and reference resource to aid in the writing of your own work. Any use of the work that does not include an appropriate citation is banned.

If you are the owner of this work and don’t want it to be published on AssignBuster, request its removal.

Request Removal
Cite this Essay

References

AssignBuster. (2022) 'New estimates for the jensen gap using s -convexity with applications'. 3 November.

Reference

AssignBuster. (2022, November 3). New estimates for the jensen gap using s -convexity with applications. Retrieved from https://assignbuster.com/new-estimates-for-the-jensen-gap-using-s-convexity-with-applications/

References

AssignBuster. 2022. "New estimates for the jensen gap using s -convexity with applications." November 3, 2022. https://assignbuster.com/new-estimates-for-the-jensen-gap-using-s-convexity-with-applications/.

1. AssignBuster. "New estimates for the jensen gap using s -convexity with applications." November 3, 2022. https://assignbuster.com/new-estimates-for-the-jensen-gap-using-s-convexity-with-applications/.


Bibliography


AssignBuster. "New estimates for the jensen gap using s -convexity with applications." November 3, 2022. https://assignbuster.com/new-estimates-for-the-jensen-gap-using-s-convexity-with-applications/.

Work Cited

"New estimates for the jensen gap using s -convexity with applications." AssignBuster, 3 Nov. 2022, assignbuster.com/new-estimates-for-the-jensen-gap-using-s-convexity-with-applications/.

Get in Touch

Please, let us know if you have any ideas on improving New estimates for the jensen gap using s -convexity with applications, or our service. We will be happy to hear what you think: [email protected]