- Published: January 2, 2022
- Updated: November 3, 2022
- University / College: Florida International University
- Language: English
- Downloads: 14
1. Introduction
Convex functions and their generalizations play a significant role in scientific observation and calculation of various parameters in modern analysis, especially in the theory of optimization. Moreover, convex functions have some nice properties, such as differentiability, monotonicity, and continuity, which are useful in applications [ 1 – 5 ]. Interest in mathematical inequalities for convex and generalized convex functions has been growing exponentially, and research in this respect has had a significant impact on modern analysis [ 6 – 20 ]. Several mathematical inequalities have been established for s -convex functions in particular [ 21 – 28 ], one of the most important being the Jensen inequality. In this paper, we study the Jensen inequality in a more standard framework for s -convex functions.
Definition 1. 1 ( s -convexity [ 29 ]). For s > 0 and a convex subset B of a real linear space S , a function Γ : B → ℝ is said to be s -convex if the inequality
holds for all ε 1 , ε 2 ∈ B and κ 1 , κ 2 ≥ 0 with κ 1 + κ 2 = 1.
The function Γ is said to be s -concave if the inequality (1. 1) holds in the reverse direction. Obviously, for s = 1 an s -convex function becomes a convex function, which shows that s -convexity of a function is a generalization of ordinary convexity of that function.
Lemma 1. 2 ([ 29 ]). Let B be a convex subset of a real linear space S and let Γ : B → ℝ be a convex function. Then the following two statements hold :
(a) Γ is s-convex for 0 < s ≤ 1 if Γ is non-negative ;
(b) Γ is s-convex for 1 ≤ s < ∞ if Γ is non-positive .
The Green function [ 30 ]
defined on [α 1 , α 2 ] × [α 1 , α 2 ] and the integral identity
for the function will be used to obtain the main results. Note that G 1 is convex and continuous with respect to both variables.
This paper is organized as follows. In section 2 we give a bound for the Jensen gap in discrete form, which pertains to functions for which the absolute value of the second derivative is s -convex. We also derive a bound for the integral version of the Jensen gap. Then we conduct two numerical experiments that provide evidence for the tightness of the bound in the main result. We deduce a converse of the Hölder inequality from the discrete result and a bound for the Hermite-Hadamard gap from the integral result. Moreover, as a consequence of the integral result we obtain a converse of the Hölder inequality in its corresponding integral version. At the beginning of section 3 we present bounds for the Csiszár and Rényi divergences in the discrete case. Finally, we give estimates for the Shannon entropy, Kullback-Leibler divergence, χ 2 divergence, Bhattacharyya coefficient, Hellinger distance, and triangular discrimination as applications of the bound obtained for the Csiszár divergence. Conclusions are presented in the final section.
2. Main Results
Using the concept of s -convexity, we derive a bound for the Jensen gap in discrete form, which is presented in the following theorem.
Theorem 2. 1. Suppose | Γ|″ is s – convex for a function and that z i ∈ [α 1 , α 2 ] and κ i ∈ [0, ∞) for i = 1, …, n with . Then the following inequality holds :
Proof: Using (1. 3), we get
and
Equations (2. 5) and (2. 6) give
Taking the absolute value of (2. 7), we get
By applying a change of variable x = tα 1 + (1 − t )α 2 for t ∈ [0, 1] and using the convexity of G 1 ( t, x ), the inequality (2. 8) is transformed to
where The inequality (2. 9) leads to the following by using s -convexity of the function | Γ|″:
Now, by using the change of variable x = tα 1 + (1 − t )α 2 for t ∈ [0, 1], we obtain
Upon replacing z i by in (2. 11), we get
Also,
Upon replacing z i by in (2. 13), we get
The result (2. 4) is then obtained by substituting the values from (2. 11)–(2. 14) into (2. 10).
Remark 2. 2. If we use the Green function G 2 , G 3 , or G 4 instead of G 1 in Theorem 2. 1, where G 2 , G 3 , and G 4 are given in [ 30 ], we obtain the same result (2. 4).
In the following theorem, we give a bound for the Jensen gap in integral form.
Theorem 2. 3. Suppose | Γ″| is an s – convex function for , and let ξ 1 and ξ 2 be real-valued functions defined on [ c 1 , c 2 ] with ξ 1 ( y ) ∈ [α 1 , α 2 ] for all y ∈ [ c 1 , c 2 ] and such that ξ 2 , ξ 1 ξ 2 , and (Γ ◦ ξ 1 ) ξ 2 are all integrable functions on [ c 1 , c 2 ]. Then the inequality
holds provided that when ξ 2 ( y ) ∈ [0, ∞) for all y ∈ [ c 1 , c 2 ].
Proof: Using the same procedure as in the proof of Theorem 2. 1, (2. 15) can be obtained.
Example 1. Let ,, and ξ 2 ( y ) = 1 for all y ∈ [0, 1]. Then for all y ∈ [0, 1]. This shows that Γ is a convex function while | Γ″| is -convex. Also , ξ 1 ( y ) ∈ [0, 1] for all y ∈ [0, 1] and we have [α 1 , α 2 ] = [ c 1 , c 2 ] = [0, 1]. Now, the left-hand side of inequality (2. 15) gives , which shows how sharp the Jensen inequality is. The right-hand side of (2. 15) gives 0. 0274, which is very close to the true discrepancy E 1 . That is, from inequality (2. 15) we have
The difference 0. 0274 − 0. 0273 = 0. 0001 between the two sides of (2. 16) shows that the bound for the Jensen gap given by inequality (2. 15) is very close to the true value .
Example 2. Let , ξ 1 ( y ) = y , and ξ 2 ( y ) = 1 for all y ∈ [0, 1]. Then for all y ∈ [0, 1], which shows that Γ is a convex function while | Γ″| is s -convex with Also , ξ 1 ( y ) ∈ [0, 1] for all y ∈ [0, 1] and we have [α 1 , α 2 ] = [ c 1 , c 2 ] = [0, 1]. Therefore, from the left-hand side of inequality (2. 15) we obtain which shows that the Jensen inequality is quite sharp. The right-hand side of (2. 15) gives 0. 0387 , a value very close to the true discrepancy E 2 . Finally, from inequality (2. 15) we have
The difference 0. 0387 − 0. 0386 = 0. 0001 between the two sides of (2. 17) provides further evidence of the tightness of the bound for the Jensen gap given by inequality (2. 15).
As an application of Theorem 2. 1, we derive a converse of the Hölder inequality, stated in the following proposition.
Proposition 2. 4. Let q 2 > 1 and q 1 ∉ (2, 3) be such that , and let s ∈ (0, 1]. Also, let [α 1 , α 2 ] be a positive interval and let ( d 1 , …, d n ) and ( b 1 , …, b n ) be two positive n – tuples such that , with for i = 1, …, n . Then
Proof: Let for x ∈ [α 1 , α 2 ]; then and which shows that Γ and | Γ″| are convex functions. The function | Γ″| is also non-negative, so by Lemma 1. 2 it is also an s -convex function for s ∈ (0, 1]. Thus, using (2. 4) with , and we derive
By using the inequality x γ − y γ ≤ ( x − y ) γ for 0 ≤ y ≤ x and γ ∈ [0, 1] with ,, and we obtain
The inequality (2. 18) follows from (2. 19) and (2. 20).
In the following proposition, we provide a converse of the Hölder inequality in integral form as an application of Theorem 2. 3.
Proposition 2. 5. Let q 2 > 1 and q 1 ∉ (2, 3) be such that Also, let be two functions such that ,, and ζ 1 ( y )ζ 2 ( y ) are integrable on [ c 1 , c 2 ] with when [α 1 , α 2 ] ⊂ ℝ. Then the inequality
holds for s ∈ (0, 1].
Proof: Using (2. 15) with for , and and following the procedure of Proposition 2. 4, we deduce (2. 21).
As an application of Theorem 2. 3, in the following corollary we establish a bound for the Hermite-Hadamard gap.
Corollary 2. 6. Let be a function such that | ψ″| is s-convex; then
Proof: The inequality (2. 22) can be obtained by using (2. 15) with ψ = Γ, [α 1 , α 2 ] = [ c 1 , c 2 ], ξ 2 ( y ) = 1, and ξ 1 ( y ) = y for y ∈ [ c 1 , c 2 ].
3. Applications to Information Theory
Definition 3. 1 (Csiszár f -divergence [ 31 ]). Let and with for [α 1 , α 2 ] ⊂ ℝ. For a function f :[α 1 , α 2 ] → ℝ, the Csiszár f -divergence functional is defined as
Theorem 3. 2. Let be a function such that | f ″| is s – convex. Then for and the inequality
holds provided that for i = 1, …, n .
Proof: The inequality (3. 23) can easily be deduced from (2. 4) by taking , and
Definition 3. 3 (Rényi divergence [ 31 ]). For μ ≥ 0 with μ ≠ 1 and two positive probability distributions t= ( t 1 , …, t n ) and r= ( r 1 , …, r n ), the Rényi divergence is defined as
Corollary 3. 4. Let 0 < s ≤ 1 and . Then for positive probability distributions t= ( t 1 , …, t n ) and r= ( r 1 , …, r n ), the inequality
holds provided that for i = 1, …, n with μ > 1.
Proof: Let for x ∈ [α 1 , α 2 ]. Then and which shows that Γ and | Γ″| are convex functions with | Γ″| ≥ 0; so by Lemma 1. 2 the function | Γ″| is s -convex for s ∈ (0, 1]. Therefore, using (2. 4) with , and , we derive (3. 24).
Definition 3. 5 (Shannon entropy [ 31 ]). Let r= ( r 1 , …, r n ) be a positive probability distribution; then the Shannon entropy is defined as
Corollary 3. 6. Let , and let r= ( r 1 , …, r n ) be a positive probability distribution such that for i = 1, …, n with 0 < s ≤ 1. Then
Proof: Let f ( x ) = −log x for x ∈ [α 1 , α 2 ]. Then and , which shows that f and | f ″| are convex functions. Also, | f ″| is non-negative and so by Lemma 1. 2 we conclude that it is s -convex for s ∈ (0, 1]. Therefore, using (3. 23) with f ( x ) = −log x and ( t 1 , …, t n ) = (1, …, 1), we get (3. 25).
Definition 3. 7 (Kullback-Leibler divergence [ 31 ]). For two positive probability distributions t= ( t 1 , …, t n ) and r= ( r 1 , …, r n ), the Kullback-Leibler divergence is defined as
Corollary 3. 8. Let 0 < s ≤ 1 and 0 < α 1 < α 2 , and let t= ( t 1 , …, t n ) and r= ( r 1 , …, r n ) be positive probability distributions such that for i = 1, …, n . Then
Proof: Let f ( x ) = x log x for x ∈ [α 1 , α 2 ]. Then and which shows that f and | f ″| are convex functions. Also, | f ″| ≥ 0, and so Lemma 1. 2 guarantees the s -convexity of | f ″| for s ∈ (0, 1]. Therefore, using (3. 23) with f ( x ) = x log x , we get (3. 26).
Definition 3. 9 (χ 2 divergence [ 31 ]). The χ 2 divergence for two positive probability distributions t= ( t 1 , …, t n ) and r= ( r 1 , …, r n ) is defined as
Corollary 3. 10. Let 0 < s ≤ 1 and 0 < α 1 < α 2 , and lett= ( t 1 , …, t n ) andr= ( r 1 , …, r n ) be positive probability distributions such that for i = 1, …, n . Then
Proof: Let f ( x ) = ( x − 1) 2 for x ∈ [α 1 , α 2 ]. Then f ″( x ) = 2 > 0 and | f ″|″( x ) = 0, which shows that f and | f ″| are convex functions. Also, the function | f ″| is non-negative, and so Lemma 1. 2 confirms its s -convexity for s ∈ (0, 1]. Therefore, using (3. 23) with f ( x ) = ( x − 1) 2 , we obtain (3. 27).
Definition 3. 11 (Bhattacharyya coefficient [ 31 ]). For two positive probability distributions t= ( t 1 , …, t n ) and r= ( r 1 , …, r n ), the Bhattacharyya coefficient is defined as
Corollary 3. 12. Let 0 < s ≤ 1 and , and let t= ( t 1 , …, t n ) and r= ( r 1 , …, r n ) be two positive probability distributions such that for i = 1, …, n . Then
Proof: Let for x ∈ [α 1 , α 2 ]. Then and which shows that f and | f ″| are convex functions. Also, | f ″| ≥ 0 implies its s -convexity for s ∈ (0, 1] by Lemma 1. 2. Therefore, using (3. 23) with we obtain (3. 28).
Definition 3. 13 (Hellinger distance [ 31 ]). The Hellinger distance between two positive probability distributions t= ( t 1 , …, t n ) and r= ( r 1 , …, r n ) is defined as
Corollary 3. 14. Let 0 < α 1 < α 2 and 0 < s ≤ 1, and let t= ( t 1 , …, t n ) and r= ( r 1 , …, r n ) be positive probability distributions such that for i = 1, …, n . Then
Proof: Let for x ∈ [α 1 , α 2 ]. Then and which shows that f and | f ″| are convex functions. Also, | f ″| ≥ 0, and so from Lemma 1. 2 we conclude its s -convexity for s ∈ (0, 1]. Therefore, using (3. 23) with we deduce (3. 29).
Definition 3. 15 (Triangular discrimination [ 31 ]). For two positive probability distributions t= ( t 1 , …, t n ) and r= ( r 1 , …, r n ), the triangular discrimination is defined as
Corollary 3. 16. Let 0 < s ≤ 1 and 0 < α 1 < α 2 , and let t= ( t 1 , …, t n ) and r= ( r 1 , …, r n ) be positive probability distributions such that for i = 1, …, n . Then
Proof: Let for x ∈ [α 1 , α 2 ]. Then and which shows that f and | f ″| are convex functions. Also, | f ″| is non-negative, and thus s -convexity of the function | f ″| for s ∈ (0, 1] follows from Lemma 1. 2. Therefore, using (3. 23) with we get (3. 30).
Remark 3. 17. Analogously, bounds for various divergences in integral form can be derived as applications of Theorem 2. 3 .
4. Conclusion
The Jensen inequality has numerous applications in engineering, economics, computer science, information theory, and coding; it has been derived for convex and generalized convex functions. This paper presents a novel approach to bounding the Jensen gap. Some bounds are obtained for the Jensen gap via s -convex functions. Numerical experiments not only confirm the sharpness of the Jensen inequality but also provide evidence for the tightness of the bound given in (2. 15) for the Jensen gap. These experiments also show that the bound in (2. 15) gives very close estimates for the Jensen gap even when the functions are not convex. The bounds are used to obtain new estimates for the Hermite-Hadamard and Hölder inequalities. Furthermore, based on the main results, various divergences are estimated. These estimates for divergences can be applied to signal processing, magnetic resonance image analysis, image segmentation, pattern recognition, and other areas. The ideas in this paper can also be used with other inequalities and for some other classes of convex functions.
Data Availability Statement
The original contributions presented in the study are included in the article/supplementary materials, further inquiries can be directed to the corresponding author/s.
Author Contributions
MA gave the main idea. MA and SK worked on Main Results while Y-MC worked on Introduction. All authors checked carefully the whole manuscript and approved.
Conflict of Interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Acknowledgments
This work was supported by the Natural Science Foundation of China (Grant Nos. 61673169, 11301127, 11701176, 11626101, and 11601485).
References
1. Asplund E. Fréchet differentiability of convex functions. Acta Math.(1968)121: 31–47. doi: 10. 1007/BF02391908
CrossRef Full Text | Google Scholar
2. Phelps RR. Convex Functions, Monotone Operators and Differentiability, Vol. 1364 . Lecture Notes in Mathematics. Berlin: Springer-Verlag. (1989).
3. Udrişte C. Continuity of convex functions on Riemannian manifolds. Bull Math Soc Sci . (1977)21: 215–8.
4. Ger R, Kuczma M. On the boundedness and continuity of convex functions and additive functions. Aequ Math.(1970)4: 157–62. doi: 10. 1007/BF01817756
CrossRef Full Text | Google Scholar
5. Minty GJ. On the monotonicity of the gradient of a convex function. Pac J Math . (1964)14: 243–7.
6. Khan S, Adil Khan M, Chu Y-M. Converses of the Jensen inequality derived from the Green functions with applications in information theory. Math Method Appl Sci.(2020)43: 2577–87. doi: 10. 1002/mma. 6066
CrossRef Full Text | Google Scholar
7. Khan S, Adil Khan M, Chu Y-M. New converses of Jensen inequality via Green functions with applications. RACSAM . (2020)114: 1–14. doi: 10. 1007/s13398-020-00843-1
CrossRef Full Text | Google Scholar
8. Adil Khan M, Pečarić Ð, Pečarić J. New refinement of the Jensen inequality associated to certain functions with applications. J Inequal Appl.(2020)2020: 1–11. doi: 10. 1186/s13660-020-02343-7
CrossRef Full Text | Google Scholar
9. Bakula MK, Özdemir ME, Pečarić J. Hadamard type inequalities for m-convex and (α, m)-convex functions. J Inequal Pure Appl Math.(2008)9: 1–12.
10. Bombardelli M, Varošanec S. Properties of h-convex functions related to the Hermite-Hadamard-Fejér inequalities. Comput Math Appl.(2009)58: 1869–77. doi: 10. 1016/j. camwa. 2009. 07. 073
CrossRef Full Text | Google Scholar
11. Khan J, Adil Khan M, Pečarić J. On Jensen’s type inequalities via generalized majorization inequalities. Filomat . (2018)32: 5719–33. doi: 10. 2298/FIL1816719K
CrossRef Full Text | Google Scholar
12. Dragomir SS, Pearce CEM. Jensen’s inequality for quasi-convex functions. Numer Algebra Control Opt . (2012)2: 279–91. doi: 10. 3934/naco. 2012. 2. 279
CrossRef Full Text | Google Scholar
13. Wang M-K, Zhang W, Chu Y-M. Monotonicity, convexity and inequalities involving the generalized elliptic integrals. Acta Math Sci.(2019)39B: 1440–50. doi: 10. 1007/s10473-019-0520-z
PubMed Abstract | CrossRef Full Text | Google Scholar
14. Wu S-H, Chu Y-M. Schur m -power convexity of generalized geometric Bonferroni mean involving three parameters. J Inequal Appl . (2019)2019: 1–11. doi: 10. 1186/s13660-019-2013-y
CrossRef Full Text | Google Scholar
15. Jain S, Mehrez K, Baleanu D, Agarwal P. Certain Hermite-Hadamard inequalities for logarithmically convex functions with applications. Mathematics.(2019)7: 1–12. doi: 10. 3390/math7020163
CrossRef Full Text | Google Scholar
16. Agarwal P, Jleli M, Tomar M. Certain Hermite-Hadamard type inequalities via generalized k-fractional integrals. J Inequal Appl.(2017)2017: 1–10. doi: 10. 1186/s13660-017-1318-y
PubMed Abstract | CrossRef Full Text | Google Scholar
17. Agarwal P. Some inequalities involving Hadamard-type k-fractional integral operators. Math Method Appl Sci.(2017)40: 3882–91. doi: 10. 1002/mma. 4270
CrossRef Full Text | Google Scholar
18. Liu Z, Yang W, Agarwal P. Certain Chebyshev type inequalities involving the generalized fractional integral operator. J Comput Anal Appl.(2017)22: 999–1014.
19. Choi J, Agarwal P. Certain inequalities involving pathway fractional integral operators. Kyungpook Math J.(2016)56: 1161–8. doi: 10. 5666/KMJ. 2016. 56. 4. 1161
CrossRef Full Text | Google Scholar
20. Mehrez K, Agarwal P. New Hermite-Hadamard type integral inequalities for convex functions and their applications. J Comput Appl Math.(2019)350: 274–85. doi: 10. 2140/pjm. 1964. 14. 243
CrossRef Full Text | Google Scholar
21. Chen X. New convex functions in linear spaces and Jensen’s discrete inequality. J Inequal Appl.(2013)2013: 1–8. doi: 10. 1186/1029-242X-2013-472
CrossRef Full Text | Google Scholar
22. Set E. New inequalities of Ostrowski type for mappings whose derivatives are s-convex in the second sense via fractional integrals. Comput Math Appl.(2012)63: 1147–54. doi: 10. 1016/j. camwa. 2011. 12. 023
CrossRef Full Text | Google Scholar
23. Sarikaya MZ, Set E, Özdemir ME. On new inequalities of Simpson’s type for s-convex functions. Comput Math Appl.(2010)60: 2191–9. doi: 10. 1016/j. camwa. 2010. 07. 033
PubMed Abstract | CrossRef Full Text | Google Scholar
24. Alomari M, Darus M, Dragomir SS, Cerone P. Ostrowski type inequalities for functions whose derivatives are s-convex in the second sense. Appl Math Lett.(2010)23: 1071–6. doi: 10. 1016/j. aml. 2010. 04. 038
CrossRef Full Text | Google Scholar
25. Chen J, Huang X. Some new inequalities of Simpson’s type for s -convex functions via fractional integrals. Filomat . (2017)31: 4989–97. doi: 10. 2298/FIL1715989C
PubMed Abstract | CrossRef Full Text | Google Scholar
26. Özcan S, Işcan I. Some new Hermite-Hadamard type inequalities for s -convex functions and their applications. J Inequal Appl.(2019)2019: 1–11. doi: 10. 1186/s13660-019-2151-2
CrossRef Full Text | Google Scholar
27. Almutairi O, kiliçman A. Integral inequalities for s-convexity via generalized fractional integrals on fractal sets. Mathematics . (2020)8. doi: 10. 3390/math8010053
CrossRef Full Text | Google Scholar
28. Özdemir ME, Yildiz Ç, Akdemir AO, Set E. On some inequalities for s-convex functions and applications. J Inequal Appl.(2013)2013: 1–11. doi: 10. 1186/1029-242X-2013-333
CrossRef Full Text | Google Scholar
29. Adil Khan M, Hanif M, Khan ZAH, Ahmad K, Chu Y-M. Association of Jensen’s inequality for s -convex function with Csiszár divergence. J Inequal Appl . (2019)2019: 1–14. doi: 10. 1186/s13660-019-2112-9
CrossRef Full Text | Google Scholar
30. Butt SI, Mehmood N, Pečarić J. New generalizations of Popoviciu type inequalities via new green functions and Fink’s identity. Trans A Razmadze Math Inst.(2017)171: 293–303. doi: 10. 1016/j. trmi. 2017. 04. 003
PubMed Abstract | CrossRef Full Text | Google Scholar
31. Lovričević N, Pečarić Ð, Pečarić J. Zipf-Mandelbrot law, f −divergences and the Jensen-type interpolating inequalities. J Inequal Appl.(2018)2018: 1–20. doi: 10. 1186/s13660-018-1625-y