Information Inequalities via Submodularity and a Problem in Extremal Graph Theory
Abstract
:1. Introduction
2. Preliminaries and Notation
- denotes the set of natural numbers.
- denotes the set of real numbers, and denotes the set of nonnegative real numbers.
- Ø denotes the empty set.
- denotes the power set of a set (i.e., the set of all subsets of ).
- denotes the complement of a subset in .
- is an indicator of E; it is 1 if event E is satisfied, and zero otherwise.
- for every ;
- denotes an n-dimensional random vector;
- is a random vector for a nonempty subset ; if , then is an empty set, and conditioning on is void.
- Let X be a discrete random variable that takes its values on a set , and let be the probability mass function (PMF) of X. The Shannon entropy of X is given by
- The binary entropy function is given by
- Let X and Y be discrete random variables with a joint PMF , and a conditional PMF of X given Y denoted by . The conditional entropy of X given Y is defined as
- The mutual information between X and Y is symmetric in X and Y, and it is given by
- The conditional mutual information between two random variables X and Y, given a third random variable Z, is symmetric in X and Y and it is given by
- For continuous random variables, the sums in (1) and (3) are replaced with integrals, and the PMFs are replaced with probability densities. The entropy of a continuous random variable is named differential entropy.
- For an n-dimensional random vector , the entropy power of is given by
- Conditioning cannot increase the entropy, i.e.,
- Nonnegativity of the (conditional) mutual information: In light of (5) and (8), with equality if and only if X and Y are independent. More generally, with equality if and only if X and Y are conditionally independent given Z.
- (a)
- The set function , given byis a rank function.
- (b)
- The set function , given byis supermodular, monotonically increasing, and .
- (c)
- The set function , given byis submodular, , but f is not a rank function. The latter holds since the equality , for all , implies that f is not a monotonic function.
- (d)
- Let be disjoint subsets, and let the entries of the random vector be conditionally independent given . Then, the set function given byis a rank function.
- (e)
- Let be independent random variables, and let be given byThen, f is a rank function.
3. Inequalities via Submodularity
3.1. A New Methodology
- (a)
- If f is submodular, and g is monotonically increasing and convex, then the sequence is monotonically decreasing, i.e.,In particular,
- (b)
- If f is submodular, and g is monotonically decreasing and concave, then the sequence is monotonically increasing.
- (c)
- If f is supermodular, and g is monotonically increasing and concave, then the sequence is monotonically increasing.
- (d)
- If f is supermodular, and g is monotonically decreasing and convex, then the sequence is monotonically decreasing.
- f is a rank function,
- or there is such that with ,
- is a sequence such that for all with ,
- (a)
- For and
- (b)
- If f is also monotonically increasing (i.e., f is a rank function), then for
- (a)
- Setting in (25) implies that, for all ,
- (b)
- Consequently, setting in (28) gives
- (a)
- The sequences
- (b)
- The sequence
- (c)
- For every , the sequences
3.2. Connections to a Generalized Version of Shearer’s Lemma and Other Results in the Literature
- (a)
- If f is non-negative and submodular, and every element in Ω is included in at least of the subsets , then
- (b)
- If f is a rank function, , and every element in is included in at least of the subsets , then
- (a)
- For every ,and equivalently,
- (b)
- For every ,
4. Proofs
4.1. Proof of Theorem 1
4.2. Proof of Corollary 1
- if ;
- scales like if with for some .
4.3. Proof of Corollary 2
5. A Problem in Extremal Graph Theory
5.1. Problem Formulation
5.2. Problem Motivation
5.3. Analysis
- (a)
- If and for any such that , then there are at least vectors whose subvectors coincide with , i.e., the integer satisfiesBy definition, the integer always exists, andIf no information is available about the value of , then it can be taken by default to be equal to 2 (since by assumption the two vectors and satisfy the equality ).
- (b)
- If and for any such that , then there are at least vectors whose subvectors coincide with , i.e., the integer satisfiesBy definition, the integer always exists, andLikewise, if no information is available about the value of , then it can be taken by default to be equal to 1 (since satisfies the requirement about its subvector in (94)).
- If , then
- If , then
5.4. Comparison of Bounds
- (a)
- For , let the integers and (be, preferably, the maximal possible values to) satisfy the requirements in (92) and (94), respectively. Then,
- (b)
- A loosened bound, which only depends on the cardinality of the set , is obtained by setting the default values and . It is then given byand, if , then the (overall) number of edges in G satisfies
- (c)
- The refined upper bound on the RHS of (119) and the loosened upper bound on the RHS of (120) improve the trivial bound , if and only if or , respectively (see Example 1).
5.5. Influence of Fixed-Size Subsets of Bits
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
Appendix A. Proof of Proposition 1
- .
- Submodularity: If , thenThis proves the submodularity of f, while also showing that
- Monotonicity: If , then
- , and .
- Supermodularity: If , then
- Monotonicity: If , then
- .
- Submodularity: If , thenConsequently, combining (A7) and (A8) gives
- Monotonicity: If , then
- .
- Submodularity: Let . DefineFrom the independence of the random variables , it follows that and W are independent. Hence, we getCombining (A12) and (A13) gives (11).
- Monotonicity: If , then since are independent random variables, (A11) implies that U and W are independent and . Hence,
Appendix B. Proof of Proposition 4
References
- Cover, T.M.; Thomas, J.A. Elements of Information Theory, 2nd ed.; John Wiley & Sons: Hoboken, NJ, USA, 2006. [Google Scholar]
- Dembo, A.; Cover, T.M.; Thomas, J.A. Information theoretic inequalities. IEEE Trans. Inf. Theory 1991, 37, 1501–1518. [Google Scholar] [CrossRef] [Green Version]
- Chan, T. Recent progresses in characterising information inequalities. Entropy 2011, 13, 379–401. [Google Scholar] [CrossRef]
- Martin, S.; Padró, C.; Yang, A. Secret sharing, rank inequalities, and information inequalities. IEEE Trans. Inf. Theory 2016, 62, 599–610. [Google Scholar] [CrossRef]
- Babu, S.A.; Radhakrishnan, J. An entropy-based proof for the Moore bound for irregular graphs. In Perspectives on Computational Complexity; Agrawal, M., Arvind, V., Eds.; Birkhäuser: Cham, Switzerland, 2014; pp. 173–182. [Google Scholar]
- Boucheron, S.; Lugosi, G.; Massart, P. Concentration Inequalities - A Nonasymptotic Theory of Independence; Oxford University Press: Oxford, UK, 2013. [Google Scholar]
- Chung, F.R.K.; Graham, L.R.; Frankl, P.; Shearer, J.B. Some intersection theorems for ordered sets and graphs. J. Comb. Theory Ser. A 1986, 43, 23–37. [Google Scholar] [CrossRef] [Green Version]
- Erdos, P.; Rényi, A. On two problems of information theory. Publ. Math. Inst. Hung. Acad. Sci. 1963, 8, 241–254. [Google Scholar]
- Friedgut, E. Hypergraphs, entropy and inequalities. Am. Math. Mon. 2004, 111, 749–760. [Google Scholar] [CrossRef] [Green Version]
- Jukna, S. Extremal Combinatorics with Applications in Computer Science, 2nd ed.; Springer: Berlin/Heidelberg, Germany, 2011. [Google Scholar]
- Kaced, T.; Romashchenko, A.; Vereshchagin, N. A conditional information inequality and its combinatorial applications. IEEE Trans. Inf. Theory 2018, 64, 3610–3615. [Google Scholar] [CrossRef]
- Kahn, J. An entropy approach to the hard-core model on bipartite graphs. Comb. Comput. 2001, 10, 219–237. [Google Scholar] [CrossRef]
- Kahn, J. Entropy, independent sets and antichains: A new approach to Dedekind’s problem. Proc. Am. Math. Soc. 2001, 130, 371–378. [Google Scholar] [CrossRef]
- Madiman, M.; Marcus, A.W.; Tetali, P. Entropy and set cardinality inequalities for partition-determined functions. Random Struct. Algorithms 2012, 40, 399–424. [Google Scholar] [CrossRef] [Green Version]
- Madiman, M.; Marcus, A.W.; Tetali, P. Information–theoretic inequalities in additive combinatorics. In Proceedings of the 2010 IEEE Information Theory Workshop, Cairo, Egypt, 6–8 January 2010. [Google Scholar]
- Pippenger, N. An information–theoretic method in combinatorial theory. J. Comb. Ser. A 1977, 23, 99–104. [Google Scholar] [CrossRef] [Green Version]
- Pippenger, N. Entropy and enumeration of boolean functions. IEEE Trans. Inf. Theory 1999, 45, 2096–2100. [Google Scholar] [CrossRef]
- Radhakrishnan, J. An entropy proof of Bregman’s theorem. J. Comb. Theory Ser. A 1997, 77, 161–164. [Google Scholar] [CrossRef] [Green Version]
- Radhakrishnan, J. Entropy and counting. In Computational Mathematics, Modelling and Algorithms; Narosa Publishers: New Delhi, India, 2001; pp. 1–25. [Google Scholar]
- Sason, I. A generalized information–theoretic approach for bounding the number of independent sets in bipartite graphs. Entropy 2021, 23, 270. [Google Scholar] [CrossRef]
- Sason, I. Entropy-based proofs of combinatorial results on bipartite graphs. In Proceedings of the 2021 IEEE International Symposium on Information Theory, Melbourne, Australia, 12–20 July 2021; pp. 3225–3230. [Google Scholar]
- Madiman, M.; Tetali, P. Information inequalities for joint distributions, interpretations and applications. IEEE Trans. Inf. Theory 2010, 56, 2699–2713. [Google Scholar] [CrossRef]
- Bach, F. Learning with submodular functions: A convex optimization perspective. Found. Trends Mach. Learn. 2013, 6, 145–373. [Google Scholar] [CrossRef] [Green Version]
- Chen, Q.; Cheng, M.; Bai, B. Matroidal entropy functions: A quartet of theories of information, matroid, design and coding. Entropy 2021, 23, 323. [Google Scholar] [CrossRef]
- Fujishige, S. Polymatroidal dependence structure of a set of random variables. Inf. Control. 1978, 39, 55–72. [Google Scholar] [CrossRef] [Green Version]
- Fujishige, S. Submodular Functions and Optimization, 2nd ed.; Annals of Discrete Mathematics Series; Elsevier: Amsterdam, The Netherlands, 2005; Volume 58. [Google Scholar]
- Iyer, R.; Khargonkar, N.; Bilems, J.; Asnani, H. Generalized submodular information measures: Theoretical properties, examples, optimization algorithms, and applications. IEEE Trans. Inf. Theory 2022, 68, 752–781. [Google Scholar] [CrossRef]
- Krause, A.; Guestrin, C. Near-optimal nonmyopic value of information in graphical models. In Proceedings of the Twenty-First Conference on Uncertainty in Artificial Intelligence (UAI 2005), Edinburgh, UK, 26–29 July 2005; pp. 324–331. [Google Scholar]
- Lovász, L. Submodular functions and convexity. In Mathematical Programming The State of the Art; Bachem, A., Korte, B., Grotschel, M., Eds.; Springer: Berlin/Heidelberg, Germany, 1983; pp. 235–257. [Google Scholar]
- Tian, C. Inequalities for entropies of sets of subsets of random variables. In Proceedings of the 2011 IEEE International Symposium on Information Theory, Saint Petersburg, Russia, 31 July–5 August 2011; pp. 1950–1954. [Google Scholar]
- Kishi, Y.; Ochiumi, N.; Yanagida, M. Entropy inequalities for sums over several subsets and their applications to average entropy. In Proceedings of the 2014 IEEE International Symposium on Information Theory (ISIT 2014), Honolulu, HI, USA, 30 June–4 July 2014; pp. 2824–2828. [Google Scholar]
- Shannon, C.E. A Mathematical Theory of Communication. Bell Syst. Tech. J. 1948, 27, 379–423, 623–656. [Google Scholar] [CrossRef] [Green Version]
- Madiman, M.; Mellbourne, J.; Xeng, P. Forward and reverse entropy power inequalities in convex geometry. In Convexity and Concentration; Carlen, E., Madiman, M., Werner, E.M., Eds.; IMA Volumes in Mathematics and Its Applications; Springer: Berlin/Heidelberg, Germany, 2017; Volume 161, pp. 427–485. [Google Scholar]
- Han, T.S. Nonnegative entropy measures of multivariate symmetric correlations. Inf. Control. 1978, 36, 133–156. [Google Scholar] [CrossRef] [Green Version]
- Polyanskiy, Y.; Wu, Y. Lecture Notes on Information Theory, version 5. Available online: http://people.lids.mit.edu/yp/homepage/data/itlectures_v5.pdf (accessed on 15 May 2019).
- Bollobás, B. Extremal Graph Theory; Academic Press: Cambridge, MA, USA, 1978. [Google Scholar]
- Madiman, M. On the entropy of sums. In Proceedings of the 2008 IEEE Information Theory Workshop, Porto, Portugal, 5–9 May 2008. [Google Scholar]
- Tao, T. Sumset and inverse sumset theory for Shannon entropy. Comb. Comput. 2010, 19, 603–639. [Google Scholar] [CrossRef] [Green Version]
- Kontoyiannis, I.; Madiman, M. Sumset and inverse sumset inequalities for differential entropy and mutual information. IEEE Trans. Inf. Theory 2014, 60, 4503–4514. [Google Scholar] [CrossRef] [Green Version]
- Artstein, S.; Ball, K.M.; Barthe, F.; Naor, A. Solution of Shannon’s problem on the monotonicity of entropy. J. Am. Soc. 2004, 17, 975–982. [Google Scholar] [CrossRef]
- Madiman, M.; Barron, A. Generalized entropy power inequalities and monotonicity properties of information. IEEE Trans. Inf. Theory 2007, 53, 2317–2329. [Google Scholar] [CrossRef] [Green Version]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2022 by the author. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Sason, I. Information Inequalities via Submodularity and a Problem in Extremal Graph Theory. Entropy 2022, 24, 597. https://doi.org/10.3390/e24050597
Sason I. Information Inequalities via Submodularity and a Problem in Extremal Graph Theory. Entropy. 2022; 24(5):597. https://doi.org/10.3390/e24050597
Chicago/Turabian StyleSason, Igal. 2022. "Information Inequalities via Submodularity and a Problem in Extremal Graph Theory" Entropy 24, no. 5: 597. https://doi.org/10.3390/e24050597
APA StyleSason, I. (2022). Information Inequalities via Submodularity and a Problem in Extremal Graph Theory. Entropy, 24(5), 597. https://doi.org/10.3390/e24050597