Necessary conditions for consistent set-based graphical model selection

Graphical model selection, where the goal is to estimate the graph underlying a distribution, is known to be an NP-hard problem. An important issue is to study theoretical limits on the performance of graphical model selection algorithms. In particular, given parameters of the underlying distribution, we want to find a lower bound on the number of samples required for accurate graph estimation. When deriving these theoretical bounds, it is common to treat the learning problem as a communication problem where the observations correspond to noisy messages and the decoding problem infers the graph from the observations. Current analysis of graphical model selection algorithms is limited to studying graph estimators that output a unique graph. In this paper, we consider graph estimators that output a set of graphs, leading to set-based graphical model selection (SB-GMS). This has connections to list-decoding where a decoder outputs a list of possible codewords instead of a single codeword. Our main contribution is to derive necessary conditions for accurate SB-GMS for various classes of graphical models and show reduction in the number of samples required for consistent estimation. Further, we derive necessary conditions on the cardinality of the set-based estimates given graph parameters.

[1]  T. Speed,et al.  Gaussian Markov Distributions over Finite Graphs , 1986 .

[2]  E. Levina,et al.  Joint estimation of multiple graphical models. , 2011, Biometrika.

[3]  Elchanan Mossel,et al.  Reconstruction of Markov Random Fields from Samples: Some Observations and Algorithms , 2007, SIAM J. Comput..

[4]  Martin J. Wainwright,et al.  Information-Theoretic Limits of Selecting Binary Graphical Models in High Dimensions , 2009, IEEE Transactions on Information Theory.

[5]  Vincent Y. F. Tan,et al.  Learning High-Dimensional Markov Forest Distributions: Analysis of Error Rates , 2010, J. Mach. Learn. Res..

[6]  J. Besag Spatial Interaction and the Statistical Analysis of Lattice Systems , 1974 .

[7]  Peter Elias,et al.  List decoding for noisy channels , 1957 .

[8]  Andrea Montanari,et al.  Which graphical models are difficult to learn? , 2009, NIPS.

[9]  N. Meinshausen,et al.  High-dimensional graphs and variable selection with the Lasso , 2006, math/0608017.

[10]  J. Lafferty,et al.  High-dimensional Ising model selection using ℓ1-regularized logistic regression , 2010, 1010.0311.

[11]  Young-Han Kim,et al.  State Amplification , 2008, IEEE Transactions on Information Theory.

[12]  Martin J. Wainwright,et al.  Information-theoretic bounds on model selection for Gaussian Markov random fields , 2010, 2010 IEEE International Symposium on Information Theory.

[13]  Venkatesan Guruswami,et al.  List decoding of error correcting codes , 2001 .

[14]  Alexandre d'Aspremont,et al.  Model Selection Through Sparse Max Likelihood Estimation Model Selection Through Sparse Maximum Likelihood Estimation for Multivariate Gaussian or Binary Data , 2022 .