Welcome to the new version of CaltechAUTHORS. Login is currently restricted to library staff. If you notice any issues, please email coda@library.caltech.edu
Published March 2011 | Submitted + Published
Journal Article Open

A Large-Deviation Analysis of the Maximum-Likelihood Learning of Markov Tree Structures

Abstract

The problem of maximum-likelihood (ML) estimation of discrete tree-structured distributions is considered. Chow and Liu established that ML-estimation reduces to the construction of a maximum-weight spanning tree using the empirical mutual information quantities as the edge weights. Using the theory of large-deviations, we analyze the exponent associated with the error probability of the event that the ML-estimate of the Markov tree structure differs from the true tree structure, given a set of independently drawn samples. By exploiting the fact that the output of ML-estimation is a tree, we establish that the error exponent is equal to the exponential rate of decay of a single dominant crossover event. We prove that in this dominant crossover event, a non-neighbor node pair replaces a true edge of the distribution that is along the path of edges in the true tree graph connecting the nodes in the non-neighbor pair. Using ideas from Euclidean information theory, we then analyze the scenario of ML-estimation in the very noisy learning regime and show that the error exponent can be approximated as a ratio, which is interpreted as the signal-to-noise ratio (SNR) for learning tree distributions. We show via numerical experiments that in this regime, our SNR approximation is accurate.

Additional Information

© 2011 IEEE. Manuscript received May 06, 2009; revised October 19, 2010; accepted November 18, 2010. Date of current version February 18, 2011. This work was supported in part by A*STAR, Singapore, by a MURI funded through ARO Grant W911NF-06-1-0076 and by AFOSR Grant FA9550-08-1-0180 and in part by the Army Research Office MURI Program under award W911NF-08-1-0238. The material in this paper was presented in part at the International Symposium on Information Theory (ISIT), Seoul, Korea, June 2009. V. Y. F. Tan performed this work while at MIT. The authors would like to thank the anonymous referees and Associate Editor A. Krzyzak who have helped to improve the exposition. One reviewer, in particular, helped highlight the connection of this work with robust hypothesis testing, leading to Section V-D. The authors would also like to thank Prof. L. Zheng, M. Agrawal, and A. Olshevsky for many stimulating discussions.

Attached Files

Published - 05714274.pdf

Submitted - 0905.0940.pdf

Files

05714274.pdf
Files (1.3 MB)
Name Size Download all
md5:e0d83794d4d7b4bd3aab9a5c1575bdff
821.7 kB Preview Download
md5:612b41fe17daa89bea0b6b9e9167ba33
497.4 kB Preview Download

Additional details

Created:
August 19, 2023
Modified:
October 17, 2023