资源论文Maximum Likelihood Learning With Arbitrary Treewidth via Fast-Mixing Parameter Sets

Maximum Likelihood Learning With Arbitrary Treewidth via Fast-Mixing Parameter Sets

2020-02-04 | |  76 |   37 |   0

Abstract

    Inference is typically intractable in high-treewidth undirected graphical models, making maximum likelihood learning a challenge. One way to overcome this is to restrict parameters to a tractable set, most typically the set of tree-structured parameters. This paper explores an alternative notion of a tractable set, namely a set of “fast-mixing parameters” where Markov chain Monte Carlo (MCMC) inference can be guaranteed to quickly converge to the stationary distribution. While it is common in practice to approximate the likelihood gradient using samples obtained from MCMC, such procedures lack theoretical guarantees. This paper proves that for any exponential family with bounded sufficient statistics, (not just graphical models) when parameters are constrained to a fast-mixing set, gradient descent with gradients approximated by sampling will approximate the maximum likelihood solution inside the set with high-probability. When unregularized, to find a solution image.png-accurate in log-likelihood requires a total amount of effort cubic in image.png, disregarding logarithmic factors. When ridge-regularized, strong convexity allows a solution image.png-accurate in parameter distance with effort quadratic in image.png. Both of these provide of a fully-polynomial time randomized approximation scheme.

上一篇:On the Optimality of Classifier Chain for Multi-label Classification

下一篇:Teaching Machines to Read and Comprehend Karl Moritz Hermann† Toma?s? Koc?isky?†‡ Edward Grefenstette†

用户评价
全部评价

热门资源

  • Learning to learn...

    The move from hand-designed features to learned...

  • A Mathematical Mo...

    Direct democracy, where each voter casts one vo...

  • Stratified Strate...

    In this paper we introduce Stratified Strategy ...

  • Rating-Boosted La...

    The performance of a recommendation system reli...

  • Hierarchical Task...

    We extend hierarchical task network planning wi...