Learning in Gibbsian Fields: How Accurate and How Fast Can It Be?

11 years 1 months ago
Learning in Gibbsian Fields: How Accurate and How Fast Can It Be?
?Gibbsian fields or Markov random fields are widely used in Bayesian image analysis, but learning Gibbs models is computationally expensive. The computational complexity is pronounced by the recent minimax entropy (FRAME) models which use large neighborhoods and hundreds of parameters [22]. In this paper, we present a common framework for learning Gibbs models. We identify two key factors that determine the accuracy and speed of learning Gibbs models: The efficiency of likelihood functions and the variance in approximating partition functions using Monte Carlo integration. We propose three new algorithms. In particular, we are interested in a maximum satellite likelihood estimator, which makes use of a set of precomputed Gibbs models called ?satellites? to approximate likelihood functions. This algorithm can approximately estimate the minimax entropy model for textures in seconds in a HP workstation. The performances of various learning algorithms are compared in our experiments.
Song Chun Zhu, Xiuwen Liu
Added 12 Oct 2009
Updated 30 Oct 2009
Type Conference
Year 2000
Where CVPR
Authors Song Chun Zhu, Xiuwen Liu
Comments (0)