Figure-ground segmentation using factor graphs

作者:

Highlights:

摘要

Foreground–background segmentation has recently been applied [S.X. Yu, J. Shi, Object-specific figure-ground segregation, Computer Vision and Pattern Recognition (CVPR), 2003; S. Kumar, M. Hebert, Man-made structure detection in natural images using a causal multiscale random field, Computer Vision and Pattern Recognition (CVPR), 2003] to the detection and segmentation of specific objects or structures of interest from the background as an efficient alternative to techniques such as deformable templates [A.L. Yuille, Deformable templates for face recognition. Journal of Cognitive Neuroscience 3 (1) (1991)]. We introduce a graphical model (i.e. Markov random field)-based formulation of structure-specific figure-ground segmentation based on simple geometric features extracted from an image, such as local configurations of linear features, that are characteristic of the desired figure structure. Our formulation is novel in that it is based on factor graphs, which are graphical models that encode interactions among arbitrary numbers of random variables. The ability of factor graphs to express interactions higher than pairwise order (the highest order encountered in most graphical models used in computer vision) is useful for modeling a variety of pattern recognition problems. In particular, we show how this property makes factor graphs a natural framework for performing grouping and segmentation, and demonstrate that the factor graph framework emerges naturally from a simple maximum entropy model of figure-ground segmentation.We cast our approach in a learning framework, in which the contributions of multiple grouping cues are learned from training data, and apply our framework to the problem of finding printed text in natural scenes. Experimental results are described, including a performance analysis that demonstrates the feasibility of the approach.

论文关键词:Figure-ground segmentation,Belief propagation,Factor graphs,Text detection

论文评审过程:Received 25 October 2007, Revised 6 December 2008, Accepted 3 February 2009, Available online 28 February 2009.

论文官网地址:https://doi.org/10.1016/j.imavis.2009.02.006