Kikuchi-Bayes: MArkov Networks for Classification

We propose a simple and efficient approach to building undirected probabilistic classification models (Markov networks) that extend the naive Bayes classifiers and outperform the existing directed probabilistic classifiers of similar complexity (e.g. Bayesian network with same cluster size). The models are represented as sets of cliques, not necessarily maximal, and the probability density functions can be estimated in closed form that mirrors the cluster variation method (Kikuchi approximation). We employ a highly efficient Bayesian learning algorithm, based on integrating along a hill-climb in the structure space. We present promising empirical results on 46 benchmarks.

By: Aleks Jakulin; Irina Rish

Published in: RC23602 in 2005


This Research Report is available. This report has been submitted for publication outside of IBM and will probably be copyrighted if accepted for publication. It has been issued as a Research Report for early dissemination of its contents. In view of the transfer of copyright to the outside publisher, its distribution outside of IBM prior to publication should be limited to peer communications and specific requests. After outside publication, requests should be filled only by reprints or legally obtained copies of the article (e.g., payment of royalties). I have read and understand this notice and am a member of the scientific community outside or inside of IBM seeking a single copy only.


Questions about this service can be mailed to .