Tong Zhang
Neural Computation
We consider Bayesian mixture approaches, where a predictor is constructed by forming a weighted average of hypotheses from some space of functions. While such procedures are known to lead to optimal predictors in several cases, where sufficiently accurate prior information is available, it has not been clear how they perform when some of the prior assumptions are violated. In this paper we establish data-dependent bounds for such procedures, extending previous randomized approaches such as the Gibbs algorithm to a fully Bayesian setting. The finite-sample guarantees established in this work enable the utilization of Bayesian mixture approaches in agnostic settings, where the usual assumptions of the Bayesian paradigm fail to hold. Moreover, the bounds derived can be directly applied to non-Bayesian mixture approaches such as Bagging and Boosting.
Tong Zhang
Neural Computation
Jinbo Bi, Tong Zhang, et al.
KDD 2004
Vijay S. Iyengar, Chidanand Apte, et al.
KDD 2000
Rie Kubota Ando, Mark Dredze, et al.
TREC 2005