TY - JOUR
T1 - Unsupervised Weight Parameter Estimation Method for Ensemble Learning
AU - Uchida, Masato
AU - Maehara, Yousuke
AU - Shioya, Hiroyuki
N1 - Funding Information:
A part of this paper appeared in the Post-Conference Proceedings of ICONIP 2007 [20]. This work was supported in part by the Japan Society for the Promotion of Science through a Grant-in-Aid for Scientific Research (S) (18100001).
PY - 2011/12
Y1 - 2011/12
N2 - When there are multiple trained predictors, one may want to integrate them into one predictor. However, this is challenging if the performances of the trained predictors are unknown and labeled data for evaluating their performances are not given. In this paper, a method is described that uses unlabeled data to estimate the weight parameters needed to build an ensemble predictor integrating multiple trained component predictors. It is readily derived from a mathematical model of ensemble learning based on a generalized mixture of probability density functions and corresponding information divergence measures. Numerical experiments demonstrated that the performance of our method is much better than that of simple average-based ensemble learning, even when the assumption placed on the performances of the component predictors does not hold exactly.
AB - When there are multiple trained predictors, one may want to integrate them into one predictor. However, this is challenging if the performances of the trained predictors are unknown and labeled data for evaluating their performances are not given. In this paper, a method is described that uses unlabeled data to estimate the weight parameters needed to build an ensemble predictor integrating multiple trained component predictors. It is readily derived from a mathematical model of ensemble learning based on a generalized mixture of probability density functions and corresponding information divergence measures. Numerical experiments demonstrated that the performance of our method is much better than that of simple average-based ensemble learning, even when the assumption placed on the performances of the component predictors does not hold exactly.
KW - Ensemble learning
KW - Exponential mixture model
KW - Kullback-Leibler divergence
KW - Unsupervised learning
UR - http://www.scopus.com/inward/record.url?scp=82355169807&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=82355169807&partnerID=8YFLogxK
U2 - 10.1007/s10852-011-9157-1
DO - 10.1007/s10852-011-9157-1
M3 - Article
AN - SCOPUS:82355169807
SN - 1570-1166
VL - 10
SP - 307
EP - 322
JO - Journal of Mathematical Modelling and Algorithms
JF - Journal of Mathematical Modelling and Algorithms
IS - 4
ER -