Paper:

# A MultiBoosting Based Transfer Learning Algorithm

## Xiaobo Liu^{*}, Guangjun Wang^{*}, Zhihua Cai^{**}, and Harry Zhang^{***}

^{*}School of Automation, China University of Geosciences

388 Lumo Road, Wuhan, Hubei 430074, China

^{**}School of Computer Science, China University of Geosciences

388 Lumo Road, Wuhan, Hubei 430074, China

^{***}Faculty of Computer Science, University of New Brunswick

P.O. Box 4400, Fredericton, NB E3B 5A3, Canada

Ensemble learning is sophisticated machine learning use to solve many problems in practical applications. MultiBoosting, a cutting-edge learning approach in ensemble learning, is combined with AdaBoost and wagging. It retains AdaBoost’s bias reduction while adding wagging’s variance reduction to that already obtained by AdaBoost, thus reducing the total number of errors in classification. Data characteristics do not always follow traditional machine learning rules, however, so transfer learning acts to solve this problem. We propose a TrMultiBoosting algorithm, composed of MultiBoosting and state-of-the-art transfer learning algorithm TrAdaBoost for transfer learning. We use naive bayes as the basic learning algorithm. TrMultiBoosting has proven to present a decision committee with higher prediction accuracy on UCI data sets than either TrAdaBoost or MultiBoosting.

*J. Adv. Comput. Intell. Intell. Inform.*, Vol.19, No.3, pp. 381-388, 2015.

- [1] I. H. Witten, E. Frank, and M. A. Hall, “Data Mining: Practical Machine Learning Tools and Techniques (3
^{rd}Edition),” Morgan Kaufmann, Burlington, 2011. - [2] T. G. Dietterich, “Ensemble Methods in Machine Learning,” Proc. of the 1
^{st}Int. Workshop on Multiple Classifier Systems, pp. 1-15. Springer-Verlag, London, 2000. - [3] S. Pan and Q. Yang, “A Survey on Transfer Learning,” IEEE Trans. on Knowledge and Data Engineering, Vol.22, pp. 1345-1359, 2011.
- [4] W. Dai, Q. Yang, G. Xue, et al., “Boosting for Transfer Learning,” Proc. of the 24
^{th}Annual Int. Conf. on Machine Learning (ICML’07), pp. 193-200, New York, IEEE Press, 2007. - [5] E. Eaton and M. DesJardins, “Set-Based Boosting for Instance-Level Transfer,” IEEE Int. Conf. on Data Mining Workshops, pp. 422-428, IEEE Press, Washington, 2009.
- [6] T. Kamishima, M. Hamasaki, and S. Akaho, “TrBagg: A Simple Transfer Learning Method and Its Application to Personalization in Collaborative Tagging,” 9
^{th}IEEE Int. Conf. on Data Mining (ICDM 2009), pp. 219-228, IEEE Press, Washington, 2009. - [7] L. Breiman, “Bagging predictors,” Machine Learning, Vol.24, pp. 123-140, 1996.
- [8] L. La, Q. Guo, Q. Cao, et al., “Transfer learning with reasonable boosting strategy,” Neural Computing & Applications, Vol.24, No.3-4, pp. 807-816, 2014.
- [9] G. I. Webb, “MultiBoosting: A Technique for Combining Boosting and Wagging,” Machine Learning, Vol.40, pp. 156-196, 2000.
- [10] Y. Freund and R. E. Schapire, “Experiments with a New Boosting Algorithm,” Proc. of the 13th Int. Conf. on Machine Learning (ICML 2006), pp. 148-156, 2006.
- [11] E. Bauer and R. Kohavi, “An Empirical Comparison of Voting Classification Algorithms: Bagging, Boosting, and Variants,” Machine Learning, Vol.36, pp. 105-139, 1999.
- [12] T. M. Mitchell, “Machine Learning,” McGraw Hill, New York, 1997.
- [13] Y. Shi, Z. Lan, and W. Liu, “Extending Semi-supervised Learning Methods for Inductive Transfer Learning,” 9
^{th}IEEE Int. Conf. on Data Mining (ICDM 2009), pp. 483-492, IEEE Press, Miami, 2009.

This article is published under a Creative Commons Attribution-NoDerivatives 4.0 Internationa License.