[go: up one dir, main page]
More Web Proxy on the site http://driver.im/ skip to main content
article

Eigenclassifiers for combining correlated classifiers

Published: 01 March 2012 Publication History

Abstract

In practice, classifiers in an ensemble are not independent. This paper is the continuation of our previous work on ensemble subset selection [A. Ulas, M. Semerci, O.T. Yildiz, E. Alpaydin, Incremental construction of classifier and discriminant ensembles, Information Sciences, 179 (9) (2009) 1298-1318] and has two parts: first, we investigate the effect of four factors on correlation: (i) algorithms used for training, (ii) hyperparameters of the algorithms, (iii) resampled training sets, (iv) input feature subsets. Simulations using 14 classifiers on 38 data sets indicate that hyperparameters and overlapping training sets have higher effect on positive correlation than features and algorithms. Second, we propose postprocessing before fusing using principal component analysis (PCA) to form uncorrelated eigenclassifiers from a set of correlated experts. Combining the information from all classifiers may be better than subset selection where some base classifiers are pruned before combination, because using all allows redundancy.

References

[1]
F. Alimoğlu, E. Alpaydın, Combining multiple representations and classifiers for pen-based handwritten digit recognition, in: Proceedings of the International Conference on Document Analysis and Recognition, ICDAR'97, 1997.
[2]
Alpaydın, E., Voting over multiple condensed nearest neighbors. Artificial Intelligence Review. v11 i1-5. 115-132.
[3]
Alpaydın, E., Combined 5í2 cv F test for comparing supervised classification learning algorithms. Neural Computation. v11 i8. 1885-1892.
[4]
A. Asuncion, D.J. Newman, UCI machine learning repository, <http://www.ics.uci.edu/~mlearn/MLRepository.html>, 2007.
[5]
S.D. Bay, Combining nearest neighbor classifiers through multiple feature subsets, in: Proceedings of the International Conference on Machine Learning, ICML'98, 1998.
[6]
Biggio, B., Fumera, G. and Roli, F., Multiple classifier systems for robust classifier design in adversarial environments. International Journal of Machine Learning and Cybernetics. v1. 27-41.
[7]
Breiman, L., Bagging predictors. Machine Learning. v24 i2. 123-140.
[8]
Brown, G., Wyatt, J., Harris, R. and Yao, X., Diversity creation methods: a survey and categorisation. Information Fusion. v6 i1. 5-20.
[9]
R. Caruana, A. Niculescu-Mizil, G. Crew, A. Ksikes, Ensemble selection from libraries of models, in: Proceedings of the International Conference on Machine Learning, ICML'04, 2004.
[10]
C.C. Chang C.J. Lin LIBSVM: a library for support vector machines, <http://www.csie.ntu.edu.tw/~cjlin/libsvm>, 2001.
[11]
Demir, C. and Alpaydın, E., Cost-conscious classifier ensembles. Pattern Recognition Letters. v26 i14. 2206-2214.
[12]
Demsar, J., Statistical comparisons of classifiers over multiple data sets. Journal of Machine Learning Research. v7. 1-30.
[13]
Y. Freund, R.E. Schapire, Experiments with a new boosting algorithm, in: Proceedings of the International Conference on Machine Learning, ICML'96, 1996.
[14]
Fumera, G. and Roli, F., A theoretical and experimental analysis of linear combiners for multiple classifier systems. IEEE Transactions on Pattern Analysis Machine Intelligence. v27 i6. 942-956.
[15]
Garcı¿a, S., Fernández, A., Luengo, J. and Herrera, F., Advanced nonparametric tests for multiple comparisons in the design of experiments in computational intelligence and data mining: experimental analysis of power. Information Sciences. v180 i10. 2044-2064.
[16]
Ho, T.K., The random subspace method for constructing decision forests. IEEE Transactions on Pattern Analysis and Machine Intelligence. v20 i8. 832-844.
[17]
Jacobs, R.A., Bias/variance analysis of mixtures-of-experts architectures. Neural Computation. v9 i2. 369-383.
[18]
Jain, A., Nandakumar, K. and Ross, A., Score normalization in multimodal biometric systems. Pattern Recognition. v38. 2270-2285.
[19]
Jolliffe, I.T., Discarding variables in a principal component analysis. II: Real data. Applied Statistics. v22 i1. 21-31.
[20]
Kuncheva, L.I., Combining classifiers: soft computing solutions. In: Pal, S.K. (Ed.), Pattern Recognition: From Classical to Modern Approaches, World Scientific.
[21]
Kuncheva, L.I., Combining pattern classifiers: methods and algorithms. 2004. Wiley-Interscience.
[22]
Kuncheva, L.I., Special issue on diversity in multiple classifier systems. Information Fusion. v6 i1. 1-115.
[23]
Kuncheva, L.I. and Whitaker, C.J., Measures of diversity in classifier ensembles and their relationship with the ensemble accuracy. Machine Learning. v51 i2. 181-207.
[24]
L.I. Kuncheva, C.J. Whitaker, C.A. Ship, R.P. Duin, Is independence good for combining classifiers? in: Proceedings of the 15th International Conference on Pattern Recognition, ICPR'00, 2000.
[25]
Liu, C.-L., Classifier combination based on confidence transformation. Pattern Recognition. v38. 11-28.
[26]
Mallipeddi, R., Mallipeddi, S. and Suganthan, P., Ensemble strategies with adaptive evolutionary programming. Information Sciences. v180 i9. 1571-1581.
[27]
D.D. Margineantu, T.G. Dietterich, Pruning adaptive boosting, in: Proceedings of the International Conference on Machine Learning, ICML'97, 1997.
[28]
Merz, C.J., Using correspondence analysis to combine classifiers. Machine Learning. v36 i1-2. 33-58.
[29]
Merz, C.J. and Pazzani, M.J., A principal components approach to combining regression estimates. Machine Learning. v36 i1-2. 9-32.
[30]
Partridge, D. and Yates, W.B., Engineering multiversion neural-net systems. Neural Computation. v8 i4. 869-893.
[31]
C.E. Rasmussen, R.M. Neal, G. Hinton, D. van Camp, M. Revow, Z. Ghahramani, R. Kustra, R. Tibshirani, Delve data for evaluating learning in valid experiments, <http://www.cs.toronto.edu/~delve/>, 1995-1996.
[32]
Raudys, S., Trainable fusion rules. I: large sample size case. Neural Networks. v19. 1506-1516.
[33]
Rencher, A.C., Interpretation of canonical discriminant functions, canonical variates, and principal components. The American Statistician. v46 i3. 217-225.
[34]
F. Roli, G. Giacinto, G. Vernazza, Methods for designing multiple classifier systems, in: Proceedings of the International Workshop on Multiple Classifier Systems, MCS'01, 2001.
[35]
Ruta, D. and Gabrys, B., Classifier selection for majority voting. Information Fusion. v6 i1. 63-81.
[36]
A.J.C. Sharkey, N.E. Sharkey, U. Gerecke, G.O. Chandroth, The "test and select" approach to ensemble combination, in: Proceedings of the International Workshop on Multiple Classifier Systems, MCS'00, vol. 1857, 2000.
[37]
C. Tamon, J. Xiang, On the boosting pruning problem, in: Proceedings of the European Conference on Machine Learning, ECML'00, 2000.
[38]
Ting, K.M. and Witten, I.H., Issues in stacked generalization. Journal of Artificial Intelligence Research. v10. 271-289.
[39]
Tong, D.L. and Mintram, R., Genetic algorithm-neural network (gann): a study of neural network activation functions and depth of genetic algorithm search applied to feature selection. International Journal of Machine Learning and Cybernetics. v1. 75-87.
[40]
Tumer, K. and Ghosh, J., Error correlation and error reduction in ensemble classifiers. Connection Science. v8 i3. 385-404.
[41]
A. Ulaş, Incremental construction of cost-conscious ensembles using multiple learners and representations in machine learning, Ph.D. thesis, Boğaziçi University. <http://www.cmpe.boun.edu.tr/~ulas/phdthesis.pdf>, 2009.
[42]
Ulaş, A., Semerci, M., Yıldız, O.T. and Alpaydın, E., Incremental construction of classifier and discriminant ensembles. Information Sciences. v179 i9. 1298-1318.
[43]
Wang, L.J., An improved multiple fuzzy nnc system based on mutual information and fuzzy integral. International Journal of Machine Learning and Cybernetics. v2. 25-36.
[44]
Wang, X.-Z. and Dong, C.-R., Improving generalization of fuzzy if-then rules by maximizing fuzzy entropy. IEEE Transactions on Fuzzy Systems. v17. 556-567.
[45]
Wang, X.-Z., Zhai, J.-H. and Lu, S.-X., Induction of multiple fuzzy decision trees based on rough set technique. Information Sciences. v178. 3188-3202.
[46]
Wolpert, D.H., Stacked generalization. Neural Networks. v5. 241-259.
[47]
Xia, R., Zong, C. and Li, S., Ensemble of feature sets and classification algorithms for sentiment classification. Information Sciences. v181 i6. 1138-1152.
[48]
Xiao, J., He, C., Jiang, X. and Liu, D., A dynamic classifier ensemble selection approach for noise data. Information Sciences. v180 i18. 3402-3421.
[49]
Yang, Y., Webb, G.I., Cerquides, J., Korb, K.B., Boughton, J. and Ting, K.M., To select or to weigh: a comparative study of linear combination schemes for superparent-one-dependence estimators. IEEE Transactions on Knowledge and Data Engineering. v19 i12. 1652-1665.
[50]
O.T. Yıldız, E. Alpaydın, Linear discriminant trees, in: Proceedings of the International Conference on Machine Learning, ICML'00, 2000.
[51]
O.T. Yıldız, A. Ulaş, M. Semerci, E. Alpaydın, AYSU: machine learning data sets for model combination, <http://www.cmpe.boun.edu.tr/~ulas/aysu>, 2007.
[52]
Yu, E. and Suganthan, P., Ensemble of niching algorithms. Information Sciences. v180 i15. 2815-2833.
[53]
Zhang, Y., Burer, S. and Street, W.N., Ensemble pruning via semi-definite programming. Journal of Machine Learning Research. v7. 1315-1338.
[54]
Zhou, Z.-H., Wu, J. and Tang, W., Ensembling neural networks: many could be better than all. Artificial Intelligence. v137. 239-263.

Cited By

View all
  • (2018)Classifying multimodal dataThe Handbook of Multimodal-Multisensor Interfaces10.1145/3107990.3107994(49-69)Online publication date: 1-Oct-2018
  • (2018)The Handbook of Multimodal-Multisensor InterfacesundefinedOnline publication date: 1-Oct-2018
  • (2015)Coupling different methods for overcoming the class imbalance problemNeurocomputing10.1016/j.neucom.2015.01.068158:C(48-61)Online publication date: 22-Jun-2015
  • Show More Cited By

Recommendations

Comments

Please enable JavaScript to view thecomments powered by Disqus.

Information & Contributors

Information

Published In

cover image Information Sciences: an International Journal
Information Sciences: an International Journal  Volume 187, Issue
March, 2012
305 pages

Publisher

Elsevier Science Inc.

United States

Publication History

Published: 01 March 2012

Author Tags

  1. Classifier correlation
  2. Classifier design and evaluation
  3. Machine learning

Qualifiers

  • Article

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)0
  • Downloads (Last 6 weeks)0
Reflects downloads up to 27 Jan 2025

Other Metrics

Citations

Cited By

View all
  • (2018)Classifying multimodal dataThe Handbook of Multimodal-Multisensor Interfaces10.1145/3107990.3107994(49-69)Online publication date: 1-Oct-2018
  • (2018)The Handbook of Multimodal-Multisensor InterfacesundefinedOnline publication date: 1-Oct-2018
  • (2015)Coupling different methods for overcoming the class imbalance problemNeurocomputing10.1016/j.neucom.2015.01.068158:C(48-61)Online publication date: 22-Jun-2015
  • (2015)Extended multimodal Eigenclassifiers and criteria for fusion model selectionInformation Sciences: an International Journal10.1016/j.ins.2014.11.034298:C(53-65)Online publication date: 20-Mar-2015
  • (2014)Embedded local feature selection within mixture of expertsInformation Sciences: an International Journal10.1016/j.ins.2014.01.008269(176-187)Online publication date: 1-Jun-2014

View Options

View options

Figures

Tables

Media

Share

Share

Share this Publication link

Share on social media