[go: up one dir, main page]
More Web Proxy on the site http://driver.im/ skip to main content
article

Randomizing Outputs to Increase Prediction Accuracy

Published: 01 September 2000 Publication History

Abstract

Bagging and boosting reduce error by changing both the inputs and outputs to form perturbed training sets, growing predictors on these perturbed training sets and combining them. An interesting question is whether it is possible to get comparable performance by perturbing the outputs alone. Two methods of randomizing outputs are experimented with. One is called output smearing and the other output flipping. Both are shown to consistently do better than bagging.

References

[1]
An, G. (1996). The effects of adding noise during backpropagation training on generalization performance. Neural Computation, 6, 643-674.
[2]
Breiman, L. (1996a). Bagging predictors. Machine Learning, 26(2), 123-140.
[3]
Breiman, L. (1996b). The heuristics of instability in model selection. Annals of Statistics, 24, 2350-2383.
[4]
Breiman, L. (1997). Prediction games and arcing algorithms. Technical Report 504, Statistics Department, University of California at Berkeley. Available at www.stat.berkeley.edu
[5]
Breiman, L. (1998a). Arcing classifiers (with discussion). Annals of Statistics, 26, 801-849.
[6]
Breiman, L. (1998b). Half and half bagging and hard boundary points. Technical Report 534, Statistics Dept. Univ. of Calif. at Berkeley.
[7]
Breiman, L., Friedman, J., Olshen, R., & Stone, C. (1984). Classification and Regression Trees. Chapman and Hall.
[8]
Dietterich, T. (1998). An experimental comparison of three methods for constructing ensembles of decision trees: Bagging, boosting and randomization. Machine Learning, 1-22.
[9]
Freund, Y. & Schapire, R. (1997). A decision-theoretic generalization of online learning and an application to boosting. Journal of Computer and System Sciences, 55(1), 119-139.
[10]
Freund, Y. & Schapire, R. (1996). Experiments with a new boosting algorithm. In Machine Learning: Proceedings of the Thirteenth International Conference (pp. 148-156).
[11]
Freund, Y. & Schapire, R. (in press). Discussion of "Arcing Classifiers" by L. Breiman. Annals of Statistics.
[12]
Friedman, J. (1991). Multivariate adaptive regression splines (with discussion). Annals of Statistics, 19, 1-141.
[13]
Geman, S., Bienenstock, E., & Doursat, R. (1992). Neural networks and the bias/variance dilemma. Neural Computation, 4, 1-58.
[14]
Grossamn, T. & Lapedes, A. (1993). Use of bad training data for better predictions. NIPS, 6, 343-350.

Cited By

View all
  • (2025)Ensemble deep learning techniques for time series analysis: a comprehensive review, applications, open issues, challenges, and future directionsCluster Computing10.1007/s10586-024-04684-028:1Online publication date: 1-Feb-2025
  • (2024)Condition Identification of Calcining Kiln Based on Fusion Machine Learning and Semantic WebInternational Journal on Semantic Web & Information Systems10.4018/IJSWIS.36520321:1(1-36)Online publication date: 10-Dec-2024
  • (2024)Train Once, Locate Anytime for Anyone: Adversarial Learning-based Wireless LocalizationACM Transactions on Sensor Networks10.1145/361409520:2(1-21)Online publication date: 10-Jan-2024
  • Show More Cited By

Recommendations

Comments

Please enable JavaScript to view thecomments powered by Disqus.

Information & Contributors

Information

Published In

cover image Machine Language
Machine Language  Volume 40, Issue 3
Sept. 2000
96 pages
ISSN:0885-6125
Issue’s Table of Contents

Publisher

Kluwer Academic Publishers

United States

Publication History

Published: 01 September 2000

Author Tags

  1. ensemble
  2. output variability
  3. randomization

Qualifiers

  • Article

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)0
  • Downloads (Last 6 weeks)0
Reflects downloads up to 03 Jan 2025

Other Metrics

Citations

Cited By

View all
  • (2025)Ensemble deep learning techniques for time series analysis: a comprehensive review, applications, open issues, challenges, and future directionsCluster Computing10.1007/s10586-024-04684-028:1Online publication date: 1-Feb-2025
  • (2024)Condition Identification of Calcining Kiln Based on Fusion Machine Learning and Semantic WebInternational Journal on Semantic Web & Information Systems10.4018/IJSWIS.36520321:1(1-36)Online publication date: 10-Dec-2024
  • (2024)Train Once, Locate Anytime for Anyone: Adversarial Learning-based Wireless LocalizationACM Transactions on Sensor Networks10.1145/361409520:2(1-21)Online publication date: 10-Jan-2024
  • (2023)A Survey on Deep Semi-Supervised LearningIEEE Transactions on Knowledge and Data Engineering10.1109/TKDE.2022.322021935:9(8934-8954)Online publication date: 1-Sep-2023
  • (2023)Optimum Bayesian thresholds for rebalanced classification problems using class-switching ensemblesPattern Recognition10.1016/j.patcog.2022.109158135:COnline publication date: 1-Mar-2023
  • (2023)Deep learning in food category recognitionInformation Fusion10.1016/j.inffus.2023.10185998:COnline publication date: 26-Jul-2023
  • (2022)Predicting COVID-19-Induced Lung Damage Based on Machine Learning MethodsProgramming and Computing Software10.1134/S036176882204006548:4(243-255)Online publication date: 18-Jul-2022
  • (2022)Majority-to-minority resampling for boosting-based classification under imbalanced dataApplied Intelligence10.1007/s10489-022-03585-253:4(4541-4562)Online publication date: 11-Jun-2022
  • (2022)Mutual match for semi-supervised online evolutive learningApplied Intelligence10.1007/s10489-022-03564-753:3(3336-3350)Online publication date: 28-May-2022
  • (2022)A comparison of random forest based algorithms: random credal random forest versus oblique random forestSoft Computing - A Fusion of Foundations, Methodologies and Applications10.1007/s00500-018-3628-523:21(10739-10754)Online publication date: 11-Mar-2022
  • Show More Cited By

View Options

View options

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media