Stacked Ensemble-based Learning Task

From GM-RKB
Jump to navigation Jump to search

A Stacked Ensemble-based Learning Task is an Ensemble Learning Task that combines several learning algorithms to improve prediction accuracy.



References

2020a

  • (Wikipedia, 2020) ⇒ https://en.wikipedia.org/wiki/Ensemble_learning#Stacking Retrieved:2020-1-21.
    • Stacking (sometimes called stacked generalization) involves training a learning algorithm to combine the predictions of several other learning algorithms. First, all of the other algorithms are trained using the available data, then a combiner algorithm is trained to make a final prediction using all the predictions of the other algorithms as additional inputs. If an arbitrary combiner algorithm is used, then stacking can theoretically represent any of the ensemble techniques described in this article, although, in practice, a logistic regression model is often used as the combiner.

      Stacking typically yields performance better than any single one of the trained models. [1] It has been successfully used on both supervised learning tasks (regression, [2] classification and distance learning ) and unsupervised learning (density estimation). [3] It has also been used to estimate bagging's error rate.[4] [5]. It has been reported to out-perform Bayesian model-averaging. [6] The two top-performers in the Netflix competition utilized blending, which may be considered to be a form of stacking.

  1. Wolpert, D., Stacked Generalization., Neural Networks, 5(2), pp. 241-259., 1992
  2. Breiman, L., Stacked Regression, Machine Learning, 24, 1996
  3. Smyth, P. and Wolpert, D. H., Linearly Combining Density Estimators via Stacking, Machine Learning Journal, 36, 59-83, 1999
  4. Rokach, L. (2010). “Ensemble-based classifiers". Artificial Intelligence Review. 33 (1–2): 1–39. doi:10.1007/s10462-009-9124-7
  5. Wolpert, D.H., and Macready, W.G., An Efficient Method to Estimate Bagging’s Generalization Error, Machine Learning Journal, 35, 41-55, 1999
  6. Clarke, B., Bayes model averaging and stacking when model approximation error cannot be ignored, Journal of Machine Learning Research, pp 683-712, 2003

2020b

2018

2017

2016

2013

2012

2010

2009

1999

1997

1996

1992