Alquier. 2021.
“User-Friendly Introduction to PAC-Bayes Bounds.” arXiv:2110.11216 [Cs, Math, Stat].
Bates, and Granger. 1969.
“The Combination of Forecasts.” Journal of the Operational Research Society.
Bishop, and Svensen. 2012.
“Bayesian Hierarchical Mixtures of Experts.” arXiv:1212.2447 [Cs, Stat].
Breiman. 1996.
“Stacked Regressions.” Machine Learning.
Buckland, Burnham, and Augustin. 1997.
“Model Selection: An Integral Part of Inference.” Biometrics.
Card, Zhang, and Smith. 2019.
“Deep Weighted Averaging Classifiers.” In
Proceedings of the Conference on Fairness, Accountability, and Transparency.
Claeskens, and Hjort. 2008. Model Selection and Model Averaging. Cambridge Series in Statistical and Probabilistic Mathematics.
Clyde, and George. 2004.
“Model Uncertainty.” Statistical Science.
Clyde, and Iversen. 2013.
“Bayesian Model Averaging in the M-Open Framework.” In
Bayesian Theory and Applications.
Gammerman, and Vovk. 2007.
“Hedging Predictions in Machine Learning.” The Computer Journal.
Ganaie, Hu, Malik, et al. 2022.
“Ensemble Deep Learning: A Review.” Engineering Applications of Artificial Intelligence.
Hansen. 2007.
“Least Squares Model Averaging.” Econometrica.
He, Lakshminarayanan, and Teh. 2020.
“Bayesian Deep Ensembles via the Neural Tangent Kernel.” In
Advances in Neural Information Processing Systems.
Hinne, Gronau, van den Bergh, et al. 2019.
“A Conceptual Introduction to Bayesian Model Averaging.” Preprint.
Hinton, Vinyals, and Dean. 2015.
“Distilling the Knowledge in a Neural Network.” arXiv:1503.02531 [Cs, Stat].
Hjort, and Claeskens. 2003.
“Frequentist Model Average Estimators.” Journal of the American Statistical Association.
Hoeting, Madigan, Raftery, et al. 1999.
“Bayesian Model Averaging: A Tutorial.” Statistical Science.
Hu, and Zidek. 2002.
“The Weighted Likelihood.” The Canadian Journal of Statistics / La Revue Canadienne de Statistique.
Naimi, and Balzer. 2018.
“Stacked Generalization: An Introduction to Super Learning.” European Journal of Epidemiology.
Polley. 2010.
“Super Learner In Prediction.” U.C. Berkeley Division of Biostatistics Working Paper Series.
Shen, and Huang. 2006.
“Optimal Model Assessment, Selection, and Combination.” Journal of the American Statistical Association.
Shwartz-Ziv, and Armon. 2021.
“Tabular Data: Deep Learning Is Not All You Need.” arXiv:2106.03253 [Cs].
Ting, and Witten. 1999.
“Issues in Stacked Generalization.” Journal of Artificial Intelligence Research.
van der Laan, Polley, and Hubbard. 2007.
“Super Learner.” Statistical Applications in Genetics and Molecular Biology.
Wang, Xiaofang, Kondratyuk, Christiansen, et al. 2021.
“Wisdom of Committees: An Overlooked Approach To Faster and More Accurate Models.” arXiv:2012.01988 [Cs].
Wang, Haiying, Zhang, and Zou. 2009.
“Frequentist Model Averaging Estimation: A Review.” Journal of Systems Science and Complexity.
Waterhouse, MacKay, and Robinson. 1995.
“Bayesian Methods for Mixtures of Experts.” In
Advances in Neural Information Processing Systems.
Wolpert. 1992.
“Stacked Generalization.” Neural Networks.
Zhang, Tianfang, Bokrantz, and Olsson. 2021.
“A Similarity-Based Bayesian Mixture-of-Experts Model.” arXiv:2012.02130 [Cs, Stat].