Banerjee, Merugu, Dhillon, et al. 2005. “Clustering with Bregman Divergences.” Journal of Machine Learning Research.
Buja, Stuetzle, and Shen. 2005. “Loss Functions for Binary Class Probability Estimation and Classification: Structure and Applications.”
Collins, Dasgupta, and Schapire. 2001.
“A Generalization of Principal Components Analysis to the Exponential Family.” In
Advances in Neural Information Processing Systems.
Gneiting, and Raftery. 2007.
“Strictly Proper Scoring Rules, Prediction, and Estimation.” Journal of the American Statistical Association.
Goldstein, Osher, Goldstein, et al. 2009.
“The Split Bregman Method for L1-Regularized Problems.” SIAM Journal on Imaging Sciences.
Gutmann, and Hirayama. 2011.
“Bregman Divergence as General Framework to Estimate Unnormalized Statistical Models.” In
Proceedings of the Twenty-Seventh Conference on Uncertainty in Artificial Intelligence. UAI’11.
Harremoës. 2015.
“Proper Scoring and Sufficiency.” arXiv:1507.07089 [Math, Stat].
Li, Schwab, Antholzer, et al. 2020.
“NETT: Solving Inverse Problems with Deep Neural Networks.” Inverse Problems.
Menon, and Ong. 2016.
“Linking Losses for Density Ratio and Class-Probability Estimation.” In
Proceedings of The 33rd International Conference on Machine Learning.
Nielsen. 2018.
“An Elementary Introduction to Information Geometry.” arXiv:1808.08271 [Cs, Math, Stat].
Nock, Menon, and Ong. 2016.
“A Scaled Bregman Theorem with Applications.” arXiv:1607.00360 [Cs, Stat].
Reid, and Williamson. 2010.
“Composite Binary Losses.” Journal of Machine Learning Research.
Singh, and Gordon. 2008.
“A Unified View of Matrix Factorization Models.” In
Machine Learning and Knowledge Discovery in Databases.
Sra, and Dhillon. 2006.
“Generalized Nonnegative Matrix Approximations with Bregman Divergences.” In
Advances in Neural Information Processing Systems 18.
Wibisono, Wilson, and Jordan. 2016.
“A Variational Perspective on Accelerated Methods in Optimization.” Proceedings of the National Academy of Sciences.