Agarwal, and Iii. 2011.
“Generative Kernels for Exponential Families.” In
Proceedings of the Fourteenth International Conference on Artificial Intelligence and Statistics.
Agrawal, Trippe, Huggins, et al. 2019.
“The Kernel Interaction Trick: Fast Bayesian Discovery of Pairwise Interactions in High Dimensions.” In
Proceedings of the 36th International Conference on Machine Learning.
Alaoui, and Mahoney. 2014.
“Fast Randomized Kernel Methods With Statistical Guarantees.” arXiv:1411.0306 [Cs, Stat].
Altun, Smola, and Hofmann. 2004.
“Exponential Families for Conditional Random Fields.” In
Proceedings of the 20th Conference on Uncertainty in Artificial Intelligence. UAI ’04.
Álvarez, Rosasco, and Lawrence. 2012.
“Kernels for Vector-Valued Functions: A Review.” Foundations and Trends® in Machine Learning.
Arbel, Korba, Salim, et al. 2019.
“Maximum Mean Discrepancy Gradient Flow.” In
Proceedings of the 33rd International Conference on Neural Information Processing Systems.
Aronszajn. 1950.
“Theory of Reproducing Kernels.” Transactions of the American Mathematical Society.
Bach, Francis. 2008.
“Exploring Large Feature Spaces with Hierarchical Multiple Kernel Learning.” In
Proceedings of the 21st International Conference on Neural Information Processing Systems. NIPS’08.
Bakır, Zien, and Tsuda. 2004.
“Learning to Find Graph Pre-Images.” In
Pattern Recognition. Lecture Notes in Computer Science 3175.
Balog, Lakshminarayanan, Ghahramani, et al. 2016.
“The Mondrian Kernel.” arXiv:1606.05241 [Stat].
Bosq, and Blanke. 2007. Inference and prediction in large dimensions. Wiley series in probability and statistics.
Boyer, Chambolle, De Castro, et al. 2018.
“On Representer Theorems and Convex Regularization.” arXiv:1806.09810 [Cs, Math].
Burges. 1998.
“Geometry and Invariance in Kernel Based Methods.” In
Advances in Kernel Methods - Support Vector Learning.
Canu, and Smola. 2006.
“Kernel Methods and the Exponential Family.” Neurocomputing.
Carrasco, Oncina, and Calera-Rubio. 2001.
“Stochastic Inference of Regular Tree Languages.” Machine Learning.
Cawley, and Talbot. 2005.
“A Simple Trick for Constructing Bayesian Formulations of Sparse Kernel Learning Methods.” In
Proceedings. 2005 IEEE International Joint Conference on Neural Networks, 2005.
Chwialkowski, Strathmann, and Gretton. 2016.
“A Kernel Test of Goodness of Fit.” In
Proceedings of the 33rd International Conference on International Conference on Machine Learning - Volume 48. ICML’16.
Clark, Florêncio, and Watkins. 2006.
“Languages as Hyperplanes: Grammatical Inference with String Kernels.” In
Machine Learning: ECML 2006. Lecture Notes in Computer Science 4212.
Clark, Florêncio, Watkins, et al. 2006.
“Planar Languages and Learnability.” In
Grammatical Inference: Algorithms and Applications. Lecture Notes in Computer Science 4201.
Collins, and Duffy. 2002.
“Convolution Kernels for Natural Language.” In
Advances in Neural Information Processing Systems 14.
Cortes, Haffner, and Mohri. 2004.
“Rational Kernels: Theory and Algorithms.” Journal of Machine Learning Research.
Cucker, and Smale. 2002.
“On the Mathematical Foundations of Learning.” Bulletin of the American Mathematical Society.
Cunningham, Shenoy, and Sahani. 2008.
“Fast Gaussian Process Methods for Point Process Intensity Estimation.” In
Proceedings of the 25th International Conference on Machine Learning. ICML ’08.
Curtain. 1975.
“Infinite-Dimensional Filtering.” SIAM Journal on Control.
Danafar, Fukumizu, and Gomez. 2014.
“Kernel-Based Information Criterion.” arXiv:1408.5810 [Stat].
Duvenaud, Lloyd, Grosse, et al. 2013.
“Structure Discovery in Nonparametric Regression Through Compositional Kernel Search.” In
Proceedings of the 30th International Conference on Machine Learning (ICML-13).
Evgeniou, Micchelli, and Pontil. 2005.
“Learning Multiple Tasks with Kernel Methods.” Journal of Machine Learning Research.
FitzGerald, Liukus, Rafii, et al. 2013.
“Harmonic/Percussive Separation Using Kernel Additive Modelling.” In
Irish Signals & Systems Conference 2014 and 2014 China-Ireland International Conference on Information and Communications Technologies (ISSC 2014/CIICT 2014). 25th IET.
Flaxman, Teh, and Sejdinovic. 2016.
“Poisson Intensity Estimation with Reproducing Kernels.” arXiv:1610.08623 [Stat].
Friedlander, Kailath, and Ljung. 1975.
“Scattering Theory and Linear Least Squares Estimation: Part II: Discrete-Time Problems.” In
1975 IEEE Conference on Decision and Control Including the 14th Symposium on Adaptive Processes.
Gorham, Raj, and Mackey. 2020.
“Stochastic Stein Discrepancies.” arXiv:2007.02857 [Cs, Math, Stat].
Grauman, and Darrell. 2005.
“The Pyramid Match Kernel: Discriminative Classification with Sets of Image Features.” In
Tenth IEEE International Conference on Computer Vision, 2005. ICCV 2005.
Greengard, and Strain. 1991.
“The Fast Gauss Transform.” SIAM Journal on Scientific and Statistical Computing.
Gretton, Borgwardt, Rasch, et al. 2012.
“A Kernel Two-Sample Test.” The Journal of Machine Learning Research.
Gretton, Fukumizu, Teo, et al. 2008.
“A Kernel Statistical Test of Independence.” In
Advances in Neural Information Processing Systems 20: Proceedings of the 2007 Conference.
Grosse, Salakhutdinov, Freeman, et al. 2012.
“Exploiting Compositionality to Explore a Large Space of Model Structures.” In
Proceedings of the Conference on Uncertainty in Artificial Intelligence.
Grünewälder, Gretton, and Shawe-Taylor. 2013.
“Smooth Operators.” In
Proceedings of the 30th International Conference on International Conference on Machine Learning - Volume 28. ICML’13.
Györfi, ed. 2002. A Distribution-Free Theory of Nonparametric Regression. Springer Series in Statistics.
Hofmann, Schölkopf, and Smola. 2008.
“Kernel Methods in Machine Learning.” The Annals of Statistics.
Ishikawa, Fujii, Ikeda, et al. 2018.
“Metric on Nonlinear Dynamical Systems with Perron-Frobenius Operators.” arXiv:1805.12324 [Cs, Math, Stat].
Jain. 2009. “Structure Spaces.” Journal of Machine Learning Research.
Jung. 2013.
“An RKHS Approach to Estimation with Sparsity Constraints.” In
Advances in Neural Information Processing Systems 29.
Kailath, Thomas. 1971. “The Structure of Radon-Nikodym Derivatives with Respect to Wiener and Related Measures.” The Annals of Mathematical Statistics.
———. 1971b.
“A Note on Least-Squares Estimation by the Innovations Method.” In
1971 IEEE Conference on Decision and Control.
———. 1974.
“A View of Three Decades of Linear Filtering Theory.” IEEE Transactions on Information Theory.
Kailath, T., Geesey, and Weinert. 1972.
“Some Relations Among RKHS Norms, Fredholm Equations, and Innovations Representations.” IEEE Transactions on Information Theory.
Kanagawa, and Fukumizu. 2014.
“Recovering Distributions from Gaussian RKHS Embeddings.” In
Journal of Machine Learning Research.
Kemerait, and Childers. 1972.
“Signal Detection and Extraction by Cepstrum Techniques.” IEEE Transactions on Information Theory.
Keriven, Bourrier, Gribonval, et al. 2016.
“Sketching for Large-Scale Learning of Mixture Models.” In
2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
Kiraly, and Oberhauser. 2019.
“Kernels for Sequentially Ordered Data.” Journal of Machine Learning Research.
Kloft, Rückert, and Bartlett. 2010.
“A Unifying View of Multiple Kernel Learning.” In
Machine Learning and Knowledge Discovery in Databases. Lecture Notes in Computer Science.
Klus, Bittracher, Schuster, et al. 2018.
“A Kernel-Based Approach to Molecular Conformation Analysis.” The Journal of Chemical Physics.
Kontorovich, Leonid, Cortes, and Mohri. 2006.
“Learning Linearly Separable Languages.” In
Algorithmic Learning Theory. Lecture Notes in Computer Science 4264.
Kontorovich, Leonid (Aryeh), Cortes, and Mohri. 2008.
“Kernel Methods for Learning Languages.” Theoretical Computer Science, Algorithmic Learning Theory,.
Kulis, and Grauman. 2012.
“Kernelized Locality-Sensitive Hashing.” IEEE Transactions on Pattern Analysis and Machine Intelligence.
Lawrence, Seeger, and Herbrich. 2003.
“Fast Sparse Gaussian Process Methods: The Informative Vector Machine.” In
Proceedings of the 16th Annual Conference on Neural Information Processing Systems.
Liu, Qiang, Lee, and Jordan. 2016.
“A Kernelized Stein Discrepancy for Goodness-of-Fit Tests.” In
Proceedings of The 33rd International Conference on Machine Learning.
Ljung, Kailath, and Friedlander. 1975.
“Scattering Theory and Linear Least Squares Estimation: Part I: Continuous-Time Problems.” In
1975 IEEE Conference on Decision and Control Including the 14th Symposium on Adaptive Processes.
Lloyd, Duvenaud, Grosse, et al. 2014.
“Automatic Construction and Natural-Language Description of Nonparametric Regression Models.” In
Twenty-Eighth AAAI Conference on Artificial Intelligence.
Lodhi, Saunders, Shawe-Taylor, et al. 2002.
“Text Classification Using String Kernels.” Journal of Machine Learning Research.
Lopez-Paz, Nishihara, Chintala, et al. 2016.
“Discovering Causal Signals in Images.” arXiv:1605.08179 [Cs, Stat].
Lu, Leen, Huang, et al. 2008.
“A Reproducing Kernel Hilbert Space Framework for Pairwise Time Series Distances.” In
Proceedings of the 25th International Conference on Machine Learning. ICML ’08.
Ma, Wan-Duo Kurt, Lewis, and Kleijn. 2020.
“The HSIC Bottleneck: Deep Learning Without Back-Propagation.” Proceedings of the AAAI Conference on Artificial Intelligence.
Manton, and Amblard. 2015.
“A Primer on Reproducing Kernel Hilbert Spaces.” Foundations and Trends® in Signal Processing.
Marteau-Ferey, Bach, and Rudi. 2019.
“Globally Convergent Newton Methods for Ill-Conditioned Generalized Self-Concordant Losses.” In
Advances in Neural Information Processing Systems.
———. 2020.
“Non-Parametric Models for Non-Negative Functions.” In
Proceedings of the 34th International Conference on Neural Information Processing Systems. NIPS ’20.
McFee, and Ellis. 2011.
“Analyzing Song Structure with Spectral Clustering.” In
IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
Meanti, Carratino, Rosasco, et al. 2020.
“Kernel Methods Through the Roof: Handling Billions of Points Efficiently.” In
Proceedings of the 34th International Conference on Neural Information Processing Systems. NIPS’20.
Mercer. 1909.
“Functions of Positive and Negative Type, and Their Connection with the Theory of Integral Equations.” Philosophical Transactions of the Royal Society of London. Series A, Containing Papers of a Mathematical or Physical Character.
Micchelli, and Pontil. 2005a.
“Learning the Kernel Function via Regularization.” Journal of Machine Learning Research.
———. 2005b.
“On Learning Vector-Valued Functions.” Neural Computation.
Muandet, Fukumizu, Sriperumbudur, et al. 2014.
“Kernel Mean Shrinkage Estimators.” arXiv:1405.5505 [Cs, Stat].
Muandet, Fukumizu, Sriperumbudur, et al. 2017.
“Kernel Mean Embedding of Distributions: A Review and Beyond.” Foundations and Trends® in Machine Learning.
Muller, Mika, Ratsch, et al. 2001.
“An Introduction to Kernel-Based Learning Algorithms.” IEEE Transactions on Neural Networks.
Nishiyama, and Fukumizu. 2016.
“Characteristic Kernels and Infinitely Divisible Distributions.” The Journal of Machine Learning Research.
Parzen, Emanuel. 1963. “Probability Density Functionals and Reproducing Kernel Hilbert Spaces.” In Proceedings of the Symposium on Time Series Analysis.
Poggio, and Girosi. 1990.
“Networks for Approximation and Learning.” Proceedings of the IEEE.
Rahimi, and Recht. 2007.
“Random Features for Large-Scale Kernel Machines.” In
Advances in Neural Information Processing Systems.
Rudi, Carratino, and Rosasco. 2017.
“FALKON: An Optimal Large Scale Kernel Method.” In
Proceedings of the 31st International Conference on Neural Information Processing Systems. NIPS’17.
Rue, and Held. 2005.
Gaussian Markov Random Fields: Theory and Applications. Monographs on Statistics and Applied Probability 104.
Saha, and Balamurugan. 2020.
“Learning with Operator-Valued Kernels in Reproducing Kernel Krein Spaces.” In
Advances in Neural Information Processing Systems.
Salvi, Cass, Foster, et al. 2021.
“The Signature Kernel Is the Solution of a Goursat PDE.” SIAM Journal on Mathematics of Data Science.
Salvi, Lemercier, Liu, et al. 2024.
“Higher Order Kernel Mean Embeddings to Capture Filtrations of Stochastic Processes.” In
Advances in Neural Information Processing Systems. NIPS ’21.
Särkkä. 2011.
“Linear Operators and Stochastic Partial Differential Equations in Gaussian Process Regression.” In
Artificial Neural Networks and Machine Learning – ICANN 2011. Lecture Notes in Computer Science.
Schölkopf, Herbrich, and Smola. 2001.
“A Generalized Representer Theorem.” In
Computational Learning Theory. Lecture Notes in Computer Science.
Schölkopf, Mika, Burges, et al. 1999. “Input Space Versus Feature Space in Kernel-Based Methods.” IEEE Transactions on Neural Networks.
Schölkopf, and Smola. 2002. Learning with Kernels: Support Vector Machines, Regularization, Optimization, and Beyond.
———. 2003.
“A Short Introduction to Learning with Kernels.” In
Advanced Lectures on Machine Learning. Lecture Notes in Computer Science 2600.
Schölkopf, Smola, and Müller. 1997.
“Kernel Principal Component Analysis.” In
Artificial Neural Networks — ICANN’97. Lecture Notes in Computer Science.
Schuster, Mollenhauer, Klus, et al. 2019.
“Kernel Conditional Density Operators.” arXiv:1905.11255 [Cs, Math, Stat].
Schuster, Strathmann, Paige, et al. 2017.
“Kernel Sequential Monte Carlo.” In
ECML-PKDD 2017.
Segall, Davis, and Kailath. 1975.
“Nonlinear Filtering with Counting Observations.” IEEE Transactions on Information Theory.
Segall, and Kailath. 1976.
“Orthogonal Functionals of Independent-Increment Processes.” IEEE Transactions on Information Theory.
———. 2004.
“A Tutorial on Support Vector Regression.” Statistics and Computing.
Snelson, and Ghahramani. 2005.
“Sparse Gaussian Processes Using Pseudo-Inputs.” In
Advances in Neural Information Processing Systems.
Song, Gretton, Bickson, et al. 2011.
“Kernel Belief Propagation.” In
Proceedings of the Fourteenth International Conference on Artificial Intelligence and Statistics.
Sriperumbudur, Gretton, Fukumizu, et al. 2008.
“Injective Hilbert Space Embeddings of Probability Measures.” In
Proceedings of the 21st Annual Conference on Learning Theory (COLT 2008).
Székely, and Rizzo. 2009.
“Brownian Distance Covariance.” The Annals of Applied Statistics.
Székely, Rizzo, and Bakirov. 2007.
“Measuring and Testing Dependence by Correlation of Distances.” The Annals of Statistics.
Tipping, and Nh. 2001.
“Sparse Kernel Principal Component Analysis.” In
Advances in Neural Information Processing Systems 13.
Tompkins, and Ramos. 2018.
“Fourier Feature Approximations for Periodic Kernels in Time-Series Modelling.” Proceedings of the AAAI Conference on Artificial Intelligence.
Vedaldi, and Zisserman. 2012.
“Efficient Additive Kernels via Explicit Feature Maps.” IEEE Transactions on Pattern Analysis and Machine Intelligence.
Vert, Tsuda, and Schölkopf. 2004.
“A Primer on Kernel Methods.” In
Kernel Methods in Computational Biology.
Vishwanathan, Schraudolph, Kondor, et al. 2010.
“Graph Kernels.” Journal of Machine Learning Research.
Walder, Christian, Kim, and Schölkopf. 2008.
“Sparse Multiscale Gaussian Process Regression.” In
Proceedings of the 25th International Conference on Machine Learning. ICML ’08.
Wang, Smola, and Tibshirani. 2014.
“The Falling Factorial Basis and Its Statistical Applications.” In
Proceedings of the 31st International Conference on International Conference on Machine Learning - Volume 32. ICML’14.
Weinert, Howard L. 1978.
“Statistical Methods in Optimal Curve Fitting.” Communications in Statistics - Simulation and Computation.
Williams. 2001.
“On a Connection Between Kernel PCA and Metric Multidimensional Scaling.” In
Advances in Neural Information Processing Systems 13.
Wilson, and Adams. 2013.
“Gaussian Process Kernels for Pattern Discovery and Extrapolation.” In
International Conference on Machine Learning.
Wilson, Dann, Lucas, et al. 2015.
“The Human Kernel.” arXiv:1510.07389 [Cs, Stat].
Wu, and Zhou. 2008.
“Learning with Sample Dependent Hypothesis Spaces.” Computers & Mathematics with Applications.
Xu, Wenkai, and Matsuda. 2020.
“A Stein Goodness-of-Fit Test for Directional Distributions.” In
International Conference on Artificial Intelligence and Statistics.
Xu, Jian-Wu, Paiva, Park, et al. 2008.
“A Reproducing Kernel Hilbert Space Framework for Information-Theoretic Learning.” IEEE Transactions on Signal Processing.
Yaglom. 1987a. Correlation Theory of Stationary and Related Random Functions. Volume II: Supplementary Notes and References. Springer Series in Statistics.
———. 1987b. Correlation Theory of Stationary and Related Random Functions Volume I.
———. 2004. An Introduction to the Theory of Stationary Random Functions.
Yang, Changjiang, Duraiswami, and Davis. 2004.
“Efficient Kernel Machines Using the Improved Fast Gauss Transform.” In
Advances in Neural Information Processing Systems.
Yang, Changjiang, Duraiswami, Gumerov, et al. 2003.
“Improved Fast Gauss Transform and Efficient Kernel Density Estimation.” In
Proceedings of the Ninth IEEE International Conference on Computer Vision - Volume 2. ICCV ’03.
Yang, Tianbao, Li, Mahdavi, et al. 2012.
“Nyström Method Vs Random Fourier Features: A Theoretical and Empirical Comparison.” In
Advances in Neural Information Processing Systems.
Yang, Jiyan, Sindhwani, Avron, et al. 2014.
“Quasi-Monte Carlo Feature Maps for Shift-Invariant Kernels.” arXiv:1412.8293 [Cs, Math, Stat].
Yu, Cheng, Schuurmans, et al. 2013.
“Characterizing the Representer Theorem.” In
Proceedings of the 30th International Conference on Machine Learning (ICML-13).
Zhang, Qinyi, Filippi, Gretton, et al. 2016.
“Large-Scale Kernel Methods for Independence Testing.” arXiv:1606.07892 [Stat].
Zhou, Zha, and Song. 2013.
“Learning Triggering Kernels for Multi-Dimensional Hawkes Processes.” In
Proceedings of the 30th International Conference on Machine Learning (ICML-13).