A sum or product (or outer sum, or tensor product) of Mercer kernels is still a kernel. For other operations YMMV; let us look at some.
Since talking about kernels in the abstract is confusing, I like to consider a (slightly) more concrete mental model: Which kernels are valid covariance functions for a Gaussian process? This turns out to be a useful way to think about it because it suggests some ways to combine kernels.
1 Operations preserving positivity
For example, in the case of Gaussian processes, suppose that, independently,
More generally, if
Note that the additivity in terms of kernels is not the same as additivity in terms of induced feature spaces. The induced feature map of
If
which apparently is now called a deep kernel. If
Also if
Genton (2001) uses the interpretation as covariance to construct some other useful combinations.
Let
we find that the following is a kernel
All these to various cunning combination strategies, which I may return to discuss. 🏗 Some of them are in the references. For example Duvenaud et al. (2013) position their work in the wider field:
There is a large body of work attempting to construct a rich kernel through a weighted sum of base kernels (e.g. Bach 2008; Christoudias, Urtasun, and Darrell 2009). While these approaches find the optimal solution in polynomial time, speed comes at a cost: the component kernels, as well as their hyperparameters, must be specified in advance […]
Hinton and Salakhutdinov (2008) use a deep neural network to learn an embedding; this is a flexible approach to kernel learning but relies upon finding structure in the input density,
. Instead we focus on domains where most of the interesting structure is in . Wilson and Adams (2013) derive kernels of the form $ (x − x_0 $), forming a basis for stationary kernels. These kernels share similarities with
but can express negative prior correlation, and could usefully be included in our grammar.
See Grosse et al. (2012) for a mind-melting compositional matrix factorisation diagram, constructing a search over hierarchical kernel decompositions.
Examples of existing machine learning models which fall under our framework. Arrows represent models reachable using a single production rule. Only a small fraction of the 2496 models reachable within 3 steps are shown, and not all possible arrows are shown.
Here is another list of kernel combinations that preserve positive-definiteness, but for which I have tragically lost the source (it is one of the papers here).
2 Applying kernels over different axes
I silently assumed this works because it seems obvious sometimes; but then sometimes I forget this trick, so maybe it is not obvious:
For multi-dimensional inputs we can apply kernels over any subset of the axes, and thence different kernels over different subsets of the axes.
3 Locally stationary kernels
Genton (2001) defines locally stationary kernels to mean those that have a particular structure, specifically, a kernel that can be factored into a stationary kernel
Local structure then depends on the global mean location
Other constructions might vie for the title of “locally stationary”. To check. 🏗
4 Stationary reducible kernels
See kernel warping.