Neural networks applied to graph data.
(Neural networks of course can already be *represented* as directed graphs, or applied to phenomena which arise from a causal graph but that is not what we mean here.
What we mean *here* is using information about graph topology as a feature input (and possibly output) for a neural network.
In practic this is usually some variant of applying convnets to spectral graph representations.

I am not closely following this area at the moment, so be aware content may not be current.

## Tutorials

Pantelis Elinas wrote a good tutorial. One of his motivating examples is nifty: He argues graph learning is powerful because it includes the fundamental problem of discovering knowledge graphs and thus research discovery. He recommends the following summaries: Bronstein et al. (2017); Bronstein et al. (2021); Hamilton (2020); Hamilton, Ying, and Leskovec (2018); Xia et al. (2021).

- Thomas Kipf’s summary.
- Xavier Besson’s implementation of one graph convnet.
- Chaitanya Joshi’s overviews of spatial graph convnets and his attempt to link them to attention mechanisms.

## Fun tweaks

Distance Encoding is a general class of graph-structure-related features that can be utilized by graph neural networks to improve the structural representation power. Given a node set whose structural representation is to be learnt, DE for a node over the graph is defined as a mapping of a set of landing probabilities of random walks from each node of the node set of interest to this node. Distance encoding generally includes measures such as shortest-path-distances and generalized PageRank scores. Distance encoding can be merged into the design of graph neural networks in simple but effective ways: First, we propose DEGNN that utilizes distance encoding as extra node features. We further enhance DEGNN by allowing distance encoding to control the aggregation procedure of traditional GNNs, which yields another model DEAGNN. Since distance encoding purely depends on the graph structure and is independent from node identifiers, it has inductive and generalization ability.

## Tooling

Build your models with PyTorch, TensorFlow or Apache MXNet.

Fast and memory-efficient message passing primitives for training Graph Neural Networks. Scale to giant graphs via multi-GPU acceleration and distributed training infrastructure.

Facebook’s GTN:

GTN is an open source framework for automatic differentiation with a powerful, expressive type of graph called weighted finite-state transducers (WFSTs). Just as PyTorch provides a framework for automatic differentiation with tensors, GTN provides such a framework for WFSTs. AI researchers and engineers can use GTN to more effectively train graph-based machine learning models.

I have not used GTN so I cannot say if I have field it correctly or if it is more of a computational graph learning tool.

## Background: Graph filtering

A lot of this seems to be based upon more classic linear systems theory applied to networks in the form of spectral graph theory. See signal processing on graphs.

## References

*arXiv:2104.13478 [Cs, Stat]*, May.

*IEEE Signal Processing Magazine*34 (4): 18–42.

*arXiv:1703.04818 [Cs]*, March.

*Machine Learning and the Physical Sciences Workshop at the 33rd Conference on Neural Information Processing Systems (NeurIPS)*, 6.

*Advances In Neural Information Processing Systems*.

*arXiv:2012.15000 [Cs, Stat]*, December.

*arXiv:1904.05146 [Cs, Stat]*, April.

*arXiv:2003.00982 [Cs, Stat]*, July.

*Synthesis Lectures on Artificial Intelligence and Machine Learning*14 (3): 1–159.

*arXiv:1709.05584 [Cs]*, April.

*arXiv:2010.01003 [Cs, Stat]*, October.

*arXiv:2010.13993 [Cs]*, November.

*IEEE Transactions on Signal Processing*65 (2): 274–88.

*IJCAI 2020*.

*arXiv:2009.00142 [Cs, Stat]*, October.

*arXiv:1910.08527 [Cs, Stat]*, February.

*Advances In Neural Information Processing Systems*.

*Machine Learning and the Physical Sciences Workshop at the 33rd Conference on Neural Information Processing Systems (NeurIPS)*, 11.

*IEEE Signal Processing Magazine*30 (3): 83–98.

*2011 International Conference on Distributed Computing in Sensor Systems and Workshops (DCOSS)*, June, 1–8.

*IEEE Transactions on Artificial Intelligence*2 (2): 109–27.

*arXiv:1703.06114 [Cs, Stat]*, April.

## No comments yet. Why not leave one?