This document discusses different neural network methods for processing graph-structured data. It begins by describing recurrent neural networks (RNNs) and their limitations for graphs, such as an inability to handle undirected or cyclic graphs. It then summarizes two alternative approaches: one that uses contraction maps to allow recurrent updates on arbitrary graphs, and one that employs a constructive architecture with frozen neurons to avoid issues with cycles. Both methods aim to make predictions at the node or graph level on relational data like molecules or web pages.
Related topics: