Paper ID: 2502.06751 • Published Feb 10, 2025
What makes a good feedforward computational graph?
Alex Vitvitskyi, João G. M. Araújo, Marc Lackenby, Petar Veličković
TL;DR
Get AI-generated summaries with premium
Get AI-generated summaries with premium
As implied by the plethora of literature on graph rewiring, the choice of
computational graph employed by a neural network can make a significant impact
on its downstream performance. Certain effects related to the computational
graph, such as under-reaching and over-squashing, may even render the model
incapable of learning certain functions. Most of these effects have only been
thoroughly studied in the domain of undirected graphs; however, recent years
have seen a significant rise in interest in feedforward computational graphs:
directed graphs without any back edges. In this paper, we study the desirable
properties of a feedforward computational graph, discovering two important
complementary measures: fidelity and mixing time, and evaluating a few popular
choices of graphs through the lens of these measures. Our study is backed by
both theoretical analyses of the metrics' asymptotic behaviour for various
graphs, as well as correlating these metrics to the performance of trained
neural network models using the corresponding graphs.