all AI news
GNN-VPA: A Variance-Preserving Aggregation Strategy for Graph Neural Networks
March 8, 2024, 5:41 a.m. | Lisa Schneckenreiter, Richard Freinschlag, Florian Sestak, Johannes Brandstetter, G\"unter Klambauer, Andreas Mayr
cs.LG updates on arXiv.org arxiv.org
Abstract: Graph neural networks (GNNs), and especially message-passing neural networks, excel in various domains such as physics, drug discovery, and molecular modeling. The expressivity of GNNs with respect to their ability to discriminate non-isomorphic graphs critically depends on the functions employed for message aggregation and graph-level readout. By applying signal propagation theory, we propose a variance-preserving aggregation function (VPA) that maintains expressivity, but yields improved forward and backward dynamics. Experiments demonstrate that VPA leads to increased …
abstract aggregation arxiv cs.ai cs.lg discovery domains drug discovery excel functions gnn gnns graph graph neural networks graphs modeling networks neural networks physics stat.ml strategy type variance
More from arxiv.org / cs.LG updates on arXiv.org
Jobs in AI, ML, Big Data
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Lead GNSS Data Scientist
@ Lurra Systems | Melbourne
Senior Machine Learning Engineer (MLOps)
@ Promaton | Remote, Europe
Reporting & Data Analytics Lead (Sizewell C)
@ EDF | London, GB
Data Analyst
@ Notable | San Mateo, CA