all AI news
GNN-VPA: A Variance-Preserving Aggregation Strategy for Graph Neural Networks
March 8, 2024, 5:41 a.m. | Lisa Schneckenreiter, Richard Freinschlag, Florian Sestak, Johannes Brandstetter, G\"unter Klambauer, Andreas Mayr
cs.LG updates on arXiv.org arxiv.org
Abstract: Graph neural networks (GNNs), and especially message-passing neural networks, excel in various domains such as physics, drug discovery, and molecular modeling. The expressivity of GNNs with respect to their ability to discriminate non-isomorphic graphs critically depends on the functions employed for message aggregation and graph-level readout. By applying signal propagation theory, we propose a variance-preserving aggregation function (VPA) that maintains expressivity, but yields improved forward and backward dynamics. Experiments demonstrate that VPA leads to increased …
abstract aggregation arxiv cs.ai cs.lg discovery domains drug discovery excel functions gnn gnns graph graph neural networks graphs modeling networks neural networks physics stat.ml strategy type variance
More from arxiv.org / cs.LG updates on arXiv.org
Jobs in AI, ML, Big Data
Software Engineer for AI Training Data (School Specific)
@ G2i Inc | Remote
Software Engineer for AI Training Data (Python)
@ G2i Inc | Remote
Software Engineer for AI Training Data (Tier 2)
@ G2i Inc | Remote
Data Engineer
@ Lemon.io | Remote: Europe, LATAM, Canada, UK, Asia, Oceania
Artificial Intelligence – Bioinformatic Expert
@ University of Texas Medical Branch | Galveston, TX
Lead Developer (AI)
@ Cere Network | San Francisco, US