July 28, 2022, 1:10 a.m. | Lucie Charlotte Magister, Pietro Barbiero, Dmitry Kazhdan, Federico Siciliano, Gabriele Ciravegna, Fabrizio Silvestri, Pietro Lio, Mateja Jamnik

cs.LG updates on arXiv.org arxiv.org

The opaque reasoning of Graph Neural Networks induces a lack of human trust.
Existing graph network explainers attempt to address this issue by providing
post-hoc explanations, however, they fail to make the model itself more
interpretable. To fill this gap, we introduce the Concept Encoder Module, the
first differentiable concept-discovery approach for graph networks. The
proposed approach makes graph networks explainable by design by first
discovering graph concepts and then using these to solve the task. Our results
demonstrate that …

arxiv encoding graph graph neural networks lg networks neural networks

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Senior AI & Data Engineer

@ Bertelsmann | Kuala Lumpur, 14, MY, 50400

Analytics Engineer

@ Reverse Tech | Philippines - Remote