March 19, 2024, 4:42 a.m. | Vincent Herrmann, Francesco Faccio, J\"urgen Schmidhuber

cs.LG updates on arXiv.org arxiv.org

arXiv:2403.11998v1 Announce Type: new
Abstract: Recurrent Neural Networks (RNNs) are general-purpose parallel-sequential computers. The program of an RNN is its weight matrix. How to learn useful representations of RNN weights that facilitate RNN analysis as well as downstream tasks? While the mechanistic approach directly looks at some RNN's weights to predict its behavior, the functionalist approach analyzes its overall functionality -- specifically, its input-output mapping. We consider several mechanistic approaches for RNN weights and adapt the permutation equivariant Deep Weight …

abstract analysis arxiv computers cs.lg general how to learn learn matrix network networks neural network neural networks recurrent neural network recurrent neural networks rnn tasks type weight matrix

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne