all AI news
Does Transformer Interpretability Transfer to RNNs?
April 10, 2024, 4:41 a.m. | Gon\c{c}alo Paulo, Thomas Marshall, Nora Belrose
cs.LG updates on arXiv.org arxiv.org
Abstract: Recent advances in recurrent neural network architectures, such as Mamba and RWKV, have enabled RNNs to match or exceed the performance of equal-size transformers in terms of language modeling perplexity and downstream evaluations, suggesting that future systems may be built on completely new architectures. In this paper, we examine if selected interpretability methods originally designed for transformer language models will transfer to these up-and-coming recurrent architectures. Specifically, we focus on steering model outputs via contrastive …
abstract advances architectures arxiv cs.ai cs.cl cs.lg future interpretability language mamba match modeling network neural network paper performance perplexity recurrent neural network rwkv systems terms transfer transformer transformers type
More from arxiv.org / cs.LG updates on arXiv.org
Jobs in AI, ML, Big Data
AI Research Scientist
@ Vara | Berlin, Germany and Remote
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Lead GNSS Data Scientist
@ Lurra Systems | Melbourne
Senior Machine Learning Engineer (MLOps)
@ Promaton | Remote, Europe
Robotics Technician - 3rd Shift
@ GXO Logistics | Perris, CA, US, 92571