all AI news
Simulating Weighted Automata over Sequences and Trees with Transformers
March 18, 2024, 4:47 a.m. | Michael Rizvi, Maude Lizaire, Clara Lacroce, Guillaume Rabusseau
cs.CL updates on arXiv.org arxiv.org
Abstract: Transformers are ubiquitous models in the natural language processing (NLP) community and have shown impressive empirical successes in the past few years. However, little is understood about how they reason and the limits of their computational capabilities. These models do not process data sequentially, and yet outperform sequential neural models such as RNNs. Recent work has shown that these models can compactly simulate the sequential reasoning abilities of deterministic finite automata (DFAs). This leads to …
abstract arxiv capabilities community computational cs.ai cs.cc cs.cl data however language language processing natural natural language natural language processing nlp process processing reason transformers trees type
More from arxiv.org / cs.CL updates on arXiv.org
Jobs in AI, ML, Big Data
Founding AI Engineer, Agents
@ Occam AI | New York
AI Engineer Intern, Agents
@ Occam AI | US
AI Research Scientist
@ Vara | Berlin, Germany and Remote
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Data Engineer
@ Kaseya | Bengaluru, Karnataka, India