all AI news
Transformers discover an elementary calculation system exploiting local attention and grid-like problem representation. (arXiv:2207.02536v1 [cs.LG])
July 7, 2022, 1:10 a.m. | Samuel Cognolato, Alberto Testolin
cs.LG updates on arXiv.org arxiv.org
Mathematical reasoning is one of the most impressive achievements of human
intellect but remains a formidable challenge for artificial intelligence
systems. In this work we explore whether modern deep learning architectures can
learn to solve a symbolic addition task by discovering effective arithmetic
procedures. Although the problem might seem trivial at first glance,
generalizing arithmetic knowledge to operations involving a higher number of
terms, possibly composed by longer sequences of digits, has proven extremely
challenging for neural networks. Here we …
arxiv attention elementary lg local attention representation transformers
More from arxiv.org / cs.LG updates on arXiv.org
Jobs in AI, ML, Big Data
Software Engineer for AI Training Data (School Specific)
@ G2i Inc | Remote
Software Engineer for AI Training Data (Python)
@ G2i Inc | Remote
Software Engineer for AI Training Data (Tier 2)
@ G2i Inc | Remote
Data Engineer
@ Lemon.io | Remote: Europe, LATAM, Canada, UK, Asia, Oceania
Artificial Intelligence – Bioinformatic Expert
@ University of Texas Medical Branch | Galveston, TX
Lead Developer (AI)
@ Cere Network | San Francisco, US