Oct. 10, 2022, 1:11 a.m. | Andrew J. Nam, Mustafa Abdool, Trevor Maxfield, James L. McClelland

cs.LG updates on arXiv.org arxiv.org

Out-of-distribution generalization (OODG) is a longstanding challenge for
neural networks, and is quite apparent in tasks with well-defined variables and
rules, where explicit use of the rules can solve problems independently of the
particular values of the variables. Large transformer-based language models
have pushed the boundaries on how well neural networks can generalize to novel
inputs, but their complexity obfuscates they achieve such robustness. As a step
toward understanding how transformer-based systems generalize, we explore the
question of OODG in …

arxiv curriculum curriculum learning distribution reasoning

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Data Scientist (Database Development)

@ Nasdaq | Bengaluru-Affluence