Nov. 9, 2022, 2:15 a.m. | François Charton

cs.CL updates on arXiv.org arxiv.org

Transformers can learn to perform numerical computations from examples only.
I study nine problems of linear algebra, from basic matrix operations to
eigenvalue decomposition and inversion, and introduce and discuss four encoding
schemes to represent real numbers. On all problems, transformers trained on
sets of random matrices achieve high accuracies (over 90%). The models are
robust to noise, and can generalize out of their training distribution. In
particular, models trained to predict Laplace-distributed eigenvalues
generalize to different classes of matrices: …

algebra arxiv linear linear algebra transformers

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Enterprise AI Architect

@ Oracle | Broomfield, CO, United States

Cloud Data Engineer France H/F (CDI - Confirmé)

@ Talan | Nantes, France