April 20, 2023, 9:34 p.m. | /u/GeorgeneKeck

Machine Learning www.reddit.com

I was wondering. I know most ARM processors are used in embedded devices which are not at all used for optimization tasks. However, Aarch64 architecture is being applied to more and more multi-purpose machines. Apple M1 for example. Plus they are oft used for clustering.

Certainly, Aarch64's SIMD cannot do the same thing that some odd-400-bit a gazillion parallel jigaflops of Nvidia GPUs achive. But I was thinking, with careful encoding of the floats, or just using vector floats of …

apple architecture arm arm processors autodiff clustering devices embedded embedded devices encoding example gpus machinelearning machines nvidia nvidia gpus optimization processors thinking vector

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Data Analyst (H/F)

@ Business & Decision | Montpellier, France

Machine Learning Researcher

@ VERSES | Brighton, England, United Kingdom - Remote