all AI news
[d] Apple claims M2 Ultra "can train massive ML workloads, like large transformer models."
June 5, 2023, 8:01 p.m. | /u/jl303
Machine Learning www.reddit.com
"Finally, the 32-core Neural Engine is 40% faster. And M2 Ultra can support an enormous 192GB of unified memory, which is 50% more than M1 Ultra, enabling it to do things other chips just can't do. For example, in a single system, it can train massive ML workloads, like large transformer models that the most powerful discrete GPU can't even process because it runs out of memory."
[WWDC 2023 …
apple apple silicon chips core enabling faster machinelearning massive memory silicon support training transformer transformer models
More from www.reddit.com / Machine Learning
Jobs in AI, ML, Big Data
Founding AI Engineer, Agents
@ Occam AI | New York
AI Engineer Intern, Agents
@ Occam AI | US
AI Research Scientist
@ Vara | Berlin, Germany and Remote
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Lead GNSS Data Scientist
@ Lurra Systems | Melbourne