Feb. 19, 2024, 1:23 p.m. | /u/bobfrutt

Artificial Intelligence www.reddit.com

I don't know much about inner workings of AI but I know that key components are neural networks, backpropagation, gradient descent and transformers. And apparently all that we figured out throughout the years and now we just using it on massive scale thanks to finally having computing power with all the GPUs available. So in that sense we know what's going on. But Eliezer talks like these systems are some kind of black box? How should we understand that exactly?

ai systems artificial backpropagation components eliezer eliezer yudkowsky gradient inside key mean networks neural networks systems transformers yudkowsky

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne