Sept. 18, 2023, 2:02 p.m. | /u/30299578815310

Machine Learning www.reddit.com

Recent work shows transformers are capable of performing multi-step gradient descent of mesa objectives inside of their transformer layers. This is even possible for linear transformers, which effectively perform linear optimization on deep representations of features calculated by earlier layers.

[https://arxiv.org/pdf/2309.05858.pdf](https://arxiv.org/pdf/2309.05858.pdf)

For those unfamiliar, [instrumental convergence](https://en.wikipedia.org/wiki/Instrumental_convergence) is the idea that entities with different goals will tend towards different subgoals. Examples could include gathering power, not dying, acquiring resources, etc. A famous thought experiment, known as the paperclip maximizer, is the …

convergence features gradient inside linear machinelearning mesa modern optimization shows think transformer transformers work

Data Engineer

@ Lemon.io | Remote: Europe, LATAM, Canada, UK, Asia, Oceania

Artificial Intelligence – Bioinformatic Expert

@ University of Texas Medical Branch | Galveston, TX

Lead Developer (AI)

@ Cere Network | San Francisco, US

Research Engineer

@ Allora Labs | Remote

Ecosystem Manager

@ Allora Labs | Remote

Founding AI Engineer, Agents

@ Occam AI | New York