April 28, 2024, 9:59 a.m. | /u/Agitated_Space_672

Machine Learning www.reddit.com

[https://arxiv.org/abs/2404.15758](https://arxiv.org/abs/2404.15758)

# From the abstract


We show that transformers can use meaningless filler tokens (e.g., '......') in place of a chain of thought to solve two hard algorithmic tasks they could not solve when responding without intermediate tokens. However, we find empirically that learning to use filler tokens is difficult and requires specific, dense supervision to converge

abstract chain of thought converge however intermediate machinelearning show solve supervision tasks thought tokens transformers

Data Engineer

@ Lemon.io | Remote: Europe, LATAM, Canada, UK, Asia, Oceania

Artificial Intelligence – Bioinformatic Expert

@ University of Texas Medical Branch | Galveston, TX

Lead Developer (AI)

@ Cere Network | San Francisco, US

Research Engineer

@ Allora Labs | Remote

Ecosystem Manager

@ Allora Labs | Remote

Founding AI Engineer, Agents

@ Occam AI | New York