March 7, 2024, 11:19 a.m. | /u/SunsetOneSix

Machine Learning www.reddit.com

**Paper**: [https://arxiv.org/abs/2402.07043](https://arxiv.org/abs/2402.07043)

**Abstract**:

>As AI model size grows, neural *scaling laws* have become a crucial tool to predict the improvements of large models when increasing capacity and the size of original (human or natural) training data. Yet, the widespread use of popular models means that the ecosystem of online data and text will co-evolve to progressively contain increased amounts of synthesized data. In this paper we ask: *How will the scaling laws change in the inevitable regime where synthetic data …

abstract ai model become capacity data ecosystem human improvements large models laws machinelearning natural popular scaling synthesized text tool training training data will

Data Engineer

@ Lemon.io | Remote: Europe, LATAM, Canada, UK, Asia, Oceania

Artificial Intelligence – Bioinformatic Expert

@ University of Texas Medical Branch | Galveston, TX

Lead Developer (AI)

@ Cere Network | San Francisco, US

Research Engineer

@ Allora Labs | Remote

Ecosystem Manager

@ Allora Labs | Remote

Founding AI Engineer, Agents

@ Occam AI | New York