April 23, 2024, 4:42 a.m. | Shitong Shao, Zikai Zhou, Huanran Chen, Zhiqiang Shen

cs.LG updates on arXiv.org arxiv.org

arXiv:2404.13733v1 Announce Type: new
Abstract: Dataset condensation, a concept within data-centric learning, efficiently transfers critical attributes from an original dataset to a synthetic version, maintaining both diversity and realism. This approach significantly improves model training efficiency and is adaptable across multiple application areas. Previous methods in dataset condensation have faced challenges: some incur high computational costs which limit scalability to larger datasets (e.g., MTT, DREAM, and TESLA), while others are restricted to less optimal design spaces, which could hinder potential …

abstract application arxiv challenges concept cs.ai cs.cv cs.lg data data-centric dataset design diversity efficiency multiple space synthetic training type

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Machine Learning Engineer

@ Apple | Sunnyvale, California, United States