March 27, 2024, 4:41 a.m. | Radu-Andrei Rosu, Mihaela-Elena Breaban, Henri Luchian

cs.LG updates on arXiv.org arxiv.org

arXiv:2403.17130v1 Announce Type: new
Abstract: Dataset distillation aims at synthesizing a dataset by a small number of artificially generated data items, which, when used as training data, reproduce or approximate a machine learning (ML) model as if it were trained on the entire original dataset. Consequently, data distillation methods are usually tied to a specific ML algorithm. While recent literature deals mainly with distillation of large collections of images in the context of neural network models, tabular data distillation is …

abstract arxiv classification cs.ai cs.lg data data classification dataset distillation generated labels machine machine learning small training training data type

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Reporting & Data Analytics Lead (Sizewell C)

@ EDF | London, GB

Data Analyst

@ Notable | San Mateo, CA