Feb. 23, 2024, 5:44 a.m. | Shahriar Golchin, Mihai Surdeanu

cs.LG updates on arXiv.org arxiv.org

arXiv:2308.08493v3 Announce Type: replace-cross
Abstract: Data contamination, i.e., the presence of test data from downstream tasks in the training data of large language models (LLMs), is a potential major issue in measuring LLMs' real effectiveness on other tasks. We propose a straightforward yet effective method for identifying data contamination within LLMs. At its core, our approach starts by identifying potential contamination at the instance level; using this information, our approach then assesses wider contamination at the partition level. To estimate …

abstract arxiv cs.ai cs.cl cs.cr cs.lg data issue language language models large language large language models llms major measuring tasks test tracing training training data travel type

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne