May 7, 2024, 4:43 a.m. | Yuxi Xia, Anastasiia Sedova, Pedro Henrique Luz de Araujo, Vasiliki Kougia, Lisa Nu{\ss}baumer, Benjamin Roth

cs.LG updates on arXiv.org arxiv.org

arXiv:2405.03004v1 Announce Type: cross
Abstract: Training data memorization in language models impacts model capability (generalization) and safety (privacy risk). This paper focuses on analyzing prompts' impact on detecting the memorization of 6 masked language model-based named entity recognition models. Specifically, we employ a diverse set of 400 automatically generated prompts, and a pairwise dataset where each pair consists of one person's name from the training set and another name out of the set. A prompt completed with a person's name …

abstract arxiv capability cs.cl cs.lg data diverse impact impacts language language model language models paper privacy prompts recognition risk safety set training training data type

Senior Machine Learning Engineer

@ GPTZero | Toronto, Canada

ML/AI Engineer / NLP Expert - Custom LLM Development (x/f/m)

@ HelloBetter | Remote

Doctoral Researcher (m/f/div) in Automated Processing of Bioimages

@ Leibniz Institute for Natural Product Research and Infection Biology (Leibniz-HKI) | Jena

Seeking Developers and Engineers for AI T-Shirt Generator Project

@ Chevon Hicks | Remote

Data Architect

@ S&P Global | IN - HYDERABAD SKYVIEW

Data Architect I

@ S&P Global | US - VA - CHARLOTTESVILLE 212 7TH STREET