Feb. 22, 2024, 5:42 a.m. | Robin Staab, Mark Vero, Mislav Balunovi\'c, Martin Vechev

cs.LG updates on arXiv.org arxiv.org

arXiv:2402.13846v1 Announce Type: cross
Abstract: Recent work in privacy research on large language models has shown that they achieve near human-level performance at inferring personal data from real-world online texts. With consistently increasing model capabilities, existing text anonymization methods are currently lacking behind regulatory requirements and adversarial threats. This raises the question of how individuals can effectively protect their personal data in sharing online texts. In this work, we take two steps to answer this question: We first present a …

abstract advanced adversarial anonymization arxiv capabilities cs.ai cs.cl cs.cr cs.lg data human language language models large language large language models near performance personal data privacy question raises regulatory requirements research text threats type work world

Software Engineer for AI Training Data (School Specific)

@ G2i Inc | Remote

Software Engineer for AI Training Data (Python)

@ G2i Inc | Remote

Software Engineer for AI Training Data (Tier 2)

@ G2i Inc | Remote

Data Engineer

@ Lemon.io | Remote: Europe, LATAM, Canada, UK, Asia, Oceania

Artificial Intelligence – Bioinformatic Expert

@ University of Texas Medical Branch | Galveston, TX

Lead Developer (AI)

@ Cere Network | San Francisco, US