all AI news
Investigating the Effect of Misalignment on Membership Privacy in the White-box Setting
March 14, 2024, 4:43 a.m. | Ana-Maria Cretu, Daniel Jones, Yves-Alexandre de Montjoye, Shruti Tople
cs.LG updates on arXiv.org arxiv.org
Abstract: Machine learning models have been shown to leak sensitive information about their training datasets. Models are increasingly deployed on devices, raising concerns that white-box access to the model parameters increases the attack surface compared to black-box access which only provides query access. Directly extending the shadow modelling technique from the black-box to the white-box setting has been shown, in general, not to perform better than black-box only attacks. A potential reason is misalignment, a known …
abstract arxiv box concerns cs.cr cs.lg datasets devices information leak machine machine learning machine learning models parameters privacy query surface training training datasets type
More from arxiv.org / cs.LG updates on arXiv.org
Testing the Segment Anything Model on radiology data
1 day, 18 hours ago |
arxiv.org
Calorimeter shower superresolution
1 day, 18 hours ago |
arxiv.org
Jobs in AI, ML, Big Data
Software Engineer for AI Training Data (School Specific)
@ G2i Inc | Remote
Software Engineer for AI Training Data (Python)
@ G2i Inc | Remote
Software Engineer for AI Training Data (Tier 2)
@ G2i Inc | Remote
Data Engineer
@ Lemon.io | Remote: Europe, LATAM, Canada, UK, Asia, Oceania
Artificial Intelligence – Bioinformatic Expert
@ University of Texas Medical Branch | Galveston, TX
Lead Developer (AI)
@ Cere Network | San Francisco, US