all AI news
Unintended memorisation of unique features in neural networks. (arXiv:2205.10079v1 [cs.LG])
May 23, 2022, 1:12 a.m. | John Hartley, Sotirios A. Tsaftaris
cs.CV updates on arXiv.org arxiv.org
Neural networks pose a privacy risk due to their propensity to memorise and
leak training data. We show that unique features occurring only once in
training data are memorised by discriminative multi-layer perceptrons and
convolutional neural networks trained on benchmark imaging datasets. We design
our method for settings where sensitive training data is not available, for
example medical imaging. Our setting knows the unique feature, but not the
training data, model weights or the unique feature's label. We develop a …
More from arxiv.org / cs.CV updates on arXiv.org
Jobs in AI, ML, Big Data
Senior Machine Learning Engineer (MLOps)
@ Promaton | Remote, Europe
Research Associate (Data Science/Information Engineering/Applied Mathematics/Information Technology)
@ Nanyang Technological University | NTU Main Campus, Singapore
Associate Director of Data Science and Analytics
@ Penn State University | Penn State University Park
Student Worker- Data Scientist
@ TransUnion | Israel - Tel Aviv
Vice President - Customer Segment Analytics Data Science Lead
@ JPMorgan Chase & Co. | Bengaluru, Karnataka, India
Middle/Senior Data Engineer
@ Devexperts | Sofia, Bulgaria