all AI news
Training Differentially Private Models with Secure Multiparty Computation. (arXiv:2202.02625v3 [cs.CR] UPDATED)
Sept. 5, 2022, 1:12 a.m. | Sikha Pentyala, Davis Railsback, Ricardo Maia, Rafael Dowsley, David Melanson, Anderson Nascimento, Martine De Cock
cs.LG updates on arXiv.org arxiv.org
We address the problem of learning a machine learning model from training
data that originates at multiple data owners while providing formal privacy
guarantees regarding the protection of each owner's data. Existing solutions
based on Differential Privacy (DP) achieve this at the cost of a drop in
accuracy. Solutions based on Secure Multiparty Computation (MPC) do not incur
such accuracy loss but leak information when the trained model is made publicly
available. We propose an MPC solution for training DP …
More from arxiv.org / cs.LG updates on arXiv.org
A Single-Loop Algorithm for Decentralized Bilevel Optimization
1 day, 6 hours ago |
arxiv.org
CLEANing Cygnus A deep and fast with R2D2
1 day, 6 hours ago |
arxiv.org
Jobs in AI, ML, Big Data
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Lead GNSS Data Scientist
@ Lurra Systems | Melbourne
Senior Machine Learning Engineer (MLOps)
@ Promaton | Remote, Europe
Alternant Data Engineering
@ Aspire Software | Angers, FR
Senior Software Engineer, Generative AI
@ Google | Dublin, Ireland