June 10, 2022, 1:10 a.m. | Kai Yue, Richeng Jin, Chau-Wai Wong, Dror Baron, Huaiyu Dai

cs.LG updates on arXiv.org arxiv.org

Federated learning has been proposed as a privacy-preserving machine learning
framework that enables multiple clients to collaborate without sharing raw
data. However, client privacy protection is not guaranteed by design in this
framework. Prior work has shown that the gradient sharing strategies in
federated learning can be vulnerable to data reconstruction attacks. In
practice, though, clients may not transmit raw gradients considering the high
communication cost or due to privacy enhancement requirements. Empirical
studies have demonstrated that gradient obfuscation, including …

arxiv false federated learning gradient learning security sense

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Principal Engineer, Deep Learning

@ Outrider | Remote

Data Analyst (Bangkok based, relocation provided)

@ Agoda | Bangkok (Central World Office)

Data Scientist II

@ MoEngage | Bengaluru

Machine Learning Engineer

@ Sika AG | Welwyn Garden City, United Kingdom