March 1, 2024, 5:44 a.m. | Florine W. DekkerDelft University of Technology, the Netherlands and, Zekeriya ErkinDelft University of Technology, the Netherlands and, Mauro ContiUn

cs.LG updates on arXiv.org arxiv.org

arXiv:2312.05248v2 Announce Type: replace-cross
Abstract: Decentralised learning has recently gained traction as an alternative to federated learning in which both data and coordination are distributed over its users. To preserve data confidentiality, decentralised learning relies on differential privacy, multi-party computation, or a combination thereof. However, running multiple privacy-preserving summations in sequence may allow adversaries to perform reconstruction attacks. Unfortunately, current reconstruction countermeasures either cannot trivially be adapted to the distributed setting, or add excessive amounts of noise.
In this work, …

abstract arxiv combination computation cs.cr cs.dc cs.dm cs.lg data decentralised differential differential privacy distributed federated learning multiple prevention privacy running topology type

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Principal Data Engineering Manager

@ Microsoft | Redmond, Washington, United States

Machine Learning Engineer

@ Apple | San Diego, California, United States