all AI news
Approximate Gradient Coding for Privacy-Flexible Federated Learning with Non-IID Data
April 5, 2024, 4:42 a.m. | Okko Makkonen, Sampo Niemel\"a, Camilla Hollanti, Serge Kas Hanna
cs.LG updates on arXiv.org arxiv.org
Abstract: This work focuses on the challenges of non-IID data and stragglers/dropouts in federated learning. We introduce and explore a privacy-flexible paradigm that models parts of the clients' local data as non-private, offering a more versatile and business-oriented perspective on privacy. Within this framework, we propose a data-driven strategy for mitigating the effects of label heterogeneity and client straggling on federated learning. Our solution combines both offline data sharing and approximate gradient coding techniques. Through numerical …
abstract arxiv business challenges coding cs.cr cs.dc cs.it cs.lg data explore federated learning framework gradient math.it paradigm perspective privacy stat.ml type work
More from arxiv.org / cs.LG updates on arXiv.org
Jobs in AI, ML, Big Data
Artificial Intelligence – Bioinformatic Expert
@ University of Texas Medical Branch | Galveston, TX
Lead Developer (AI)
@ Cere Network | San Francisco, US
Research Engineer
@ Allora Labs | Remote
Ecosystem Manager
@ Allora Labs | Remote
Founding AI Engineer, Agents
@ Occam AI | New York
AI Engineer Intern, Agents
@ Occam AI | US