Oct. 7, 2022, 1:11 a.m. | Jiawei Shao, Yuchang Sun, Songze Li, Jun Zhang

cs.LG updates on arXiv.org arxiv.org

Federated learning (FL) strives to enable collaborative training of machine
learning models without centrally collecting clients' private data. Different
from centralized training, the local datasets across clients in FL are
non-independent and identically distributed (non-IID). In addition, the
data-owning clients may drop out of the training process arbitrarily. These
characteristics will significantly degrade the training performance. This paper
proposes a Dropout-Resilient Secure Federated Learning (DReS-FL) framework
based on Lagrange coded computing (LCC) to tackle both the non-IID and dropout
problems. …

arxiv data data sharing dropout federated learning resilient secret

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne