Jan. 4, 2022, 2:10 a.m. | Phillip Rieger, Thien Duc Nguyen, Markus Miettinen, Ahmad-Reza Sadeghi

cs.LG updates on arXiv.org arxiv.org

Federated Learning (FL) allows multiple clients to collaboratively train a
Neural Network (NN) model on their private data without revealing the data.
Recently, several targeted poisoning attacks against FL have been introduced.
These attacks inject a backdoor into the resulting model that allows
adversary-controlled inputs to be misclassified. Existing countermeasures
against backdoor attacks are inefficient and often merely aim to exclude
deviating models from the aggregation. However, this approach also removes
benign models of clients with deviating data distributions, causing …

arxiv attacks federated learning learning

Data Scientist (m/f/x/d)

@ Symanto Research GmbH & Co. KG | Spain, Germany

Research Analyst

@ Cypris | Los Angeles, California, United States

Data Manager H/F

@ ASSYSTEM | Courbevoie, France

Software Engineer III - Java Scala BigData AWS

@ JPMorgan Chase & Co. | Bengaluru, Karnataka, India

Reference Data Specialist

@ JPMorgan Chase & Co. | Bengaluru, Karnataka, India

Data Visualization Manager

@ PatientPoint | Cincinnati, Ohio, United States