all AI news
An Interpretable Client Decision Tree Aggregation process for Federated Learning
April 4, 2024, 4:41 a.m. | Alberto Argente-Garrido, Cristina Zuheros, M. Victoria Luz\'on, Francisco Herrera
cs.LG updates on arXiv.org arxiv.org
Abstract: Trustworthy Artificial Intelligence solutions are essential in today's data-driven applications, prioritizing principles such as robustness, safety, transparency, explainability, and privacy among others. This has led to the emergence of Federated Learning as a solution for privacy and distributed machine learning. While decision trees, as self-explanatory models, are ideal for collaborative model training across multiple devices in resource-constrained environments such as federated learning environments for injecting interpretability in these models. Decision tree structure makes the aggregation …
abstract aggregation applications artificial artificial intelligence arxiv client cs.ai cs.lg data data-driven decision decision trees distributed emergence explainability federated learning intelligence machine machine learning privacy process robustness safety solution solutions transparency tree trees trustworthy type
More from arxiv.org / cs.LG updates on arXiv.org
Jobs in AI, ML, Big Data
AI Research Scientist
@ Vara | Berlin, Germany and Remote
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Lead GNSS Data Scientist
@ Lurra Systems | Melbourne
Senior Machine Learning Engineer (MLOps)
@ Promaton | Remote, Europe
Robotics Technician - 3rd Shift
@ GXO Logistics | Perris, CA, US, 92571