all AI news
An Interpretable Client Decision Tree Aggregation process for Federated Learning
April 4, 2024, 4:41 a.m. | Alberto Argente-Garrido, Cristina Zuheros, M. Victoria Luz\'on, Francisco Herrera
cs.LG updates on arXiv.org arxiv.org
Abstract: Trustworthy Artificial Intelligence solutions are essential in today's data-driven applications, prioritizing principles such as robustness, safety, transparency, explainability, and privacy among others. This has led to the emergence of Federated Learning as a solution for privacy and distributed machine learning. While decision trees, as self-explanatory models, are ideal for collaborative model training across multiple devices in resource-constrained environments such as federated learning environments for injecting interpretability in these models. Decision tree structure makes the aggregation …
abstract aggregation applications artificial artificial intelligence arxiv client cs.ai cs.lg data data-driven decision decision trees distributed emergence explainability federated learning intelligence machine machine learning privacy process robustness safety solution solutions transparency tree trees trustworthy type
More from arxiv.org / cs.LG updates on arXiv.org
Jobs in AI, ML, Big Data
Software Engineer for AI Training Data (School Specific)
@ G2i Inc | Remote
Software Engineer for AI Training Data (Python)
@ G2i Inc | Remote
Software Engineer for AI Training Data (Tier 2)
@ G2i Inc | Remote
Data Engineer
@ Lemon.io | Remote: Europe, LATAM, Canada, UK, Asia, Oceania
Artificial Intelligence – Bioinformatic Expert
@ University of Texas Medical Branch | Galveston, TX
Lead Developer (AI)
@ Cere Network | San Francisco, US