all AI news
Demystifying Why Local Aggregation Helps: Convergence Analysis of Hierarchical SGD
April 12, 2024, 4:43 a.m. | Jiayi Wang, Shiqiang Wang, Rong-Rong Chen, Mingyue Ji
cs.LG updates on arXiv.org arxiv.org
Abstract: Hierarchical SGD (H-SGD) has emerged as a new distributed SGD algorithm for multi-level communication networks. In H-SGD, before each global aggregation, workers send their updated local models to local servers for aggregations. Despite recent research efforts, the effect of local aggregation on global convergence still lacks theoretical understanding. In this work, we first introduce a new notion of "upward" and "downward" divergences. We then use it to conduct a novel analysis to obtain a worst-case …
abstract aggregation algorithm analysis arxiv communication convergence cs.dc cs.it cs.lg distributed global hierarchical math.it math.oc networks research servers type workers
More from arxiv.org / cs.LG updates on arXiv.org
Jobs in AI, ML, Big Data
Software Engineer for AI Training Data (School Specific)
@ G2i Inc | Remote
Software Engineer for AI Training Data (Python)
@ G2i Inc | Remote
Software Engineer for AI Training Data (Tier 2)
@ G2i Inc | Remote
Data Engineer
@ Lemon.io | Remote: Europe, LATAM, Canada, UK, Asia, Oceania
Artificial Intelligence – Bioinformatic Expert
@ University of Texas Medical Branch | Galveston, TX
Lead Developer (AI)
@ Cere Network | San Francisco, US