all AI news
Demystifying Why Local Aggregation Helps: Convergence Analysis of Hierarchical SGD
April 12, 2024, 4:43 a.m. | Jiayi Wang, Shiqiang Wang, Rong-Rong Chen, Mingyue Ji
cs.LG updates on arXiv.org arxiv.org
Abstract: Hierarchical SGD (H-SGD) has emerged as a new distributed SGD algorithm for multi-level communication networks. In H-SGD, before each global aggregation, workers send their updated local models to local servers for aggregations. Despite recent research efforts, the effect of local aggregation on global convergence still lacks theoretical understanding. In this work, we first introduce a new notion of "upward" and "downward" divergences. We then use it to conduct a novel analysis to obtain a worst-case …
abstract aggregation algorithm analysis arxiv communication convergence cs.dc cs.it cs.lg distributed global hierarchical math.it math.oc networks research servers type workers
More from arxiv.org / cs.LG updates on arXiv.org
Jobs in AI, ML, Big Data
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Lead GNSS Data Scientist
@ Lurra Systems | Melbourne
Senior Machine Learning Engineer (MLOps)
@ Promaton | Remote, Europe
Senior Principal, Product Strategy Operations, Cloud Data Analytics
@ Google | Sunnyvale, CA, USA; Austin, TX, USA
Data Scientist - HR BU
@ ServiceNow | Hyderabad, India