April 3, 2024, 4:43 a.m. | Tonmoy Dey, Yixin Chen, Alan Kuhnle

cs.LG updates on arXiv.org arxiv.org

arXiv:2206.09563v5 Announce Type: replace-cross
Abstract: Distributed maximization of a submodular function in the MapReduce (MR) model has received much attention, culminating in two frameworks that allow a centralized algorithm to be run in the MR setting without loss of approximation, as long as the centralized algorithm satisfies a certain consistency property - which had previously only been known to be satisfied by the standard greedy and continous greedy algorithms. A separate line of work has studied parallelizability of submodular maximization …

abstract algorithm algorithms approximation arxiv attention complexity cs.dc cs.ds cs.lg distributed frameworks function loss mapreduce scalable type

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Senior Data Engineer

@ Cint | Gurgaon, India

Data Science (M/F), setor automóvel - Aveiro

@ Segula Technologies | Aveiro, Portugal