Nov. 15, 2022, 2:12 a.m. | Pedro Soto, Haibin Guan, Jun Li

cs.LG updates on arXiv.org arxiv.org

Matrix multiplication is a fundamental operation in machine learning and is
commonly distributed into multiple parallel tasks for large datasets.
Stragglers and other failures can severely impact the overall completion time.
Recent works in coded computing provide a novel strategy to mitigate stragglers
with coded tasks, with an objective of minimizing the number of tasks needed to
recover the overall result, known as the recovery threshold. However, we
demonstrate that this combinatorial definition does not directly optimize the
probability of …

arxiv coding distributed matrix matrix multiplication random

Lead Developer (AI)

@ Cere Network | San Francisco, US

Research Engineer

@ Allora Labs | Remote

Ecosystem Manager

@ Allora Labs | Remote

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US

AI Research Scientist

@ Vara | Berlin, Germany and Remote