Nov. 1, 2022, 1:13 a.m. | Zhiying Xu, Jiafan Xu, Hongding Peng, Wei Wang, Xiaoliang Wang, Haoran Wan, Haipeng Dai, Yixu Xu, Hao Cheng, Kun Wang, Guihai Chen

cs.LG updates on arXiv.org arxiv.org

Deep learning models rely on highly optimized tensor libraries for efficient
inference on heterogeneous hardware. Current deep compilers typically
predetermine layouts of tensors and then optimize loops of operators. However,
such unidirectional and one-off workflow strictly separates graph-level
optimization and operator-level optimization into different system layers,
missing opportunities for unified tuning. This paper proposes ALT, a compiler
that performs joint graph- and operator-level optimizations for deep models.
ALT provides a generic transformation module to manipulate layouts and loops
with easy-to-use …

arxiv boosting breaking deep learning deep learning performance graph performance

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Data Engineer

@ Kaseya | Bengaluru, Karnataka, India