March 7, 2024, 5:41 a.m. | Xinwei Ou, Ce Zhu, Xiaolin Huang, Yipeng Liu

cs.LG updates on arXiv.org arxiv.org

arXiv:2403.03473v1 Announce Type: new
Abstract: Second-order methods can converge much faster than first-order methods by incorporating second-order derivates or statistics, but they are far less prevalent in deep learning due to their computational inefficiency. To handle this, many of the existing solutions focus on reducing the size of the matrix to be inverted. However, it is still needed to perform the inverse operator in each iteration. In this paper, we present a fast natural gradient descent (FNGD) method, which only …

abstract arxiv computational converge cs.cv cs.lg deep learning faster focus free gradient matrix natural solutions statistics the matrix type

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne