March 19, 2024, 4:48 a.m. | Beichen Zhang, Xiaoxing Wang, Xiaohan Qin, Junchi Yan

cs.CV updates on arXiv.org arxiv.org

arXiv:2403.11380v1 Announce Type: new
Abstract: Supernet is a core component in many recent Neural Architecture Search (NAS) methods. It not only helps embody the search space but also provides a (relative) estimation of the final performance of candidate architectures. Thus, it is critical that the top architectures ranked by a supernet should be consistent with those ranked by true performance, which is known as the order-preserving ability. In this work, we analyze the order-preserving ability on the whole search space …

abstract architecture architectures arxiv boosting core cs.cv fine-tuning nas neural architecture search performance search space type

AI Engineer Intern, Agents

@ Occam AI | US

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Lead Data Modeler

@ Sherwin-Williams | Cleveland, OH, United States