all AI news
FALCON: FLOP-Aware Combinatorial Optimization for Neural Network Pruning
March 13, 2024, 4:41 a.m. | Xiang Meng, Wenyu Chen, Riade Benbaki, Rahul Mazumder
cs.LG updates on arXiv.org arxiv.org
Abstract: The increasing computational demands of modern neural networks present deployment challenges on resource-constrained devices. Network pruning offers a solution to reduce model size and computational cost while maintaining performance. However, most current pruning methods focus primarily on improving sparsity by reducing the number of nonzero parameters, often neglecting other deployment costs such as inference time, which are closely related to the number of floating-point operations (FLOPs). In this paper, we propose FALCON, a novel combinatorial-optimization-based …
abstract arxiv challenges computational cost cs.lg current deployment devices falcon focus however modern network networks neural network neural networks optimization parameters performance pruning reduce solution sparsity type
More from arxiv.org / cs.LG updates on arXiv.org
Jobs in AI, ML, Big Data
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Lead GNSS Data Scientist
@ Lurra Systems | Melbourne
Senior Machine Learning Engineer (MLOps)
@ Promaton | Remote, Europe
Business Data Scientist, gTech Ads
@ Google | Mexico City, CDMX, Mexico
Lead, Data Analytics Operations
@ Zocdoc | Pune, Maharashtra, India