Feb. 23, 2024, 5:43 a.m. | Miaoxin Wang, Xiao Wu, Jun Lin, Zhongfeng Wang

cs.LG updates on arXiv.org arxiv.org

arXiv:2402.14307v1 Announce Type: cross
Abstract: Convolutional neural networks (CNNs) with large kernels, drawing inspiration from the key operations of vision transformers (ViTs), have demonstrated impressive performance in various vision-based applications. To address the issue of computational efficiency degradation in existing designs for supporting large-kernel convolutions, an FPGA-based inference accelerator is proposed for the efficient deployment of CNNs with arbitrary kernel sizes. Firstly, a Z-flow method is presented to optimize the computing data flow by maximizing data reuse opportunity. Besides, the …

abstract accelerator applications arxiv cnns computational convolutional neural networks cs.ar cs.lg designs efficiency enabling fpga inference inspiration issue kernel key networks neural networks operations performance support the key transformers type vision vision transformers

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Business Data Scientist, gTech Ads

@ Google | Mexico City, CDMX, Mexico

Lead, Data Analytics Operations

@ Zocdoc | Pune, Maharashtra, India