Feb. 5, 2024, 6:46 a.m. | Mohammadreza Tayaranian Seyyed Hasan Mozafari James J. Clark Brett Meyer Warren Gross

cs.CV updates on arXiv.org arxiv.org

SWIN transformer is a prominent vision transformer model that has state-of-the-art accuracy in image classification tasks. Despite this success, its unique architecture causes slower inference compared with similar deep neural networks. Integer quantization of the model is one of the methods used to improve its inference latency. However, state-of-the-art has not been able to fully quantize the model. In this work, we improve upon the inference latency of the state-of-the-art methods by removing the floating-point operations, which are associated with …

accuracy architecture art classification cs.ai cs.cv faster image inference latency networks neural networks quantization state success swin swin transformer tasks transformer transformer model vision vision transformer model

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Global Data Architect, AVP - State Street Global Advisors

@ State Street | Boston, Massachusetts

Data Engineer

@ NTT DATA | Pune, MH, IN