all AI news
Instance-Aware Group Quantization for Vision Transformers
April 2, 2024, 7:43 p.m. | Jaehyeon Moon, Dohyung Kim, Junyong Cheon, Bumsub Ham
cs.LG updates on arXiv.org arxiv.org
Abstract: Post-training quantization (PTQ) is an efficient model compression technique that quantizes a pretrained full-precision model using only a small calibration set of unlabeled samples without retraining. PTQ methods for convolutional neural networks (CNNs) provide quantization results comparable to full-precision counterparts. Directly applying them to vision transformers (ViTs), however, incurs severe performance degradation, mainly due to the differences in architectures between CNNs and ViTs. In particular, the distribution of activations for each channel vary drastically according …
abstract arxiv cnns compression convolutional neural networks cs.cv cs.lg however instance networks neural networks precision quantization results retraining samples set small them training transformers type vision vision transformers
More from arxiv.org / cs.LG updates on arXiv.org
Jobs in AI, ML, Big Data
Senior Machine Learning Engineer
@ GPTZero | Toronto, Canada
ML/AI Engineer / NLP Expert - Custom LLM Development (x/f/m)
@ HelloBetter | Remote
Doctoral Researcher (m/f/div) in Automated Processing of Bioimages
@ Leibniz Institute for Natural Product Research and Infection Biology (Leibniz-HKI) | Jena
Seeking Developers and Engineers for AI T-Shirt Generator Project
@ Chevon Hicks | Remote
Real World Evidence Research Analyst
@ Novartis | Dublin (Novartis Global Service Center (NGSC))
Senior DataOps Engineer
@ Winterthur Gas & Diesel AG | Winterthur, CH