April 19, 2024, 4:44 a.m. | Kyunghwan Shim, Jaewoong Yun, Shinkook Choi

cs.CV updates on arXiv.org arxiv.org

arXiv:2404.11630v1 Announce Type: new
Abstract: Multi-head self-attention (MSA) is a key component of Vision Transformers (ViTs), which have achieved great success in various vision tasks. However, their high computational cost and memory footprint hinder their deployment on resource-constrained devices. Conventional pruning approaches can only compress and accelerate the MSA module using head pruning, although the head is not an atomic unit. To address this issue, we propose a novel graph-aware neuron-level pruning method, Structured Neuron-level Pruning (SNP). SNP prunes neurons …

abstract arxiv attention computational cost cs.ai cs.cv deployment devices head hinder however key memory multi-head neuron pruning self-attention success tasks transformers type vision vision transformers

Data Engineer

@ Lemon.io | Remote: Europe, LATAM, Canada, UK, Asia, Oceania

Artificial Intelligence – Bioinformatic Expert

@ University of Texas Medical Branch | Galveston, TX

Lead Developer (AI)

@ Cere Network | San Francisco, US

Research Engineer

@ Allora Labs | Remote

Ecosystem Manager

@ Allora Labs | Remote

Founding AI Engineer, Agents

@ Occam AI | New York