all AI news
Vision Transformer Adapter for Dense Predictions. (arXiv:2205.08534v2 [cs.CV] UPDATED)
May 19, 2022, 1:10 a.m. | Zhe Chen, Yuchen Duan, Wenhai Wang, Junjun He, Tong Lu, Jifeng Dai, Yu Qiao
cs.CV updates on arXiv.org arxiv.org
This work investigates a simple yet powerful adapter for Vision Transformer
(ViT). Unlike recent visual transformers that introduce vision-specific
inductive biases into their architectures, ViT achieves inferior performance on
dense prediction tasks due to lacking prior information of images. To solve
this issue, we propose a Vision Transformer Adapter (ViT-Adapter), which can
remedy the defects of ViT and achieve comparable performance to vision-specific
models by introducing inductive biases via an additional architecture.
Specifically, the backbone in our framework is a …
More from arxiv.org / cs.CV updates on arXiv.org
Jobs in AI, ML, Big Data
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Lead GNSS Data Scientist
@ Lurra Systems | Melbourne
Senior Machine Learning Engineer (MLOps)
@ Promaton | Remote, Europe
Alternant Data Engineering
@ Aspire Software | Angers, FR
Senior Software Engineer, Generative AI
@ Google | Dublin, Ireland