Oct. 25, 2022, 1:16 a.m. | Chiyu Zhang, Jun Yang, Lei Wang, Zaiyan Dai

cs.CV updates on arXiv.org arxiv.org

This paper presents a new hierarchical vision Transformer for image style
transfer, called Strips Window Attention Transformer (S2WAT), which serves as
an encoder of encoder-transfer-decoder architecture. With hierarchical
features, S2WAT can leverage proven techniques in other fields of computer
vision, such as feature pyramid networks (FPN) or U-Net, to image style
transfer in future works. However, the existing window-based Transformers will
cause a problem that the stylized images will be grid-like when introducing
them into image style transfer directly. To …

arxiv attention hierarchical image style transfer transfer transformer vision

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Alternant Data Engineering

@ Aspire Software | Angers, FR

Senior Software Engineer, Generative AI

@ Google | Dublin, Ireland