March 7, 2024, 5:45 a.m. | Bingyan Liu, Chengyu Wang, Tingfeng Cao, Kui Jia, Jun Huang

cs.CV updates on arXiv.org arxiv.org

arXiv:2403.03431v1 Announce Type: new
Abstract: Deep Text-to-Image Synthesis (TIS) models such as Stable Diffusion have recently gained significant popularity for creative Text-to-image generation. Yet, for domain-specific scenarios, tuning-free Text-guided Image Editing (TIE) is of greater importance for application developers, which modify objects or object properties in images by manipulating feature components in attention layers during the generation process. However, little is known about what semantic meanings these attention layers have learned and which parts of the attention maps contribute to …

abstract application arxiv attention creative cs.cv developers diffusion domain editing free image image generation images importance object objects self-attention stable diffusion synthesis text text-to-image type understanding

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Machine Learning Research Scientist

@ d-Matrix | San Diego, Ca