Feb. 12, 2024, 5:45 a.m. | David Yan Winnie Zhang Luxin Zhang Anmol Kalia Dingkang Wang Ankit Ramchandani Miao Liu Albert

cs.CV updates on arXiv.org arxiv.org

We introduce animated stickers, a video diffusion model which generates an animation conditioned on a text prompt and static sticker image. Our model is built on top of the state-of-the-art Emu text-to-image model, with the addition of temporal layers to model motion. Due to the domain gap, i.e. differences in visual and motion style, a model which performed well on generating natural videos can no longer generate vivid videos when applied to stickers. To bridge this gap, we employ a …

animated animation art cs.cv differences diffusion diffusion model domain emu gap image life prompt state sticker stickers temporal text text-to-image video video diffusion visual

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Director, Clinical Data Science

@ Aura | Remote USA

Research Scientist, AI (PhD)

@ Meta | Menlo Park, CA | New York City