March 21, 2024, 4:42 a.m. | Yumeng Li, William Beluch, Margret Keuper, Dan Zhang, Anna Khoreva

cs.LG updates on arXiv.org arxiv.org

arXiv:2403.13501v1 Announce Type: cross
Abstract: Despite tremendous progress in the field of text-to-video (T2V) synthesis, open-sourced T2V diffusion models struggle to generate longer videos with dynamically varying and evolving content. They tend to synthesize quasi-static videos, ignoring the necessary visual change-over-time implied in the text prompt. At the same time, scaling these models to enable longer, more dynamic video synthesis often remains computationally intractable. To address this challenge, we introduce the concept of Generative Temporal Nursing (GTN), where we aim …

abstract arxiv change cs.ai cs.cv cs.lg cs.mm diffusion diffusion models dynamic generate generative nursing progress prompt scaling struggle synthesis temporal text text-to-video type video videos visual

Software Engineer for AI Training Data (School Specific)

@ G2i Inc | Remote

Software Engineer for AI Training Data (Python)

@ G2i Inc | Remote

Software Engineer for AI Training Data (Tier 2)

@ G2i Inc | Remote

Data Engineer

@ Lemon.io | Remote: Europe, LATAM, Canada, UK, Asia, Oceania

Artificial Intelligence – Bioinformatic Expert

@ University of Texas Medical Branch | Galveston, TX

Lead Developer (AI)

@ Cere Network | San Francisco, US