March 6, 2024, 5:43 a.m. | Hongkai Zheng, Weili Nie, Arash Vahdat, Anima Anandkumar

cs.LG updates on arXiv.org arxiv.org

arXiv:2306.09305v2 Announce Type: replace-cross
Abstract: We propose an efficient approach to train large diffusion models with masked transformers. While masked transformers have been extensively explored for representation learning, their application to generative learning is less explored in the vision domain. Our work is the first to exploit masked training to reduce the training cost of diffusion models significantly. Specifically, we randomly mask out a high proportion (e.g., 50%) of patches in diffused input images during training. For masked training, we …

abstract application arxiv cs.ai cs.cv cs.lg diffusion diffusion models domain exploit generative reduce representation representation learning train training transformers type vision work

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Senior Machine Learning Engineer

@ Samsara | Canada - Remote