all AI news
A-SDM: Accelerating Stable Diffusion through Redundancy Removal and Performance Optimization
March 6, 2024, 5:46 a.m. | Jinchao Zhu, Yuxuan Wang, Xiaobing Tu, Siyuan Pan, Pengfei Wan, Gao Huang
cs.CV updates on arXiv.org arxiv.org
Abstract: The Stable Diffusion Model (SDM) is a popular and efficient text-to-image (t2i) generation and image-to-image (i2i) generation model. Although there have been some attempts to reduce sampling steps, model distillation, and network quantization, these previous methods generally retain the original network architecture. Billion scale parameters and high computing requirements make the research of model architecture adjustment scarce. In this work, we first explore the computational redundancy part of the network, and then prune the redundancy …
abstract architecture arxiv billion cs.cv diffusion diffusion model distillation image image-to-image model distillation network network architecture optimization parameters performance popular quantization reduce redundancy sampling scale stable diffusion text text-to-image through type
More from arxiv.org / cs.CV updates on arXiv.org
Compact 3D Scene Representation via Self-Organizing Gaussian Grids
1 day, 4 hours ago |
arxiv.org
Fingerprint Matching with Localized Deep Representation
1 day, 4 hours ago |
arxiv.org
Jobs in AI, ML, Big Data
Founding AI Engineer, Agents
@ Occam AI | New York
AI Engineer Intern, Agents
@ Occam AI | US
AI Research Scientist
@ Vara | Berlin, Germany and Remote
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Lead GNSS Data Scientist
@ Lurra Systems | Melbourne