April 19, 2024, 4:41 a.m. | Thibault Castells, Hyoung-Kyu Song, Bo-Kyeong Kim, Shinkook Choi

cs.LG updates on arXiv.org arxiv.org

arXiv:2404.11936v1 Announce Type: new
Abstract: Latent Diffusion Models (LDMs) have emerged as powerful generative models, known for delivering remarkable results under constrained computational resources. However, deploying LDMs on resource-limited devices remains a complex issue, presenting challenges such as memory consumption and inference speed. To address this issue, we introduce LD-Pruner, a novel performance-preserving structured pruning method for compressing LDMs. Traditional pruning methods for deep neural networks are not tailored to the unique characteristics of LDMs, such as the high computational …

abstract arxiv challenges computational consumption cs.ai cs.cv cs.lg devices diffusion diffusion models generative generative models however inference insights issue latent diffusion models memory memory consumption presenting pruning resources results speed type

Data Engineer

@ Lemon.io | Remote: Europe, LATAM, Canada, UK, Asia, Oceania

Artificial Intelligence – Bioinformatic Expert

@ University of Texas Medical Branch | Galveston, TX

Lead Developer (AI)

@ Cere Network | San Francisco, US

Research Engineer

@ Allora Labs | Remote

Ecosystem Manager

@ Allora Labs | Remote

Founding AI Engineer, Agents

@ Occam AI | New York