HyperAIHyperAI

Command Palette

Search for a command to run...

GMem: A Modular Approach for Ultra-Efficient Generative Models

Yi Tang Peng Sun Zhenglin Cheng Tao Lin

Abstract

Recent studies indicate that the denoising process in deep generative diffusion models implicitly learns and memorizes semantic information from the data distribution. These findings suggest that capturing more complex data distributions requires larger neural networks, leading to a substantial increase in computational demands, which in turn become the primary bottleneck in both training and inference of diffusion models. To this end, we introduce GMem: A Modular Approach for Ultra-Efficient Generative Models. Our approach GMem decouples the memory capacity from model and implements it as a separate, immutable memory set that preserves the essential semantic information in the data. The results are significant: GMem enhances both training, sampling efficiency, and diversity generation. This design on one hand reduces the reliance on network for memorize complex data distribution and thus enhancing both training and sampling efficiency. On ImageNet at 256×256256 \times 256256×256 resolution, GMem achieves a 50×50\times50× training speedup compared to SiT, reaching FID =7.66=7.66=7.66 in fewer than 282828 epochs (4\sim 44 hours training time), while SiT requires 140014001400 epochs. Without classifier-free guidance, GMem achieves state-of-the-art (SoTA) performance FID =1.53=1.53=1.53 in 160160160 epochs with only 20\sim 2020 hours of training, outperforming LightningDiT which requires 800800800 epochs and 95\sim 9595 hours to attain FID =2.17=2.17=2.17.


Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing

HyperAI Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp