Command Palette

Search for a command to run...

6시간 전

DeCo: 종단간 이미지 생성을 위한 주파수 분리 픽셀 확산

Zehong Ma Longhui Wei Shuai Wang Shiliang Zhang Qi Tian

DeCo: 종단간 이미지 생성을 위한 주파수 분리 픽셀 확산

초록

픽셀 디퓨전(Pixel diffusion)은 엔드투엔드(End-to-End) 방식으로 픽셀 공간에서 이미지를 직접 생성하는 것을 목표로 합니다. 이러한 접근 방식은 2단계 잠재 디퓨전(Latent diffusion) 모델에서 발생하는 VAE(Variational Autoencoder)의 한계를 극복하며, 더 높은 모델 용량을 제공합니다. 그러나 기존의 픽셀 디퓨전 모델들은 단일 디퓨전 트랜스포머(DiT) 내에서 고주파 신호와 저주파 의미 정보(Low-frequency semantics)를 동시에 모델링해야 하기 때문에 훈련 및 추론 속도가 느리다는 단점이 있습니다.이에 우리는 보다 효율적인 픽셀 디퓨전 패러다임을 구현하기 위해 주파수 분리(Frequency-DeCoupled) 픽셀 디퓨전 프레임워크를 제안합니다. 고주파와 저주파 요소의 생성을 분리한다는 직관을 바탕으로, 우리는 DiT의 시맨틱 가이던스(Semantic guidance)를 조건으로 하여 고주파 세부 정보를 생성하는 경량 픽셀 디코더를 활용합니다. 이를 통해 결과적으로 DiT는 저주파 의미 정보를 모델링하는 데 더욱 집중할 수 있게 됩니다.또한, 우리는 시각적으로 두드러지는 주파수를 강조하고 중요하지 않은 주파수는 억제하는 주파수 인식 플로우 매칭(Frequency-aware flow-matching) 손실을 도입했습니다. 광범위한 실험 결과, 우리가 제안한 DeCo는 픽셀 디퓨전 모델 중 우수한 성능을 입증하며 ImageNet 데이터셋에서 1.62(256x256 해상도) 및 2.22(512x512 해상도)의 FID를 달성하여 잠재 디퓨전 방식과의 격차를 크게 좁혔습니다. 아울러, 사전 학습된 우리의 텍스트-이미지 생성 모델은 GenEval 시스템 수준 비교에서 0.86이라는 최고 수준의 종합 점수를 기록했습니다. 관련 코드는 https://github.com/Zehong-Ma/DeCo 에서 확인하실 수 있습니다.

Summarization

Researchers from Peking University, Huawei Inc., and Nanjing University introduce DeCo, a frequency-decoupled pixel diffusion framework that closes the performance gap with latent diffusion methods by employing a lightweight decoder for high-frequency details and a frequency-aware flow-matching loss to enable efficient semantic modeling.

Introduction

High-fidelity image generation is currently dominated by diffusion models, typically categorized into latent diffusion (which compresses images via VAEs) and pixel diffusion (which models raw pixels directly). While pixel diffusion avoids the compression artifacts and training instability associated with VAEs, it faces a significant computational hurdle: jointly modeling complex high-frequency details and low-frequency semantics within a massive pixel space is inefficient and prone to noise interference.

Current pixel diffusion methods often rely on a single Diffusion Transformer (DiT) to handle all frequency components simultaneously. The authors identify that high-frequency noise in this setup distracts the model, preventing it from effectively learning low-frequency semantic structures and leading to degraded image quality.

To address this, the authors introduce DeCo, a frequency-decoupled framework that splits the generation process based on frequency components. By assigning semantic modeling to the DiT and detail reconstruction to a specialized decoder, the system optimizes both aspects independently.

Key innovations include:

  • Decoupled Architecture: The model utilizes a DiT to capture low-frequency semantics from downsampled inputs, while a lightweight pixel decoder uses these semantic cues to reconstruct high-frequency details from high-resolution inputs.
  • Frequency-Aware Loss: The authors implement a novel Flow-Matching loss inspired by JPEG compression, which uses Discrete Cosine Transform (DCT) and adaptive weighting to prioritize visually salient frequencies while suppressing imperceptible high-frequency noise.
  • Enhanced Pixel Fidelity: By removing the reliance on VAE compression and specializing the model components, DeCo achieves superior FID scores on ImageNet, effectively closing the performance gap between pixel-space and latent-space diffusion models.

Method

The authors leverage a frequency-decoupled framework, DeCo, to improve pixel diffusion by separating the modeling of low-frequency semantics from high-frequency signal generation. The overall architecture, as illustrated in the framework diagram, consists of two primary components: a Diffusion Transformer (DiT) and a lightweight pixel decoder. The DiT operates on downsampled, small-scale inputs to model low-frequency semantics, while the pixel decoder generates high-frequency details conditioned on the semantic cues from the DiT. This decoupling allows the DiT to specialize in semantic modeling without being distracted by high-frequency noise, leading to more efficient training and improved visual fidelity.

The pixel decoder is a lightweight, attention-free network composed of NNN linear decoder blocks and several linear projection layers. It directly processes the full-resolution noised image xtx_txt without downsampling. All operations are local and linear, enabling efficient high-frequency modeling. The dense query construction begins by concatenating the noised pixels with their corresponding positional embeddings pos\text{pos}pos and linearly projecting them with WinW_{\text{in}}Win to form dense query vectors h0h_0h0. For each decoder block, the DiT output ccc is linearly upsampled and reshaped to match the spatial resolution of xtx_txt, yielding cupc_{\text{up}}cup. A multi-layer perceptron (MLP) generates modulation parameters α,β,γ\alpha, \beta, \gammaα,β,γ for adaptive layer norm (AdaLN), which are used to modulate the dense decoder queries. The final output of the decoder is mapped to the pixel space via a linear projection and rearrangement operation to predict the pixel velocity vθ(xt,t,y)v_{\theta}(x_t, t, y)vθ(xt,t,y).

To further emphasize visually salient frequencies and suppress insignificant high-frequency components, the authors introduce a frequency-aware flow-matching (FM) loss. This loss reweights different frequency components using adaptive weights derived from JPEG perceptual priors. The transformation from the spatial domain to the frequency domain is achieved by converting the color space to YCbCr and applying a block-wise 8×88 \times 88×8 discrete cosine transform (DCT). The adaptive weights are generated from the normalized reciprocal of the scaled quantization tables QcurQ_{\text{cur}}Qcur, which are designed based on the human visual system's sensitivity to different frequencies. Frequencies with smaller quantization intervals are considered more perceptually important and are assigned higher weights. The frequency-aware FM loss is defined as the expectation of the weighted squared difference between the predicted and ground-truth frequency-domain velocities. The overall training objective combines the standard pixel-level flow-matching loss, the frequency-aware FM loss, and the REPA alignment loss.

Experiment

  • DCT Energy Spectra Analysis: Validates that DeCo effectively decouples frequencies by maintaining high-frequency components in the pixel decoder while shifting them away from the DiT, which focuses on low-frequency semantics.
  • Ablation Studies: Confirm the critical role of the multi-scale input strategy and AdaLN-based interaction, identifying optimal settings with a pixel decoder hidden size of 32, depth of 3, and patch size of 1.
  • Class-to-Image Generation (ImageNet 256x256): The model achieves a leading FID of 1.62 and an inference latency of 1.05s, significantly surpassing RDM (38.4s) and PixelFlow (9.78s) in efficiency while matching two-stage latent diffusion models.
  • Training Efficiency: DeCo demonstrates a 10x improvement in training efficiency, achieving an FID of 2.57 in just 80 epochs, which exceeds the baseline's performance at 800 epochs.
  • Class-to-Image Generation (ImageNet 512x512): The method attains a superior FID of 2.22, delivering performance comparable to DiT-XL/2 and SiT-XL/2.
  • Text-to-Image Generation: On the GenEval benchmark, DeCo achieves an overall score of 0.86, outperforming prominent models such as SD3, FLUX.1-dev, and BLIP3o, as well as pixel diffusion methods like PixNerd.

The authors use a baseline DiT model and compare it to their DeCo framework, which replaces the final two DiT blocks with a pixel decoder. Results show that DeCo achieves significantly better generation quality, reducing FID from 61.10 to 31.35 and increasing IS from 16.81 to 48.35, while maintaining comparable training and inference costs.

The authors use a multi-scale input strategy and AdaLN-based interaction to effectively decouple high-frequency signals from the DiT, allowing the pixel decoder to model fine details while the DiT focuses on low-frequency semantics. Results show that DeCo achieves superior performance on both class-to-image and text-to-image generation tasks, outperforming existing pixel diffusion models and achieving competitive results with two-stage latent diffusion methods.

The authors use DeCo to achieve state-of-the-art performance in text-to-image generation, outperforming models like SD3 and FLUX.1-dev on the GenEval benchmark with an overall score of 0.86, while also achieving competitive results on DPG-Bench. DeCo surpasses other end-to-end pixel diffusion methods such as PixelFlow and PixNerd, demonstrating that pixel-based diffusion can match two-stage models in quality with lower computational costs.

The authors analyze the impact of the pixel decoder's hidden size on performance, showing that a hidden size of 32 achieves the best results with an FID of 34.12, sFID of 10.41, and IS of 46.44. A smaller size of 16 leads to lower performance, while increasing to 64 provides no further improvement.

Results show that the pixel decoder depth of 3 achieves the best performance, with a FID of 34.12, sFID of 10.41, and IS of 46.44, outperforming both shallower and deeper configurations.

AI로 AI 구축

아이디어에서 출시까지 — 무료 AI 공동 코딩, 즉시 사용 가능한 환경, 최적 가격 GPU로 AI 개발을 가속화하세요.

AI 공동 코딩
즉시 사용 가능한 GPU
최적 가격
시작하기

Hyper Newsletters

최신 정보 구독하기
한국 시간 매주 월요일 오전 9시 에 이번 주의 최신 업데이트를 메일로 발송합니다
이메일 서비스 제공: MailChimp