What Matters for Diffusion-Friendly Latent Manifold? Prior-Aligned Autoencoders for Latent Diffusion
Abstract
Research investigates latent manifold properties for diffusion models and proposes a Prior-Aligned AutoEncoder that explicitly optimizes latent space structure for improved generative modeling.
Tokenizers are a crucial component of latent diffusion models, as they define the latent space in which diffusion models operate. However, existing tokenizers are primarily designed to improve reconstruction fidelity or inherit pretrained representations, leaving unclear what kind of latent space is truly friendly for generative modeling. In this paper, we study this question from the perspective of latent manifold organization. By constructing controlled tokenizer variants, we identify three key properties of a diffusion-friendly latent manifold: coherent spatial structure, local manifold continuity, and global manifold semantics. We find that these properties are more consistent with downstream generation quality than reconstruction fidelity. Motivated by this finding, we propose the Prior-Aligned AutoEncoder (PAE), which explicitly shapes the latent manifold instead of leaving diffusion-friendly manifold to emerge indirectly from reconstruction or inheritance. Specifically, PAE leverages refined priors derived from VFMs and perturbation-based regularization to turn spatial structure, local continuity, and global semantics into explicit training objectives. On ImageNet 256x256, PAE improves both training efficiency and generation quality over existing tokenizers, reaching performance comparable to RAE with up to 13x faster convergence under the same training setup and achieving a new state-of-the-art gFID of 1.03. These results highlight the importance of organizing the latent manifold for latent diffusion models.
Community
š PAE: State-of-the-Art Latent Diffusion with 13x Faster Training!
We are excited to present Prior-Aligned AutoEncoder (PAE), a new paradigm for constructing diffusion-friendly latent manifolds! By explicitly shaping the latent space geometry, PAE breaks the trade-off between reconstruction fidelity and generation learnability.
š„ Highlights:
State-of-the-Art Quality: Achieves a new SOTA gFID of 1.03 on ImageNet 256Ć256, surpassing strong baselines like RAE and FAE.
Unprecedented Efficiency: Enables 13x faster convergence for downstream DiT training. Reaches competitive performance in just 80 epochs compared to 800+ epochs for previous methods.
Diffusion-Friendly Manifold: Explicitly optimizes three key geometric properties: Spatial Structure Coherence, Local Manifold Continuity, and Global Semantic Organization, ensuring smooth and semantically consistent latent spaces.
Robust Few-Step Sampling: Maintains high-quality generation (gFID 1.05) with only 45 denoising steps, thanks to improved local continuity.
Available Models: Pre-trained tokenizers based on DINOv2, SigLIP2, DINOv3, and MAE backbones! Code & Models are fully open-sourced on HuggingFace and ModelScope.
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- Taming Sampling Perturbations with Variance Expansion Loss for Latent Diffusion Models (2026)
- RPiAE: A Representation-Pivoted Autoencoder Enhancing Both Image Generation and Editing (2026)
- Latent-Compressed Variational Autoencoder for Video Diffusion Models (2026)
- Coevolving Representations in Joint Image-Feature Diffusion (2026)
- TC-AE: Unlocking Token Capacity for Deep Compression Autoencoders (2026)
- End-to-End Training for Unified Tokenization and Latent Denoising (2026)
- Repurposing Geometric Foundation Models for Multi-view Diffusion (2026)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment: @librarian-bot recommend
Get this paper in your agent:
hf papers read 2605.07915 Don't have the latest CLI?
curl -LsSf https://hf.co/cli/install.sh | bash Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper
