Get the latest tech news

Faster convergence for diffusion models


Representation Alignment for Generation: Training Diffusion Transformers Is Easier Than You Think

Specifically, we introduce REPresentation Alignment (REPA), a simple regularization technique built on recent diffusion transformer architectures. This, in turn, allows the later layers of the diffusion transformers to focus on capturing high-frequency details based on the aligned representations, further improving generation performance. We also examine the scalability of REPA by varying pretrained encoders and diffusion transformer model sizes, showing that aligning with better visual representations leads to improved generation and linear probing results.

Get the Android app

Or read this on Hacker News

Read more on:

Photo of Diffusion Models

Diffusion Models

Photo of 20x

20x

Photo of faster convergence

faster convergence

Related news:

News photo

Pulsar: Secure Steganography for Diffusion Models

News photo

Tutorial on diffusion models for imaging and vision

News photo

Diffusion models are real-time game engines