|
|
--- |
|
|
license: mit |
|
|
pipeline_tag: unconditional-image-generation |
|
|
--- |
|
|
|
|
|
# RAE: Diffusion Transformers with Representation Autoencoders |
|
|
|
|
|
This repository contains the official PyTorch checkpoints for Representation Autoencoders. |
|
|
|
|
|
Representation Autoencoders (RAE) are a class of autoencoders that utilize pretrained, frozen representation encoders such as DINOv2 and SigLIP2 as encoders with trained ViT decoders. RAE can be used in a two-stage training pipeline for high-fidelity image synthesis, where a Stage 2 diffusion model is trained on the latent space of a pretrained RAE to generate images. |
|
|
|
|
|
Website: https://rae-dit.github.io/ |
|
|
|
|
|
Code: https://github.com/bytetriper/RAE |
|
|
|
|
|
Paper: https://huggingface.co/papers/2510.11690 |