readout-guidance.github.io - Readout Guidance: Learning Control from Diffusion Features

Description: We train very small networks called readout heads to predict useful properties and guide image generation.

guidance (655) readout (5) diffusion hyperfeatures (2) readout guidance (1) diffusion features (1)

Example domain paragraphs

We present Readout Guidance, a method for controlling text-to-image diffusion models with learned signals. Readout Guidance uses readout heads , lightweight networks trained to extract signals from the features of a pre-trained, frozen diffusion model at every timestep. These readouts can encode single-image properties, such as pose, depth, and edges; or higher-order properties that relate multiple images, such as correspondence and appearance similarity. Furthermore, by comparing the readout estimates to a

These readouts can also be used for controlled image generation---by guiding the readout towards some desired value. Readouts for single-image concepts, such as pose and depth, enable spatially aligned control. Readouts for relative concepts between two images, such as appearance similarity and correspondence, enable cross-image controls, such as drag-based manipulation, identity-consistent generation, or image variations.

Links to readout-guidance.github.io (1)