Abstract
The proposed Flow Map Trajectory Tilting (FMTT) algorithm improves diffusion models at test-time by leveraging flow maps to better align with user-specified rewards, enabling more effective sampling and image editing.
A common recipe to improve diffusion models at test-time so that samples score highly against a user-specified reward is to introduce the gradient of the reward into the dynamics of the diffusion itself. This procedure is often ill posed, as user-specified rewards are usually only well defined on the data distribution at the end of generation. While common workarounds to this problem are to use a denoiser to estimate what a sample would have been at the end of generation, we propose a simple solution to this problem by working directly with a flow map. By exploiting a relationship between the flow map and velocity field governing the instantaneous transport, we construct an algorithm, Flow Map Trajectory Tilting (FMTT), which provably performs better ascent on the reward than standard test-time methods involving the gradient of the reward. The approach can be used to either perform exact sampling via importance weighting or principled search that identifies local maximizers of the reward-tilted distribution. We demonstrate the efficacy of our approach against other look-ahead techniques, and show how the flow map enables engagement with complicated reward functions that make possible new forms of image editing, e.g. by interfacing with vision language models.
Community
TL;DR: We introduce Flow Map Trajectory Tilting (FMTT), which uses a flow-map look-ahead to apply rewards throughout generative diffusion trajectories, enabling more effective test-time optimization than denoiser-based guidance and achieving better sampling/search under complex rewards (including VLM-based ones).
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- Effective Test-Time Scaling of Discrete Diffusion through Iterative Refinement (2025)
- SteeringTTA: Guiding Diffusion Trajectories for Robust Test-Time-Adaptation (2025)
- Understanding Sampler Stochasticity in Training Diffusion Models for RLHF (2025)
- ProxT2I: Efficient Reward-Guided Text-to-Image Generation via Proximal Diffusion (2025)
- Fine-Tuning Diffusion Models via Intermediate Distribution Shaping (2025)
- Reinforcing Diffusion Models by Direct Group Preference Optimization (2025)
- Proximal Diffusion Neural Sampler (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper