Recovering Hidden Reward in Diffusion-Based Policies
Abstract
EnergyFlow unifies generative action modeling with inverse reinforcement learning by parameterizing an energy function whose gradient serves as a denoising field, enabling reward extraction without adversarial training while improving policy generalization through structural constraints.
This paper introduces EnergyFlow, a framework that unifies generative action modeling with inverse reinforcement learning by parameterizing a scalar energy function whose gradient is the denoising field. We establish that under maximum-entropy optimality, the score function learned via denoising score matching recovers the gradient of the expert's soft Q-function, enabling reward extraction without adversarial training. Formally, we prove that constraining the learned field to be conservative reduces hypothesis complexity and tightens out-of-distribution generalization bounds. We further characterize the identifiability of recovered rewards and bound how score estimation errors propagate to action preferences. Empirically, EnergyFlow achieves state-of-the-art imitation performance on various manipulation tasks while providing an effective reward signal for downstream reinforcement learning that outperforms both adversarial IRL methods and likelihood-based alternatives. These results show that the structural constraints required for valid reward extraction simultaneously serve as beneficial inductive biases for policy generalization. The code is available at https://github.com/sotaagi/EnergyFlow.
Community
ENERGYFLOW unifies diffusion-based imitation learning and inverse reinforcement learning by learning a conservative energy field whose gradient drives action generation while exposing a recoverable reward signal, improving manipulation performance, downstream RL, and out-of-distribution robustness.
curious how robust the reward recovery is when the max-entropy assumption is only approximately satisfied in real expert data. the core move—using the energy gradient as the denoising field and tying it to the soft q gradient under a conservative energy—feels like a nice bridge between diffusion modeling and reward learning. i’m especially curious about the identifiability caveat: the paper notes a state-dependent offset that prevents global recovery; would a tiny learned centering term or a structured prior still preserve the conservative guarantee while yielding a recoverable global reward? the arxivlens breakdown helped me parse the energy-to-gradient mapping and the 1d temporal u-net choice; quick refresher here if others want the same vibe: https://arxivlens.com/PaperView/Details/recovering-hidden-reward-in-diffusion-based-policies-6944-7dae7923. it would be neat to see how this scales to truly high-dimensional action spaces or more severe distribution shifts, to test whether the integrability bias consistently improves out-of-sample generalization without trading off top performance.
Get this paper in your agent:
hf papers read 2605.00623 Don't have the latest CLI?
curl -LsSf https://hf.co/cli/install.sh | bash Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper