Skip to content

Latest commit

 

History

History
405 lines (323 loc) · 21.8 KB

README.md

File metadata and controls

405 lines (323 loc) · 21.8 KB

Awesome Diffusion Model in RL

Awesome docs visitor badge GitHub stars GitHub forks GitHub license

This is a collection of research papers for Diffusion Model in RL. And the repository will be continuously updated to track the frontier of Diffusion RL.

Welcome to follow and star!

Table of Contents

Overview of Diffusion Model in RL

The Diffusion Model in RL was introduced by “Planning with Diffusion for Flexible Behavior Synthesis” by Janner, Michael, et al. It casts trajectory optimization as a diffusion probabilistic model that plans by iteratively refining trajectories.

image info

There is another way: "Diffusion Policies as an Expressive Policy Class for Offline Reinforcement Learning" by Wang, Z. proposed Diffusion Model as policy-optimization in offline RL, et al. Specifically, Diffusion-QL forms policy as a conditional diffusion model with states as the condition from the offline policy-optimization perspective.

image info

Advantage

  1. Bypass the need for bootstrapping for long term credit assignment.
  2. Avoid undesirable short-sighted behaviors due to the discounting future rewards.
  3. Enjoy the diffusion models widely used in language and vision, which are easy to scale and adapt to multi-modal data.

Papers

format:
- [title](paper link) [links]
  - author1, author2, and author3...
  - publisher
  - key 
  - code 
  - experiment environment

Arxiv

ICML 2024

CVPR 2024

ICLR 2024

NeurIPS 2023

ICML 2023

ICLR 2023

ICRA 2023

NeurIPS 2022

ICML 2022

Contributing

Our purpose is to make this repo even better. If you are interested in contributing, please refer to HERE for instructions in contribution.

License

Awesome Diffusion Model in RL is released under the Apache 2.0 license.