This repository contains the reference implementation of the Distributional Successor Measure, presented as a spotlight at ICML 2024.
A Distributional Analogue to the Successor Representation
by Harley Wiltzer* & Jesse Farebrother*, Arthur Gretton, Yunhao Tang, André Baretto, Will Dabney, Marc G. Bellemare, and Mark Rowland.
distributional-sr.mp4
The Distributional Successor Measure (DSM) a new approach for distributional reinforcement learning which proposes a clean separation of transition structure and reward in the learning process. Analogous to how the successor representation (SR) describes the expected consequences of behaving according to a given policy, our distributional successor measure describes the distributional consequences of this behaviour. This repository contains the code for learning a
This project makes heavy use of Jax, Flax, Optax, and Fiddle. We use pdm to manage our dependencies. With the lockfile pdm.lock
you should be able to faithfully instantiate the same environment we used to train our
pdm venv create
pdm install
If you're looking to build upon this project you might want to read more about how to use the Fiddle configuration library here: JesseFarebro/fiddle-demo.
The following command will train a policy on the desired environment before generating a dataset of transitions from the learned policy. For example,
python -m sr.scripts.make_dataset --env Pendulum-v1 --dataset_path datasets/pendulum/sac/dataset.pkl --policy_path datasets/pendulum/sac/policy
NOTE: The policy will be cached and if you don't specify the --force
flag it will skip the policy optimization step.
To train the
python -m dsm --workdir logdir
where logdir
will store checkpoints of the saved model. Plots of the learned return distributions and various metrics will be logged periodically throughout training. These plots and metrics can be found in the experiment tracker (defaults to Aim).
You can switch how the experiment is logged either using Weights & Biases or Aim with the flag --metric_writer {wandb, aim}
. Specific options for each of these methods can be configured via --wandb.{save_code,tags,name,group,mode}
and --aim.{repo=,experiment,log_system_params}
respectively.
To run the local Aim server you can simply run: pdm run aim
and then navigate to the provided URL.
If you build on our work or find it useful, please cite it using the following bibtex.
@inproceedings{wiltzer2024dsm,
title={A Distributional Analogue to the Successor Representation},
author={Harley Wiltzer and Jesse Farebrother and Arthur Gretton and Yunhao Tang and Andr\'e Barreto and Will Dabney and Marc G Bellemare and Mark Rowland},
booktitle = {International Conference on Machine Learning (ICML)},
year={2024},
}
This project is licensed under the MIT License - see the LICENSE.md file for details.