Need help with PlaNet?
Click the “chat” button below for chat support from the developer who created it, or find similar developers for support.

About the developer

242 Stars 43 Forks MIT License 50 Commits 2 Opened issues


Deep Planning Network: Control from pixels by latent planning with learned dynamics

Services available


Need anything else?

Contributors list

No Data


MIT License

PlaNet: A Deep Planning Network for Reinforcement Learning [1]. Supports symbolic/visual observation spaces. Supports some Gym environments (including classic control/non-MuJoCo environments, so DeepMind Control Suite/MuJoCo are optional dependencies). Hyperparameters have been taken from the original work and are tuned for DeepMind Control Suite, so would need tuning for any other domains (such as the Gym environments).

Run with
. For best performance with DeepMind Control Suite, try setting environment variable
(see instructions and details here).

Results and pretrained models can be found in the releases.


To install all dependencies with Anaconda run

conda env create -f environment.yml
and use
source activate planet
to activate the environment.




[1] Learning Latent Dynamics for Planning from Pixels

We use cookies. If you continue to browse the site, you agree to the use of cookies. For more information on our use of cookies please see our Privacy Policy.