This is the implementation of SRMRL and this repository is based on garage.
To install locally, you will need to first install MuJoCo. Set LD_LIBRARY_PATH
to point to both the MuJoCo binaries (/$HOME/.mujoco/mujoco200/bin
) as well as the gpu drivers (something like /usr/lib/nvidia-390
, you can find your version by running nvidia-smi
).
Clone this repo and construct a virtual environment via pipenv install -r requirements.txt
. Then activate the virtual environment with pipenv shell
.
The implementation of SRMRL is placed in garage/torch/algos/srmrl.py
. Add the garage
package into your python path with export PYTHONPATH=.:$PYTHONPATH
in SRMRL directory.
To run the experiments, you can use the scripts in script
folder with
./script/[filename]
The script assumes you are in the SRMRL
directory. Or you can run other tasks using
python example/[filename].py --env_name=[env_name]
The output files including log file and model parameters will be placed in ./data/local/experiment/[EXP_NAME]
.
The output log file can be visualized with tensorboard.
The file progress.csv
contains statistics logged over the course of training.
We recommend viskit
for visualizing learning curves: https://github.com/vitchyr/viskit.
After training, use sim_policy.py
to visualize the learned policy:
python sim_policy.py --model_dir=[output_dir]
This script generate images and gif file of trajectories.