diff --git a/docs/rl-algorithms/ppo-isaacgymenvs.md b/docs/rl-algorithms/ppo-isaacgymenvs.md index 80a1eb3a8..dbec854fe 100644 --- a/docs/rl-algorithms/ppo-isaacgymenvs.md +++ b/docs/rl-algorithms/ppo-isaacgymenvs.md @@ -183,7 +183,7 @@ Additionally, `charts/consecutive_successes` means the number of consecutive epi To run benchmark experiments, see :material-github: [benchmark/ppo.sh](https://github.com/vwxyzjn/cleanrl/blob/master/benchmark/ppo.sh). Specifically, execute the following command: - + Below are the average episodic returns for `ppo_continuous_action_isaacgym.py`. To ensure the quality of the implementation, we compared the results against [Denys88/rl_games](https://github.com/Denys88/rl_games)' PPO and present the training time (units being `s (seconds), m (minutes)`). The hardware used is a NVIDIA RTX A6000 in a 24 core machine.