lunar_lander_ppo / replay.mp4
mustious7's picture
first PPO model, n_steps = 1024, batch_size = 64, n_epochs = 4,gamma = 0.999
88c379e
download
history contribute delete
158 kB