Commit ·
dfca58a
1
Parent(s): 01ac34e
Adding Lunar Lander Trained agent
Browse files- README.md +1 -1
- config.json +1 -1
- ppo-LunarLander-v2.zip +2 -2
- ppo-LunarLander-v2/data +18 -18
- ppo-LunarLander-v2/policy.optimizer.pth +1 -1
- ppo-LunarLander-v2/policy.pth +1 -1
- replay.mp4 +0 -0
- results.json +1 -1
README.md
CHANGED
|
@@ -16,7 +16,7 @@ model-index:
|
|
| 16 |
type: LunarLander-v2
|
| 17 |
metrics:
|
| 18 |
- type: mean_reward
|
| 19 |
-
value:
|
| 20 |
name: mean_reward
|
| 21 |
verified: false
|
| 22 |
---
|
|
|
|
| 16 |
type: LunarLander-v2
|
| 17 |
metrics:
|
| 18 |
- type: mean_reward
|
| 19 |
+
value: 251.87 +/- 17.44
|
| 20 |
name: mean_reward
|
| 21 |
verified: false
|
| 22 |
---
|
config.json
CHANGED
|
@@ -1 +1 @@
|
|
| 1 |
-
{"policy_class": {":type:": "<class 'abc.ABCMeta'>", ":serialized:": "gAWVOwAAAAAAAACMIXN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi5wb2xpY2llc5SMEUFjdG9yQ3JpdGljUG9saWN5lJOULg==", "__module__": "stable_baselines3.common.policies", "__doc__": "\n Policy class for actor-critic algorithms (has both policy and value prediction).\n Used by A2C, PPO and the likes.\n\n :param observation_space: Observation space\n :param action_space: Action space\n :param lr_schedule: Learning rate schedule (could be constant)\n :param net_arch: The specification of the policy and value networks.\n :param activation_fn: Activation function\n :param ortho_init: Whether to use or not orthogonal initialization\n :param use_sde: Whether to use State Dependent Exploration or not\n :param log_std_init: Initial value for the log standard deviation\n :param full_std: Whether to use (n_features x n_actions) parameters\n for the std instead of only (n_features,) when using gSDE\n :param use_expln: Use ``expln()`` function instead of ``exp()`` to ensure\n a positive standard deviation (cf paper). It allows to keep variance\n above zero and prevent it from growing too fast. In practice, ``exp()`` is usually enough.\n :param squash_output: Whether to squash the output using a tanh function,\n this allows to ensure boundaries when using gSDE.\n :param features_extractor_class: Features extractor to use.\n :param features_extractor_kwargs: Keyword arguments\n to pass to the features extractor.\n :param share_features_extractor: If True, the features extractor is shared between the policy and value networks.\n :param normalize_images: Whether to normalize images or not,\n dividing by 255.0 (True by default)\n :param optimizer_class: The optimizer to use,\n ``th.optim.Adam`` by default\n :param optimizer_kwargs: Additional keyword arguments,\n excluding the learning rate, to pass to the optimizer\n ", "__init__": "<function ActorCriticPolicy.__init__ at 0x7cdcf61ed5a0>", "_get_constructor_parameters": "<function ActorCriticPolicy._get_constructor_parameters at 0x7cdcf61ed630>", "reset_noise": "<function ActorCriticPolicy.reset_noise at 0x7cdcf61ed6c0>", "_build_mlp_extractor": "<function ActorCriticPolicy._build_mlp_extractor at 0x7cdcf61ed750>", "_build": "<function ActorCriticPolicy._build at 0x7cdcf61ed7e0>", "forward": "<function ActorCriticPolicy.forward at 0x7cdcf61ed870>", "extract_features": "<function ActorCriticPolicy.extract_features at 0x7cdcf61ed900>", "_get_action_dist_from_latent": "<function ActorCriticPolicy._get_action_dist_from_latent at 0x7cdcf61ed990>", "_predict": "<function ActorCriticPolicy._predict at 0x7cdcf61eda20>", "evaluate_actions": "<function ActorCriticPolicy.evaluate_actions at 0x7cdcf61edab0>", "get_distribution": "<function ActorCriticPolicy.get_distribution at 0x7cdcf61edb40>", "predict_values": "<function ActorCriticPolicy.predict_values at 0x7cdcf61edbd0>", "__abstractmethods__": "frozenset()", "_abc_impl": "<_abc._abc_data object at 0x7cdcf635d1c0>"}, "verbose": 1, "policy_kwargs": {}, "num_timesteps": 1015808, "_total_timesteps": 1000500, "_num_timesteps_at_start": 0, "seed": null, "action_noise": null, "start_time": 1703594279421748573, "learning_rate": 0.0003, "tensorboard_log": null, "_last_obs": {":type:": "<class 'numpy.ndarray'>", ":serialized:": "gAWVdQIAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJYAAgAAAAAAABptFD735W8/hfhaPvK77b6oq3E+Hm37vAAAAAAAAAAA85tvPmJPQT9TgHM9hS3Yvlctij7NQ+69AAAAAAAAAAAzkfw8uNvNu2+3JT0x0II9BGU2vQWU7zsAAIA/AACAPwB8drw9IHO7PWFVPH1ZhzzuwdW8gnZoPQAAgD8AAIA/zY4uPCYsvT8iIeM8nymcvn0gYjze6wk9AAAAAAAAAAANjdi99k5rPdJeKj7/LpG+r7InPM5EeT0AAAAAAAAAADO07r3sFZ4/urDsvolGCb9WtCC+eyUxvgAAAAAAAAAAAGS6PEg3nroyuoG1IqwyrgKD6jm2m7k0AACAPwAAgD8zQ6k7Kdx/utO0nrNQfGcuzyL/ugrOsTMAAIA/AACAPwDP+L0pWE+6+NQfOXM4jTTm0sU5XbQ5uAAAgD8AAIA/2mPJvge2dj96rLy+mfoUv/jNsb5qM7k9AAAAAAAAAAAAI9m9J7UGPiWiBj4M8Ui+1S4Dvdo57zwAAAAAAAAAAGbeyrtcE026RrzNOE+gyTO73+o3mMzytwAAgD8AAIA/Gm3UPTulrz5aVV6+maKxvihH0r3a6Wu9AAAAAAAAAADmuaS9bxKWPvGdhD05cZ6+g5MCPGBHqj0AAAAAAAAAAKbngr2txB4+af4oPZKKT76ZYik7QTM6uwAAAAAAAAAAlIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUiYiHlFKUKEsDjAE8lE5OTkr/////Sv////9LAHSUYksQSwiGlIwBQ5R0lFKULg=="}, "_last_episode_starts": {":type:": "<class 'numpy.ndarray'>", ":serialized:": "gAWVgwAAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJYQAAAAAAAAAAAAAAAAAAAAAAAAAAAAAACUjAVudW1weZSMBWR0eXBllJOUjAJiMZSJiIeUUpQoSwOMAXyUTk5OSv////9K/////0sAdJRiSxCFlIwBQ5R0lFKULg=="}, "_last_original_obs": null, "_episode_num": 0, "use_sde": false, "sde_sample_freq": -1, "_current_progress_remaining": -0.015300349825087434, "_stats_window_size": 100, "ep_info_buffer": {":type:": "<class 'collections.deque'>", ":serialized:": "gAWVKAwAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKUKH2UKIwBcpRHQHAB++IuXeGMAWyUTRMBjAF0lEdAkdGMEidJ8XV9lChoBkdAcHZnDiwSrmgHTUQBaAhHQJHUQxQBPsR1fZQoaAZHQHF1wSSNfgJoB00XAWgIR0CR1ZY/mknDdX2UKGgGR0BELFvQ4S6EaAdL0WgIR0CR1LnjABT5dX2UKGgGR0BuMCr/82rGaAdNIAFoCEdAkdYj41xbS3V9lChoBkdAbtgNFSbYsmgHTQQBaAhHQJHWe4/eLvV1fZQoaAZHQGAOq4YrJ8xoB03oA2gIR0CR1sRwIdELdX2UKGgGR0BygHDAJswdaAdNDAFoCEdAkdfOVX3g1nV9lChoBkdAcBwfOD8Lr2gHS/VoCEdAkdgnR5TqB3V9lChoBkdActKlANXo1WgHTToBaAhHQJHZxLUTcqR1fZQoaAZHQHGHfcFhXsBoB00fAWgIR0CR2t86V+qjdX2UKGgGR0BfbU4BFNL2aAdN6ANoCEdAkdny+QEIPnV9lChoBkdAcIIeZof0VmgHTQIBaAhHQJHZ96mfoRt1fZQoaAZHQG8+fSH/LkloB0vdaAhHQJHbWGQCCBh1fZQoaAZHQHOGrSuyNXJoB000AWgIR0CR2wQokRjCdX2UKGgGR0BheCWiUPhAaAdN6ANoCEdAkdxxQzk6tHV9lChoBkdAcLKa4c3l0mgHS+poCEdAkdzTEehf0HV9lChoBkdAc5b+RHPNV2gHTX8BaAhHQJHd4XYUWVN1fZQoaAZHQFmfUqQRwqBoB03oA2gIR0CR3eKVY6n0dX2UKGgGR0BwEkb2lEZ0aAdL5GgIR0CR3nQwsXizdX2UKGgGR0BwehDKHO8kaAdNPwFoCEdAkd+hEroW6HV9lChoBkdAcx7JtSAH3WgHTSABaAhHQJHe6dsi0OV1fZQoaAZHQHDKoRywOe9oB0v7aAhHQJHfXf1pTMt1fZQoaAZHQHKCN+LFXJZoB0vsaAhHQJHg4cOskpt1fZQoaAZHQG7AOhkAggZoB0vyaAhHQJHitnOB19x1fZQoaAZHQHC4ebNKRMhoB00OAWgIR0CR4kvF3pwCdX2UKGgGR0BxIjZh8YygaAdNEQFoCEdAkeNY0/GEPHV9lChoBkdAcmnZaV2RrGgHTS0BaAhHQJHjX4oJAt51fZQoaAZHQHAwLcCYCyRoB00qAWgIR0CR5OvQ4S6EdX2UKGgGR0Bxe63d9Dx9aAdNFgFoCEdAkeZdZvDP4XV9lChoBkdAcR8m78Nx2mgHTQkBaAhHQJHnaUPhAGB1fZQoaAZHQHJEgCCBf8doB01iAWgIR0CR6QZKWcBmdX2UKGgGR0Bv/7hHbypaaAdL8GgIR0CR6I88cMmXdX2UKGgGR0Bx3T6YVqN7aAdNIQFoCEdAkekYKx9oe3V9lChoBkdAb05depn6EmgHTRYBaAhHQJHq11PnB+F1fZQoaAZHQHKE4h+vyLBoB00kAWgIR0CR6i8VHnU2dX2UKGgGR0Bx+T2WY4Q0aAdNKAFoCEdAkevDbFjur3V9lChoBkdAb2FmJWNm2GgHS95oCEdAkeyWVzIV/XV9lChoBkdAcGXMxXXAdmgHS/9oCEdAke3YqTbFj3V9lChoBkdAcA5GxUvPC2gHTTYBaAhHQJHvJas6q811fZQoaAZHQHJ9YZQ53khoB00xAWgIR0CR7nlK9PDYdX2UKGgGR0BvMvNiYsunaAdNZAFoCEdAke8HmaH9FXV9lChoBkdAcTsAZKnNxGgHTQkBaAhHQJHvOxcE/0N1fZQoaAZHQHC/EJfICEJoB0vyaAhHQJHwmLZSNwR1fZQoaAZHQG/pBZQpF1BoB00gAWgIR0CR8T1UlzEKdX2UKGgGR0Bxumj9GZuyaAdNCAFoCEdAkgNGus90R3V9lChoBkdAcRMrEtNBW2gHS+hoCEdAkgLLlq8DjnV9lChoBkdAcP8JzT4L1GgHTRMBaAhHQJIDCG0u14R1fZQoaAZHQHGqa7EpAlhoB00YAWgIR0CSA4UXHim3dX2UKGgGR0BxiphH9WIXaAdNDgFoCEdAkgS7iyY5UHV9lChoBkdAcga2xptaZGgHTSoBaAhHQJIGLN3W4Ex1fZQoaAZHQG5QgKWszVNoB00KAWgIR0CSBzvoNd7fdX2UKGgGR0BgP/m1YyO8aAdN6ANoCEdAkge/IKc/dXV9lChoBkdAbjgam4y44WgHTQ0BaAhHQJIIto6CDmN1fZQoaAZHQG/3sPjGT9toB01GAWgIR0CSCAqbSZ0CdX2UKGgGR0ByGCj9GZuyaAdL7mgIR0CSCTLlmvnsdX2UKGgGR0BrHbjHXEqEaAdNvwNoCEdAkgiFSKm8/XV9lChoBkdAcXJs67ulXWgHTTQBaAhHQJIJNxrBTGZ1fZQoaAZHQG7N+/5+H8FoB00HAWgIR0CSCo1QqI8AdX2UKGgGR0BxiAqslsxgaAdNRwFoCEdAkgpE52hZhnV9lChoBkdAb6/XQMQVbmgHTRIBaAhHQJIL9Zr56+p1fZQoaAZHQHFP9MsYl6ZoB00HAWgIR0CSCx2JSBK+dX2UKGgGR0ByERvitJWeaAdL/2gIR0CSCxlD4QBgdX2UKGgGR0Bw3ngXMyJsaAdNGQFoCEdAkgwxCY1HfHV9lChoBkdATpt/4Irvs2gHS85oCEdAkg1h5cC5mXV9lChoBkdAbc+LNwBHTmgHTSQBaAhHQJINqQRwqAl1fZQoaAZHQHEI52ECeVdoB0vxaAhHQJIOqhGpdbB1fZQoaAZHQGz5zErGza9oB00AAWgIR0CSEBpSrHU+dX2UKGgGR0Bx5PPdEb5uaAdL+WgIR0CSDyxNZeRgdX2UKGgGR0BxWa6e5Fw2aAdL9WgIR0CSEDR64UeudX2UKGgGR0BstdHnU2DQaAdNMwFoCEdAkg9OQMhHLHV9lChoBkdAcYbWnTAnD2gHS/poCEdAkhGcxbjcVXV9lChoBkdAcOuXWOIZZWgHTR0BaAhHQJIRVBRhttR1fZQoaAZHQG84mI9C/oJoB004AWgIR0CSEWXzDn/2dX2UKGgGR0Bx770voNd7aAdL/WgIR0CSEk4LThHcdX2UKGgGR0BxvPQkX1rZaAdNBgFoCEdAkhKgFTvRZ3V9lChoBkdAcRtqMm4RVmgHTRABaAhHQJITwpLEk0J1fZQoaAZHQHB1zv7WNFVoB00vAWgIR0CSEvWl/H5rdX2UKGgGR0BTQME7nxJ/aAdLs2gIR0CSFNCNCJGfdX2UKGgGR0BwiRXnyNGWaAdNGgFoCEdAkhTdEw35vnV9lChoBkdAcMYf5k9U0mgHTSQBaAhHQJIXEbMotth1fZQoaAZHQHAGb4nF5v9oB0vxaAhHQJIWyivgWJt1fZQoaAZHQG88dv0h/y5oB0vraAhHQJIYOgoPTXt1fZQoaAZHQG/C9yLhrFhoB00OAWgIR0CSGYQv6CUYdX2UKGgGR0BwrHO0LMLXaAdNEgFoCEdAkhjvEXLvC3V9lChoBkdAcKiq1gH/tWgHS+9oCEdAkhqHwTdtVXV9lChoBkdAOjM6/7BO6GgHS9JoCEdAkhsFtoBaLXV9lChoBkdAcAW1AJLM92gHTSMBaAhHQJIcqLbYbsF1fZQoaAZHQHAaW+9Jz1doB0vyaAhHQJIb8wpON5t1fZQoaAZHQHK+FMRHww1oB0v/aAhHQJIeJE+gUUR1fZQoaAZHQHGB+P7vXshoB00+AWgIR0CSHdAUcn3MdX2UKGgGR0BvWSbx3FDOaAdN5wFoCEdAkh7KBNEgGXV9lChoBkdAbq6/3WWhRWgHTRgBaAhHQJIeaxUvPC51fZQoaAZHQHEWpof0VahoB00HAWgIR0CSH5bWEsasdX2UKGgGR0ByG7TfBN21aAdNCgFoCEdAkh+xiw0O3HV9lChoBkdAYe1G9YfW+WgHTegDaAhHQJIhXIS13MZ1fZQoaAZHQHEFaCxu89RoB00PAWgIR0CSImLQXyiFdX2UKGgGR0ByiZQ3xWkraAdNOwFoCEdAkiLM1CPZI3V9lChoBkdAcjXM+u/1x2gHTQwBaAhHQJIjKVdHDrJ1fZQoaAZHQG/gpjlPrOZoB01BAWgIR0CSInfmcOLBdWUu"}, "ep_success_buffer": {":type:": "<class 'collections.deque'>", ":serialized:": "gAWVIAAAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKULg=="}, "_n_updates": 248, "observation_space": {":type:": "<class 'gymnasium.spaces.box.Box'>", ":serialized:": "gAWVcAIAAAAAAACMFGd5bW5hc2l1bS5zcGFjZXMuYm94lIwDQm94lJOUKYGUfZQojAVkdHlwZZSMBW51bXB5lGgFk5SMAmY0lImIh5RSlChLA4wBPJROTk5K/////0r/////SwB0lGKMDWJvdW5kZWRfYmVsb3eUjBJudW1weS5jb3JlLm51bWVyaWOUjAtfZnJvbWJ1ZmZlcpSTlCiWCAAAAAAAAAABAQEBAQEBAZRoB4wCYjGUiYiHlFKUKEsDjAF8lE5OTkr/////Sv////9LAHSUYksIhZSMAUOUdJRSlIwNYm91bmRlZF9hYm92ZZRoECiWCAAAAAAAAAABAQEBAQEBAZRoFEsIhZRoGHSUUpSMBl9zaGFwZZRLCIWUjANsb3eUaBAoliAAAAAAAAAAAAC0wgAAtMIAAKDAAACgwNsPScAAAKDAAAAAgAAAAICUaApLCIWUaBh0lFKUjARoaWdolGgQKJYgAAAAAAAAAAAAtEIAALRCAACgQAAAoEDbD0lAAACgQAAAgD8AAIA/lGgKSwiFlGgYdJRSlIwIbG93X3JlcHKUjFtbLTkwLiAgICAgICAgLTkwLiAgICAgICAgIC01LiAgICAgICAgIC01LiAgICAgICAgIC0zLjE0MTU5MjcgIC01LgogIC0wLiAgICAgICAgIC0wLiAgICAgICBdlIwJaGlnaF9yZXBylIxTWzkwLiAgICAgICAgOTAuICAgICAgICAgNS4gICAgICAgICA1LiAgICAgICAgIDMuMTQxNTkyNyAgNS4KICAxLiAgICAgICAgIDEuICAgICAgIF2UjApfbnBfcmFuZG9tlE51Yi4=", "dtype": "float32", "bounded_below": "[ True True True True True True True True]", "bounded_above": "[ True True True True True True True True]", "_shape": [8], "low": "[-90. -90. -5. -5. -3.1415927 -5.\n -0. -0. ]", "high": "[90. 90. 5. 5. 3.1415927 5.\n 1. 1. ]", "low_repr": "[-90. -90. -5. -5. -3.1415927 -5.\n -0. -0. ]", "high_repr": "[90. 90. 5. 5. 3.1415927 5.\n 1. 1. ]", "_np_random": null}, "action_space": {":type:": "<class 'gymnasium.spaces.discrete.Discrete'>", ":serialized:": "gAWV1QAAAAAAAACMGWd5bW5hc2l1bS5zcGFjZXMuZGlzY3JldGWUjAhEaXNjcmV0ZZSTlCmBlH2UKIwBbpSMFW51bXB5LmNvcmUubXVsdGlhcnJheZSMBnNjYWxhcpSTlIwFbnVtcHmUjAVkdHlwZZSTlIwCaTiUiYiHlFKUKEsDjAE8lE5OTkr/////Sv////9LAHSUYkMIBAAAAAAAAACUhpRSlIwFc3RhcnSUaAhoDkMIAAAAAAAAAACUhpRSlIwGX3NoYXBllCloCmgOjApfbnBfcmFuZG9tlE51Yi4=", "n": "4", "start": "0", "_shape": [], "dtype": "int64", "_np_random": null}, "n_envs": 16, "n_steps": 1024, "gamma": 0.999, "gae_lambda": 0.98, "ent_coef": 0.01, "vf_coef": 0.5, "max_grad_norm": 0.5, "batch_size": 64, "n_epochs": 4, "clip_range": {":type:": "<class 'function'>", ":serialized:": "gAWVxQIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwOX21ha2VfZnVuY3Rpb26Uk5QoaACMDV9idWlsdGluX3R5cGWUk5SMCENvZGVUeXBllIWUUpQoSwFLAEsASwFLAUsTQwSIAFMAlE6FlCmMAV+UhZSMSS91c3IvbG9jYWwvbGliL3B5dGhvbjMuMTAvZGlzdC1wYWNrYWdlcy9zdGFibGVfYmFzZWxpbmVzMy9jb21tb24vdXRpbHMucHmUjARmdW5jlEuEQwIEAZSMA3ZhbJSFlCl0lFKUfZQojAtfX3BhY2thZ2VfX5SMGHN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbpSMCF9fbmFtZV9flIwec3RhYmxlX2Jhc2VsaW5lczMuY29tbW9uLnV0aWxzlIwIX19maWxlX1+UjEkvdXNyL2xvY2FsL2xpYi9weXRob24zLjEwL2Rpc3QtcGFja2FnZXMvc3RhYmxlX2Jhc2VsaW5lczMvY29tbW9uL3V0aWxzLnB5lHVOTmgAjBBfbWFrZV9lbXB0eV9jZWxslJOUKVKUhZR0lFKUjBxjbG91ZHBpY2tsZS5jbG91ZHBpY2tsZV9mYXN0lIwSX2Z1bmN0aW9uX3NldHN0YXRllJOUaB99lH2UKGgWaA2MDF9fcXVhbG5hbWVfX5SMGWNvbnN0YW50X2ZuLjxsb2NhbHM+LmZ1bmOUjA9fX2Fubm90YXRpb25zX1+UfZSMDl9fa3dkZWZhdWx0c19flE6MDF9fZGVmYXVsdHNfX5ROjApfX21vZHVsZV9flGgXjAdfX2RvY19flE6MC19fY2xvc3VyZV9flGgAjApfbWFrZV9jZWxslJOURz/JmZmZmZmahZRSlIWUjBdfY2xvdWRwaWNrbGVfc3VibW9kdWxlc5RdlIwLX19nbG9iYWxzX1+UfZR1hpSGUjAu"}, "clip_range_vf": null, "normalize_advantage": true, "target_kl": null, "lr_schedule": {":type:": "<class 'function'>", ":serialized:": "gAWVxQIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwOX21ha2VfZnVuY3Rpb26Uk5QoaACMDV9idWlsdGluX3R5cGWUk5SMCENvZGVUeXBllIWUUpQoSwFLAEsASwFLAUsTQwSIAFMAlE6FlCmMAV+UhZSMSS91c3IvbG9jYWwvbGliL3B5dGhvbjMuMTAvZGlzdC1wYWNrYWdlcy9zdGFibGVfYmFzZWxpbmVzMy9jb21tb24vdXRpbHMucHmUjARmdW5jlEuEQwIEAZSMA3ZhbJSFlCl0lFKUfZQojAtfX3BhY2thZ2VfX5SMGHN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbpSMCF9fbmFtZV9flIwec3RhYmxlX2Jhc2VsaW5lczMuY29tbW9uLnV0aWxzlIwIX19maWxlX1+UjEkvdXNyL2xvY2FsL2xpYi9weXRob24zLjEwL2Rpc3QtcGFja2FnZXMvc3RhYmxlX2Jhc2VsaW5lczMvY29tbW9uL3V0aWxzLnB5lHVOTmgAjBBfbWFrZV9lbXB0eV9jZWxslJOUKVKUhZR0lFKUjBxjbG91ZHBpY2tsZS5jbG91ZHBpY2tsZV9mYXN0lIwSX2Z1bmN0aW9uX3NldHN0YXRllJOUaB99lH2UKGgWaA2MDF9fcXVhbG5hbWVfX5SMGWNvbnN0YW50X2ZuLjxsb2NhbHM+LmZ1bmOUjA9fX2Fubm90YXRpb25zX1+UfZSMDl9fa3dkZWZhdWx0c19flE6MDF9fZGVmYXVsdHNfX5ROjApfX21vZHVsZV9flGgXjAdfX2RvY19flE6MC19fY2xvc3VyZV9flGgAjApfbWFrZV9jZWxslJOURz8zqSowVTJhhZRSlIWUjBdfY2xvdWRwaWNrbGVfc3VibW9kdWxlc5RdlIwLX19nbG9iYWxzX1+UfZR1hpSGUjAu"}, "system_info": {"OS": "Linux-6.1.58+-x86_64-with-glibc2.35 # 1 SMP PREEMPT_DYNAMIC Sat Nov 18 15:31:17 UTC 2023", "Python": "3.10.12", "Stable-Baselines3": "2.0.0a5", "PyTorch": "2.1.0+cu121", "GPU Enabled": "True", "Numpy": "1.23.5", "Cloudpickle": "2.2.1", "Gymnasium": "0.28.1", "OpenAI Gym": "0.25.2"}}
|
|
|
|
| 1 |
+
{"policy_class": {":type:": "<class 'abc.ABCMeta'>", ":serialized:": "gAWVOwAAAAAAAACMIXN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi5wb2xpY2llc5SMEUFjdG9yQ3JpdGljUG9saWN5lJOULg==", "__module__": "stable_baselines3.common.policies", "__doc__": "\n Policy class for actor-critic algorithms (has both policy and value prediction).\n Used by A2C, PPO and the likes.\n\n :param observation_space: Observation space\n :param action_space: Action space\n :param lr_schedule: Learning rate schedule (could be constant)\n :param net_arch: The specification of the policy and value networks.\n :param activation_fn: Activation function\n :param ortho_init: Whether to use or not orthogonal initialization\n :param use_sde: Whether to use State Dependent Exploration or not\n :param log_std_init: Initial value for the log standard deviation\n :param full_std: Whether to use (n_features x n_actions) parameters\n for the std instead of only (n_features,) when using gSDE\n :param use_expln: Use ``expln()`` function instead of ``exp()`` to ensure\n a positive standard deviation (cf paper). It allows to keep variance\n above zero and prevent it from growing too fast. In practice, ``exp()`` is usually enough.\n :param squash_output: Whether to squash the output using a tanh function,\n this allows to ensure boundaries when using gSDE.\n :param features_extractor_class: Features extractor to use.\n :param features_extractor_kwargs: Keyword arguments\n to pass to the features extractor.\n :param share_features_extractor: If True, the features extractor is shared between the policy and value networks.\n :param normalize_images: Whether to normalize images or not,\n dividing by 255.0 (True by default)\n :param optimizer_class: The optimizer to use,\n ``th.optim.Adam`` by default\n :param optimizer_kwargs: Additional keyword arguments,\n excluding the learning rate, to pass to the optimizer\n ", "__init__": "<function ActorCriticPolicy.__init__ at 0x7e12e3cdb6d0>", "_get_constructor_parameters": "<function ActorCriticPolicy._get_constructor_parameters at 0x7e12e3cdb760>", "reset_noise": "<function ActorCriticPolicy.reset_noise at 0x7e12e3cdb7f0>", "_build_mlp_extractor": "<function ActorCriticPolicy._build_mlp_extractor at 0x7e12e3cdb880>", "_build": "<function ActorCriticPolicy._build at 0x7e12e3cdb910>", "forward": "<function ActorCriticPolicy.forward at 0x7e12e3cdb9a0>", "extract_features": "<function ActorCriticPolicy.extract_features at 0x7e12e3cdba30>", "_get_action_dist_from_latent": "<function ActorCriticPolicy._get_action_dist_from_latent at 0x7e12e3cdbac0>", "_predict": "<function ActorCriticPolicy._predict at 0x7e12e3cdbb50>", "evaluate_actions": "<function ActorCriticPolicy.evaluate_actions at 0x7e12e3cdbbe0>", "get_distribution": "<function ActorCriticPolicy.get_distribution at 0x7e12e3cdbc70>", "predict_values": "<function ActorCriticPolicy.predict_values at 0x7e12e3cdbd00>", "__abstractmethods__": "frozenset()", "_abc_impl": "<_abc._abc_data object at 0x7e12e3e77440>"}, "verbose": 1, "policy_kwargs": {}, "num_timesteps": 1015808, "_total_timesteps": 1001000, "_num_timesteps_at_start": 0, "seed": null, "action_noise": null, "start_time": 1703800446569541592, "learning_rate": 0.0003, "tensorboard_log": null, "_last_obs": {":type:": "<class 'numpy.ndarray'>", ":serialized:": "gAWVdQIAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJYAAgAAAAAAAE1Mq72vDyc+mJe9PQW5Gr6aWwm9n2BJPQAAAAAAAAAAra1YvjQqTj9+myK+BPrkvmCUMb5ztjA8AAAAAAAAAACaTya+vA51P4Jkyr7i9f6+wTRKvrcHML4AAAAAAAAAAMYHBL40iZ8+zKAXPiGxfb5ZjAS9oPoLvQAAAAAAAAAAYK+CPjIHTT/6kC8+YGSsvu/+iT7enb67AAAAAAAAAACA27E9j3Y5ur08f7rfIce1tDQ7unDkkTkAAIA/AACAP2YwIT3hsu05kCF4PC6OSbz+wZe64txqPAAAAAAAAAAAgHsrvrQnYD5mVqs+/i9mvhEwaj1G1UY9AAAAAAAAAACzHz894Wyxutt68TXxp+gwy0Unuoq5GrUAAIA/AACAP0Cohr6uT2E/5dKEveGf3b70yEi+RQ2iPQAAAAAAAAAAowuAPsMuXT85wTI+XUPSvsergz5XbaK8AAAAAAAAAADNlOu7b05TP4iqn7tQYfW+nqoHvEpcgb0AAAAAAAAAABpTPj2PJmW6iOSHOM3MhTPj6u65tI+ftwAAgD8AAIA/zQiuvRjIij9m1yS+h3P3vrHH370GIpu7AAAAAAAAAAA6ghI+Z1vHPqvLkDuAuWW+Zm37PMVkrjwAAAAAAAAAANNiVz7hHvy8sC8vOwuwy7nTBGG+zqhsugAAgD8AAIA/lIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUiYiHlFKUKEsDjAE8lE5OTkr/////Sv////9LAHSUYksQSwiGlIwBQ5R0lFKULg=="}, "_last_episode_starts": {":type:": "<class 'numpy.ndarray'>", ":serialized:": "gAWVgwAAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJYQAAAAAAAAAAAAAAAAAAAAAAAAAAAAAACUjAVudW1weZSMBWR0eXBllJOUjAJiMZSJiIeUUpQoSwOMAXyUTk5OSv////9K/////0sAdJRiSxCFlIwBQ5R0lFKULg=="}, "_last_original_obs": null, "_episode_num": 0, "use_sde": false, "sde_sample_freq": -1, "_current_progress_remaining": -0.014793206793206837, "_stats_window_size": 100, "ep_info_buffer": {":type:": "<class 'collections.deque'>", ":serialized:": "gAWVJAwAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKUKH2UKIwBcpRHQG+HrVWjoISMAWyUTQ0BjAF0lEdAsI90JTl1bXV9lChoBkdAcIC7L+xW1mgHS/JoCEdAsI91as6q83V9lChoBkdAcmdYUnG83GgHTSABaAhHQLCPdY7aIvd1fZQoaAZHQHBAwtrbg0loB00aAWgIR0Cwj3oPXkHVdX2UKGgGR0Bx6Httygf2aAdNOQFoCEdAsI+gVLzwt3V9lChoBkdAcj1U/OdGzGgHTRABaAhHQLCP6L6k6911fZQoaAZHQHFqL6ciGFloB0vuaAhHQLCQGO58Sf11fZQoaAZHQHCGozi0fHRoB00bAWgIR0CwkB6Df3vhdX2UKGgGR0ByBYQ04zacaAdL+WgIR0CwkCGitaIOdX2UKGgGR0Bxypy6tknUaAdL8WgIR0CwkFGo3rD7dX2UKGgGR0BwNNwT/Q0GaAdL9GgIR0CwkFdHhCMQdX2UKGgGR0BwObbmEGqxaAdL9mgIR0CwkGkgjhUBdX2UKGgGR0BwxfenAIppaAdNCwFoCEdAsJDqy2QXAXV9lChoBkdAcITvFFUhm2gHTUsBaAhHQLCRKEkB0ZF1fZQoaAZHQHH72n4wh4doB002AWgIR0CwkTwc5sCUdX2UKGgGR0BtQBGlQ/HHaAdNCAFoCEdAsJF7fzjFQ3V9lChoBkdAbDdNLUTcqWgHTSUBaAhHQLCRytFKCg91fZQoaAZHQG9P+fqX4TNoB00QAWgIR0CwkdQdKdxydX2UKGgGR0BwqUxHoX9BaAdNKwFoCEdAsJHUKQaJh3V9lChoBkdAcrK+JgsshGgHTSwBaAhHQLCR1/yXlbN1fZQoaAZHQG4OdXcQAdZoB01VAWgIR0CwkeiL/CIldX2UKGgGR0Bx/7FbVz6raAdL6WgIR0CwkhR6KLsKdX2UKGgGR0BuPLLyMDOkaAdNBQFoCEdAsJIc6IWP93V9lChoBkdAcW9oLG7z1GgHTQkBaAhHQLCScqLjxTd1fZQoaAZHQHAt96Tnq3VoB00ZAWgIR0CwkpfugHu7dX2UKGgGR0Bw/zhDPWxyaAdNBQFoCEdAsJLVxYJVsHV9lChoBkdAcIKJm/WUbGgHTToBaAhHQLCTOWw/xDt1fZQoaAZHQHK1vqcEvCdoB005AWgIR0Cwk0Ajlgc+dX2UKGgGR0Bxibru6VdHaAdL7mgIR0Cwk3v1HvtudX2UKGgGR0BwhSA4GUwBaAdNIAFoCEdAsJOyAG0NSnV9lChoBkdAcoFqIrOJL2gHTQABaAhHQLCTwvCMxXZ1fZQoaAZHQHD1jEaVD8doB0v6aAhHQLCT/3MINVl1fZQoaAZHQHIy8D8tPHloB0vkaAhHQLCUF4z7/GV1fZQoaAZHQHDnhn8KohpoB00HAWgIR0CwlHaYeDFqdX2UKGgGR0BwbPe67NB4aAdNBgFoCEdAsJSQg1WKdnV9lChoBkdAcZsaURnOB2gHTRcBaAhHQLCUo90ihWZ1fZQoaAZHQHGulIRRMvhoB00nAWgIR0CwlNKkAPupdX2UKGgGR0BuZqFEiMYNaAdL7WgIR0CwlNvC2tuDdX2UKGgGR0BwRaRaHKwIaAdNEwFoCEdAsJToemvW6XV9lChoBkdAcPyC04R282gHTQUBaAhHQLCVQbB42TB1fZQoaAZHQHN3mMS9M9NoB00hAWgIR0CwlaIrvsqsdX2UKGgGR0BxLCB/ZuhsaAdNFwFoCEdAsJo0i+tbLXV9lChoBkdAbwJUe+23KGgHTS4BaAhHQLCaaUoKD011fZQoaAZHQHMjGuDBdldoB00jAWgIR0CwmoTvy9VWdX2UKGgGR0Btrj2xptaZaAdNvAFoCEdAsJqjTiKiwnV9lChoBkdAUZ7+zdDYy2gHS7loCEdAsJrC2a2F4HV9lChoBkdAcoYWv8qFy2gHTQ8BaAhHQLCaxvG6wt91fZQoaAZHQHCtNZaFEiNoB00MAWgIR0CwmtO0svqUdX2UKGgGR0Bw0GIeo1k2aAdL82gIR0Cwmv88YAKfdX2UKGgGR0BwewoOQQtjaAdNSwFoCEdAsJsTpfQa73V9lChoBkdAcwVz8gpz92gHTRoBaAhHQLCbO3o9s8B1fZQoaAZHQHGicIqslsxoB01sAWgIR0Cwm0XzQNTcdX2UKGgGR0BurCSowVTKaAdL+GgIR0Cwm0mkN4JNdX2UKGgGR0BtOGU0Nz8xaAdNAAFoCEdAsJtJnZkCm3V9lChoBkdAbvuMJhOQAGgHTSIBaAhHQLCbY2FFlTZ1fZQoaAZHQHIgg9FF2FFoB00FAWgIR0Cwm5HLRrrPdX2UKGgGR0Bx6/8uSOinaAdNEgFoCEdAsJwAtPHktHV9lChoBkdAbzZTsIE8rGgHS/poCEdAsJwJ/oaDPHV9lChoBkdAcawDxLCemWgHS+RoCEdAsJxGhew9q3V9lChoBkdAcH/2MKkVOGgHS+toCEdAsJx1H+ZPVXV9lChoBkdAcO5fVqesgmgHTRsBaAhHQLCcpaZQYUF1fZQoaAZHQHFYo9HMEA5oB0v3aAhHQLCcpdLQHA11fZQoaAZHQG4E09QoCuFoB00VAWgIR0CwnN2OIZZTdX2UKGgGR0BxoKTLW7OFaAdL42gIR0CwnO9tALRbdX2UKGgGR0BwIyHO8kD7aAdNVgFoCEdAsJ0NM8HObHV9lChoBkdASgcRJ2+wkmgHS8BoCEdAsJ0aCXhOxnV9lChoBkdAcNm3iJfplmgHTQ8BaAhHQLCdIY4Qz1t1fZQoaAZHQHDZ2Ef1YhdoB0v4aAhHQLCdKRgZ0jl1fZQoaAZHQHHBZyhi9ZloB00kAWgIR0CwnTZb2USqdX2UKGgGR0BuRvXumaYvaAdNKAFoCEdAsJ2fybx3FHV9lChoBkdAcU/05EMLGGgHTUQBaAhHQLCdtEjPfKp1fZQoaAZHQG4DU6xPfsNoB0v+aAhHQLCeAuSfUWl1fZQoaAZHQGwJZh8YyftoB0v+aAhHQLCeUg4Otnx1fZQoaAZHQHFQaC6H0shoB01IAWgIR0CwnrpaA4GVdX2UKGgGR0BwhhmapgkUaAdNCgFoCEdAsJ7UGUwBYHV9lChoBkdAcy8w5/9YOmgHS/RoCEdAsJ7YnlXA/XV9lChoBkdAc3s8r7O3UmgHTSwBaAhHQLCe88WbgCR1fZQoaAZHQHFQSiqQzUJoB0vvaAhHQLCe/lY2bXp1fZQoaAZHQHH3ZHuqm0poB00NAWgIR0Cwnx+rdWQwdX2UKGgGR0BxFcn/kvK2aAdNQAFoCEdAsJ9O6NEPUnV9lChoBkdAcA6fO2RaHWgHTScBaAhHQLCftGyX2M91fZQoaAZHQHCy6TOgQH1oB0v0aAhHQLCf4gcLjPx1fZQoaAZHQHIB2Ef1YhdoB006AWgIR0CwoAPfsNUgdX2UKGgGR0BuXTQeFL39aAdNTwFoCEdAsKAdyp71I3V9lChoBkdAbSs8+zMRpWgHTQ0BaAhHQLCgSggow251fZQoaAZHQHF4w7YChexoB0v4aAhHQLCge71qWTp1fZQoaAZHQHBbd92HLzRoB016AWgIR0CwoIbILgGbdX2UKGgGR0BwBx4A0bcXaAdL/WgIR0CwoONHc1wYdX2UKGgGR0BxWWBGx2SuaAdL6WgIR0CwoUJPVNHpdX2UKGgGR0BsdKpR4yGjaAdNIQFoCEdAsKHIlAu7H3V9lChoBkdAcZzh3aBZp2gHTQUBaAhHQLCiAByS3b51fZQoaAZHQHK2lUhmoR9oB00rAWgIR0Cwog7d30PIdX2UKGgGR0Bxx3M7lq8EaAdNIQFoCEdAsKIjAxi5NHV9lChoBkdAcBOX/YJ3PmgHTQkBaAhHQLCipeK8+Rp1fZQoaAZHQHCb0QPI4l1oB00IAWgIR0Cwosik43m3dX2UKGgGR0BuXSHARChOaAdL/WgIR0CwospvcafjdX2UKGgGR0BwvSTpxFRYaAdL5WgIR0Cwosyz5XU6dX2UKGgGR0Bx+GCK77KraAdNQgFoCEdAsKLbZ6D5CXV9lChoBkdAbbg4XoC+12gHS/BoCEdAsKMJIatLc3V9lChoBkdAcFhYCyQgcWgHTYMBaAhHQLCjDFCb+cZ1ZS4="}, "ep_success_buffer": {":type:": "<class 'collections.deque'>", ":serialized:": "gAWVIAAAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKULg=="}, "_n_updates": 248, "observation_space": {":type:": "<class 'gymnasium.spaces.box.Box'>", ":serialized:": "gAWVcAIAAAAAAACMFGd5bW5hc2l1bS5zcGFjZXMuYm94lIwDQm94lJOUKYGUfZQojAVkdHlwZZSMBW51bXB5lGgFk5SMAmY0lImIh5RSlChLA4wBPJROTk5K/////0r/////SwB0lGKMDWJvdW5kZWRfYmVsb3eUjBJudW1weS5jb3JlLm51bWVyaWOUjAtfZnJvbWJ1ZmZlcpSTlCiWCAAAAAAAAAABAQEBAQEBAZRoB4wCYjGUiYiHlFKUKEsDjAF8lE5OTkr/////Sv////9LAHSUYksIhZSMAUOUdJRSlIwNYm91bmRlZF9hYm92ZZRoECiWCAAAAAAAAAABAQEBAQEBAZRoFEsIhZRoGHSUUpSMBl9zaGFwZZRLCIWUjANsb3eUaBAoliAAAAAAAAAAAAC0wgAAtMIAAKDAAACgwNsPScAAAKDAAAAAgAAAAICUaApLCIWUaBh0lFKUjARoaWdolGgQKJYgAAAAAAAAAAAAtEIAALRCAACgQAAAoEDbD0lAAACgQAAAgD8AAIA/lGgKSwiFlGgYdJRSlIwIbG93X3JlcHKUjFtbLTkwLiAgICAgICAgLTkwLiAgICAgICAgIC01LiAgICAgICAgIC01LiAgICAgICAgIC0zLjE0MTU5MjcgIC01LgogIC0wLiAgICAgICAgIC0wLiAgICAgICBdlIwJaGlnaF9yZXBylIxTWzkwLiAgICAgICAgOTAuICAgICAgICAgNS4gICAgICAgICA1LiAgICAgICAgIDMuMTQxNTkyNyAgNS4KICAxLiAgICAgICAgIDEuICAgICAgIF2UjApfbnBfcmFuZG9tlE51Yi4=", "dtype": "float32", "bounded_below": "[ True True True True True True True True]", "bounded_above": "[ True True True True True True True True]", "_shape": [8], "low": "[-90. -90. -5. -5. -3.1415927 -5.\n -0. -0. ]", "high": "[90. 90. 5. 5. 3.1415927 5.\n 1. 1. ]", "low_repr": "[-90. -90. -5. -5. -3.1415927 -5.\n -0. -0. ]", "high_repr": "[90. 90. 5. 5. 3.1415927 5.\n 1. 1. ]", "_np_random": null}, "action_space": {":type:": "<class 'gymnasium.spaces.discrete.Discrete'>", ":serialized:": "gAWV1QAAAAAAAACMGWd5bW5hc2l1bS5zcGFjZXMuZGlzY3JldGWUjAhEaXNjcmV0ZZSTlCmBlH2UKIwBbpSMFW51bXB5LmNvcmUubXVsdGlhcnJheZSMBnNjYWxhcpSTlIwFbnVtcHmUjAVkdHlwZZSTlIwCaTiUiYiHlFKUKEsDjAE8lE5OTkr/////Sv////9LAHSUYkMIBAAAAAAAAACUhpRSlIwFc3RhcnSUaAhoDkMIAAAAAAAAAACUhpRSlIwGX3NoYXBllCloCmgOjApfbnBfcmFuZG9tlE51Yi4=", "n": "4", "start": "0", "_shape": [], "dtype": "int64", "_np_random": null}, "n_envs": 16, "n_steps": 1024, "gamma": 0.999, "gae_lambda": 0.98, "ent_coef": 0.01, "vf_coef": 0.5, "max_grad_norm": 0.5, "batch_size": 64, "n_epochs": 4, "clip_range": {":type:": "<class 'function'>", ":serialized:": "gAWVxQIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwOX21ha2VfZnVuY3Rpb26Uk5QoaACMDV9idWlsdGluX3R5cGWUk5SMCENvZGVUeXBllIWUUpQoSwFLAEsASwFLAUsTQwSIAFMAlE6FlCmMAV+UhZSMSS91c3IvbG9jYWwvbGliL3B5dGhvbjMuMTAvZGlzdC1wYWNrYWdlcy9zdGFibGVfYmFzZWxpbmVzMy9jb21tb24vdXRpbHMucHmUjARmdW5jlEuEQwIEAZSMA3ZhbJSFlCl0lFKUfZQojAtfX3BhY2thZ2VfX5SMGHN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbpSMCF9fbmFtZV9flIwec3RhYmxlX2Jhc2VsaW5lczMuY29tbW9uLnV0aWxzlIwIX19maWxlX1+UjEkvdXNyL2xvY2FsL2xpYi9weXRob24zLjEwL2Rpc3QtcGFja2FnZXMvc3RhYmxlX2Jhc2VsaW5lczMvY29tbW9uL3V0aWxzLnB5lHVOTmgAjBBfbWFrZV9lbXB0eV9jZWxslJOUKVKUhZR0lFKUjBxjbG91ZHBpY2tsZS5jbG91ZHBpY2tsZV9mYXN0lIwSX2Z1bmN0aW9uX3NldHN0YXRllJOUaB99lH2UKGgWaA2MDF9fcXVhbG5hbWVfX5SMGWNvbnN0YW50X2ZuLjxsb2NhbHM+LmZ1bmOUjA9fX2Fubm90YXRpb25zX1+UfZSMDl9fa3dkZWZhdWx0c19flE6MDF9fZGVmYXVsdHNfX5ROjApfX21vZHVsZV9flGgXjAdfX2RvY19flE6MC19fY2xvc3VyZV9flGgAjApfbWFrZV9jZWxslJOURz/JmZmZmZmahZRSlIWUjBdfY2xvdWRwaWNrbGVfc3VibW9kdWxlc5RdlIwLX19nbG9iYWxzX1+UfZR1hpSGUjAu"}, "clip_range_vf": null, "normalize_advantage": true, "target_kl": null, "lr_schedule": {":type:": "<class 'function'>", ":serialized:": "gAWVxQIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwOX21ha2VfZnVuY3Rpb26Uk5QoaACMDV9idWlsdGluX3R5cGWUk5SMCENvZGVUeXBllIWUUpQoSwFLAEsASwFLAUsTQwSIAFMAlE6FlCmMAV+UhZSMSS91c3IvbG9jYWwvbGliL3B5dGhvbjMuMTAvZGlzdC1wYWNrYWdlcy9zdGFibGVfYmFzZWxpbmVzMy9jb21tb24vdXRpbHMucHmUjARmdW5jlEuEQwIEAZSMA3ZhbJSFlCl0lFKUfZQojAtfX3BhY2thZ2VfX5SMGHN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbpSMCF9fbmFtZV9flIwec3RhYmxlX2Jhc2VsaW5lczMuY29tbW9uLnV0aWxzlIwIX19maWxlX1+UjEkvdXNyL2xvY2FsL2xpYi9weXRob24zLjEwL2Rpc3QtcGFja2FnZXMvc3RhYmxlX2Jhc2VsaW5lczMvY29tbW9uL3V0aWxzLnB5lHVOTmgAjBBfbWFrZV9lbXB0eV9jZWxslJOUKVKUhZR0lFKUjBxjbG91ZHBpY2tsZS5jbG91ZHBpY2tsZV9mYXN0lIwSX2Z1bmN0aW9uX3NldHN0YXRllJOUaB99lH2UKGgWaA2MDF9fcXVhbG5hbWVfX5SMGWNvbnN0YW50X2ZuLjxsb2NhbHM+LmZ1bmOUjA9fX2Fubm90YXRpb25zX1+UfZSMDl9fa3dkZWZhdWx0c19flE6MDF9fZGVmYXVsdHNfX5ROjApfX21vZHVsZV9flGgXjAdfX2RvY19flE6MC19fY2xvc3VyZV9flGgAjApfbWFrZV9jZWxslJOURz8zqSowVTJhhZRSlIWUjBdfY2xvdWRwaWNrbGVfc3VibW9kdWxlc5RdlIwLX19nbG9iYWxzX1+UfZR1hpSGUjAu"}, "system_info": {"OS": "Linux-6.1.58+-x86_64-with-glibc2.35 # 1 SMP PREEMPT_DYNAMIC Sat Nov 18 15:31:17 UTC 2023", "Python": "3.10.12", "Stable-Baselines3": "2.0.0a5", "PyTorch": "2.1.0+cu121", "GPU Enabled": "True", "Numpy": "1.23.5", "Cloudpickle": "2.2.1", "Gymnasium": "0.28.1", "OpenAI Gym": "0.25.2"}}
|
ppo-LunarLander-v2.zip
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f14ac144281991acddd2bfbe9b5ccb4fa3e95763c7ff7740cddaae6cc955c45d
|
| 3 |
+
size 148028
|
ppo-LunarLander-v2/data
CHANGED
|
@@ -4,34 +4,34 @@
|
|
| 4 |
":serialized:": "gAWVOwAAAAAAAACMIXN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi5wb2xpY2llc5SMEUFjdG9yQ3JpdGljUG9saWN5lJOULg==",
|
| 5 |
"__module__": "stable_baselines3.common.policies",
|
| 6 |
"__doc__": "\n Policy class for actor-critic algorithms (has both policy and value prediction).\n Used by A2C, PPO and the likes.\n\n :param observation_space: Observation space\n :param action_space: Action space\n :param lr_schedule: Learning rate schedule (could be constant)\n :param net_arch: The specification of the policy and value networks.\n :param activation_fn: Activation function\n :param ortho_init: Whether to use or not orthogonal initialization\n :param use_sde: Whether to use State Dependent Exploration or not\n :param log_std_init: Initial value for the log standard deviation\n :param full_std: Whether to use (n_features x n_actions) parameters\n for the std instead of only (n_features,) when using gSDE\n :param use_expln: Use ``expln()`` function instead of ``exp()`` to ensure\n a positive standard deviation (cf paper). It allows to keep variance\n above zero and prevent it from growing too fast. In practice, ``exp()`` is usually enough.\n :param squash_output: Whether to squash the output using a tanh function,\n this allows to ensure boundaries when using gSDE.\n :param features_extractor_class: Features extractor to use.\n :param features_extractor_kwargs: Keyword arguments\n to pass to the features extractor.\n :param share_features_extractor: If True, the features extractor is shared between the policy and value networks.\n :param normalize_images: Whether to normalize images or not,\n dividing by 255.0 (True by default)\n :param optimizer_class: The optimizer to use,\n ``th.optim.Adam`` by default\n :param optimizer_kwargs: Additional keyword arguments,\n excluding the learning rate, to pass to the optimizer\n ",
|
| 7 |
-
"__init__": "<function ActorCriticPolicy.__init__ at
|
| 8 |
-
"_get_constructor_parameters": "<function ActorCriticPolicy._get_constructor_parameters at
|
| 9 |
-
"reset_noise": "<function ActorCriticPolicy.reset_noise at
|
| 10 |
-
"_build_mlp_extractor": "<function ActorCriticPolicy._build_mlp_extractor at
|
| 11 |
-
"_build": "<function ActorCriticPolicy._build at
|
| 12 |
-
"forward": "<function ActorCriticPolicy.forward at
|
| 13 |
-
"extract_features": "<function ActorCriticPolicy.extract_features at
|
| 14 |
-
"_get_action_dist_from_latent": "<function ActorCriticPolicy._get_action_dist_from_latent at
|
| 15 |
-
"_predict": "<function ActorCriticPolicy._predict at
|
| 16 |
-
"evaluate_actions": "<function ActorCriticPolicy.evaluate_actions at
|
| 17 |
-
"get_distribution": "<function ActorCriticPolicy.get_distribution at
|
| 18 |
-
"predict_values": "<function ActorCriticPolicy.predict_values at
|
| 19 |
"__abstractmethods__": "frozenset()",
|
| 20 |
-
"_abc_impl": "<_abc._abc_data object at
|
| 21 |
},
|
| 22 |
"verbose": 1,
|
| 23 |
"policy_kwargs": {},
|
| 24 |
"num_timesteps": 1015808,
|
| 25 |
-
"_total_timesteps":
|
| 26 |
"_num_timesteps_at_start": 0,
|
| 27 |
"seed": null,
|
| 28 |
"action_noise": null,
|
| 29 |
-
"start_time":
|
| 30 |
"learning_rate": 0.0003,
|
| 31 |
"tensorboard_log": null,
|
| 32 |
"_last_obs": {
|
| 33 |
":type:": "<class 'numpy.ndarray'>",
|
| 34 |
-
":serialized:": "
|
| 35 |
},
|
| 36 |
"_last_episode_starts": {
|
| 37 |
":type:": "<class 'numpy.ndarray'>",
|
|
@@ -41,11 +41,11 @@
|
|
| 41 |
"_episode_num": 0,
|
| 42 |
"use_sde": false,
|
| 43 |
"sde_sample_freq": -1,
|
| 44 |
-
"_current_progress_remaining": -0.
|
| 45 |
"_stats_window_size": 100,
|
| 46 |
"ep_info_buffer": {
|
| 47 |
":type:": "<class 'collections.deque'>",
|
| 48 |
-
":serialized:": "
|
| 49 |
},
|
| 50 |
"ep_success_buffer": {
|
| 51 |
":type:": "<class 'collections.deque'>",
|
|
|
|
| 4 |
":serialized:": "gAWVOwAAAAAAAACMIXN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi5wb2xpY2llc5SMEUFjdG9yQ3JpdGljUG9saWN5lJOULg==",
|
| 5 |
"__module__": "stable_baselines3.common.policies",
|
| 6 |
"__doc__": "\n Policy class for actor-critic algorithms (has both policy and value prediction).\n Used by A2C, PPO and the likes.\n\n :param observation_space: Observation space\n :param action_space: Action space\n :param lr_schedule: Learning rate schedule (could be constant)\n :param net_arch: The specification of the policy and value networks.\n :param activation_fn: Activation function\n :param ortho_init: Whether to use or not orthogonal initialization\n :param use_sde: Whether to use State Dependent Exploration or not\n :param log_std_init: Initial value for the log standard deviation\n :param full_std: Whether to use (n_features x n_actions) parameters\n for the std instead of only (n_features,) when using gSDE\n :param use_expln: Use ``expln()`` function instead of ``exp()`` to ensure\n a positive standard deviation (cf paper). It allows to keep variance\n above zero and prevent it from growing too fast. In practice, ``exp()`` is usually enough.\n :param squash_output: Whether to squash the output using a tanh function,\n this allows to ensure boundaries when using gSDE.\n :param features_extractor_class: Features extractor to use.\n :param features_extractor_kwargs: Keyword arguments\n to pass to the features extractor.\n :param share_features_extractor: If True, the features extractor is shared between the policy and value networks.\n :param normalize_images: Whether to normalize images or not,\n dividing by 255.0 (True by default)\n :param optimizer_class: The optimizer to use,\n ``th.optim.Adam`` by default\n :param optimizer_kwargs: Additional keyword arguments,\n excluding the learning rate, to pass to the optimizer\n ",
|
| 7 |
+
"__init__": "<function ActorCriticPolicy.__init__ at 0x7e12e3cdb6d0>",
|
| 8 |
+
"_get_constructor_parameters": "<function ActorCriticPolicy._get_constructor_parameters at 0x7e12e3cdb760>",
|
| 9 |
+
"reset_noise": "<function ActorCriticPolicy.reset_noise at 0x7e12e3cdb7f0>",
|
| 10 |
+
"_build_mlp_extractor": "<function ActorCriticPolicy._build_mlp_extractor at 0x7e12e3cdb880>",
|
| 11 |
+
"_build": "<function ActorCriticPolicy._build at 0x7e12e3cdb910>",
|
| 12 |
+
"forward": "<function ActorCriticPolicy.forward at 0x7e12e3cdb9a0>",
|
| 13 |
+
"extract_features": "<function ActorCriticPolicy.extract_features at 0x7e12e3cdba30>",
|
| 14 |
+
"_get_action_dist_from_latent": "<function ActorCriticPolicy._get_action_dist_from_latent at 0x7e12e3cdbac0>",
|
| 15 |
+
"_predict": "<function ActorCriticPolicy._predict at 0x7e12e3cdbb50>",
|
| 16 |
+
"evaluate_actions": "<function ActorCriticPolicy.evaluate_actions at 0x7e12e3cdbbe0>",
|
| 17 |
+
"get_distribution": "<function ActorCriticPolicy.get_distribution at 0x7e12e3cdbc70>",
|
| 18 |
+
"predict_values": "<function ActorCriticPolicy.predict_values at 0x7e12e3cdbd00>",
|
| 19 |
"__abstractmethods__": "frozenset()",
|
| 20 |
+
"_abc_impl": "<_abc._abc_data object at 0x7e12e3e77440>"
|
| 21 |
},
|
| 22 |
"verbose": 1,
|
| 23 |
"policy_kwargs": {},
|
| 24 |
"num_timesteps": 1015808,
|
| 25 |
+
"_total_timesteps": 1001000,
|
| 26 |
"_num_timesteps_at_start": 0,
|
| 27 |
"seed": null,
|
| 28 |
"action_noise": null,
|
| 29 |
+
"start_time": 1703800446569541592,
|
| 30 |
"learning_rate": 0.0003,
|
| 31 |
"tensorboard_log": null,
|
| 32 |
"_last_obs": {
|
| 33 |
":type:": "<class 'numpy.ndarray'>",
|
| 34 |
+
":serialized:": "gAWVdQIAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJYAAgAAAAAAAE1Mq72vDyc+mJe9PQW5Gr6aWwm9n2BJPQAAAAAAAAAAra1YvjQqTj9+myK+BPrkvmCUMb5ztjA8AAAAAAAAAACaTya+vA51P4Jkyr7i9f6+wTRKvrcHML4AAAAAAAAAAMYHBL40iZ8+zKAXPiGxfb5ZjAS9oPoLvQAAAAAAAAAAYK+CPjIHTT/6kC8+YGSsvu/+iT7enb67AAAAAAAAAACA27E9j3Y5ur08f7rfIce1tDQ7unDkkTkAAIA/AACAP2YwIT3hsu05kCF4PC6OSbz+wZe64txqPAAAAAAAAAAAgHsrvrQnYD5mVqs+/i9mvhEwaj1G1UY9AAAAAAAAAACzHz894Wyxutt68TXxp+gwy0Unuoq5GrUAAIA/AACAP0Cohr6uT2E/5dKEveGf3b70yEi+RQ2iPQAAAAAAAAAAowuAPsMuXT85wTI+XUPSvsergz5XbaK8AAAAAAAAAADNlOu7b05TP4iqn7tQYfW+nqoHvEpcgb0AAAAAAAAAABpTPj2PJmW6iOSHOM3MhTPj6u65tI+ftwAAgD8AAIA/zQiuvRjIij9m1yS+h3P3vrHH370GIpu7AAAAAAAAAAA6ghI+Z1vHPqvLkDuAuWW+Zm37PMVkrjwAAAAAAAAAANNiVz7hHvy8sC8vOwuwy7nTBGG+zqhsugAAgD8AAIA/lIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUiYiHlFKUKEsDjAE8lE5OTkr/////Sv////9LAHSUYksQSwiGlIwBQ5R0lFKULg=="
|
| 35 |
},
|
| 36 |
"_last_episode_starts": {
|
| 37 |
":type:": "<class 'numpy.ndarray'>",
|
|
|
|
| 41 |
"_episode_num": 0,
|
| 42 |
"use_sde": false,
|
| 43 |
"sde_sample_freq": -1,
|
| 44 |
+
"_current_progress_remaining": -0.014793206793206837,
|
| 45 |
"_stats_window_size": 100,
|
| 46 |
"ep_info_buffer": {
|
| 47 |
":type:": "<class 'collections.deque'>",
|
| 48 |
+
":serialized:": "gAWVJAwAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKUKH2UKIwBcpRHQG+HrVWjoISMAWyUTQ0BjAF0lEdAsI90JTl1bXV9lChoBkdAcIC7L+xW1mgHS/JoCEdAsI91as6q83V9lChoBkdAcmdYUnG83GgHTSABaAhHQLCPdY7aIvd1fZQoaAZHQHBAwtrbg0loB00aAWgIR0Cwj3oPXkHVdX2UKGgGR0Bx6Httygf2aAdNOQFoCEdAsI+gVLzwt3V9lChoBkdAcj1U/OdGzGgHTRABaAhHQLCP6L6k6911fZQoaAZHQHFqL6ciGFloB0vuaAhHQLCQGO58Sf11fZQoaAZHQHCGozi0fHRoB00bAWgIR0CwkB6Df3vhdX2UKGgGR0ByBYQ04zacaAdL+WgIR0CwkCGitaIOdX2UKGgGR0Bxypy6tknUaAdL8WgIR0CwkFGo3rD7dX2UKGgGR0BwNNwT/Q0GaAdL9GgIR0CwkFdHhCMQdX2UKGgGR0BwObbmEGqxaAdL9mgIR0CwkGkgjhUBdX2UKGgGR0BwxfenAIppaAdNCwFoCEdAsJDqy2QXAXV9lChoBkdAcITvFFUhm2gHTUsBaAhHQLCRKEkB0ZF1fZQoaAZHQHH72n4wh4doB002AWgIR0CwkTwc5sCUdX2UKGgGR0BtQBGlQ/HHaAdNCAFoCEdAsJF7fzjFQ3V9lChoBkdAbDdNLUTcqWgHTSUBaAhHQLCRytFKCg91fZQoaAZHQG9P+fqX4TNoB00QAWgIR0CwkdQdKdxydX2UKGgGR0BwqUxHoX9BaAdNKwFoCEdAsJHUKQaJh3V9lChoBkdAcrK+JgsshGgHTSwBaAhHQLCR1/yXlbN1fZQoaAZHQG4OdXcQAdZoB01VAWgIR0CwkeiL/CIldX2UKGgGR0Bx/7FbVz6raAdL6WgIR0CwkhR6KLsKdX2UKGgGR0BuPLLyMDOkaAdNBQFoCEdAsJIc6IWP93V9lChoBkdAcW9oLG7z1GgHTQkBaAhHQLCScqLjxTd1fZQoaAZHQHAt96Tnq3VoB00ZAWgIR0CwkpfugHu7dX2UKGgGR0Bw/zhDPWxyaAdNBQFoCEdAsJLVxYJVsHV9lChoBkdAcIKJm/WUbGgHTToBaAhHQLCTOWw/xDt1fZQoaAZHQHK1vqcEvCdoB005AWgIR0Cwk0Ajlgc+dX2UKGgGR0Bxibru6VdHaAdL7mgIR0Cwk3v1HvtudX2UKGgGR0BwhSA4GUwBaAdNIAFoCEdAsJOyAG0NSnV9lChoBkdAcoFqIrOJL2gHTQABaAhHQLCTwvCMxXZ1fZQoaAZHQHD1jEaVD8doB0v6aAhHQLCT/3MINVl1fZQoaAZHQHIy8D8tPHloB0vkaAhHQLCUF4z7/GV1fZQoaAZHQHDnhn8KohpoB00HAWgIR0CwlHaYeDFqdX2UKGgGR0BwbPe67NB4aAdNBgFoCEdAsJSQg1WKdnV9lChoBkdAcZsaURnOB2gHTRcBaAhHQLCUo90ihWZ1fZQoaAZHQHGulIRRMvhoB00nAWgIR0CwlNKkAPupdX2UKGgGR0BuZqFEiMYNaAdL7WgIR0CwlNvC2tuDdX2UKGgGR0BwRaRaHKwIaAdNEwFoCEdAsJToemvW6XV9lChoBkdAcPyC04R282gHTQUBaAhHQLCVQbB42TB1fZQoaAZHQHN3mMS9M9NoB00hAWgIR0CwlaIrvsqsdX2UKGgGR0BxLCB/ZuhsaAdNFwFoCEdAsJo0i+tbLXV9lChoBkdAbwJUe+23KGgHTS4BaAhHQLCaaUoKD011fZQoaAZHQHMjGuDBdldoB00jAWgIR0CwmoTvy9VWdX2UKGgGR0Btrj2xptaZaAdNvAFoCEdAsJqjTiKiwnV9lChoBkdAUZ7+zdDYy2gHS7loCEdAsJrC2a2F4HV9lChoBkdAcoYWv8qFy2gHTQ8BaAhHQLCaxvG6wt91fZQoaAZHQHCtNZaFEiNoB00MAWgIR0CwmtO0svqUdX2UKGgGR0Bw0GIeo1k2aAdL82gIR0Cwmv88YAKfdX2UKGgGR0BwewoOQQtjaAdNSwFoCEdAsJsTpfQa73V9lChoBkdAcwVz8gpz92gHTRoBaAhHQLCbO3o9s8B1fZQoaAZHQHGicIqslsxoB01sAWgIR0Cwm0XzQNTcdX2UKGgGR0BurCSowVTKaAdL+GgIR0Cwm0mkN4JNdX2UKGgGR0BtOGU0Nz8xaAdNAAFoCEdAsJtJnZkCm3V9lChoBkdAbvuMJhOQAGgHTSIBaAhHQLCbY2FFlTZ1fZQoaAZHQHIgg9FF2FFoB00FAWgIR0Cwm5HLRrrPdX2UKGgGR0Bx6/8uSOinaAdNEgFoCEdAsJwAtPHktHV9lChoBkdAbzZTsIE8rGgHS/poCEdAsJwJ/oaDPHV9lChoBkdAcawDxLCemWgHS+RoCEdAsJxGhew9q3V9lChoBkdAcH/2MKkVOGgHS+toCEdAsJx1H+ZPVXV9lChoBkdAcO5fVqesgmgHTRsBaAhHQLCcpaZQYUF1fZQoaAZHQHFYo9HMEA5oB0v3aAhHQLCcpdLQHA11fZQoaAZHQG4E09QoCuFoB00VAWgIR0CwnN2OIZZTdX2UKGgGR0BxoKTLW7OFaAdL42gIR0CwnO9tALRbdX2UKGgGR0BwIyHO8kD7aAdNVgFoCEdAsJ0NM8HObHV9lChoBkdASgcRJ2+wkmgHS8BoCEdAsJ0aCXhOxnV9lChoBkdAcNm3iJfplmgHTQ8BaAhHQLCdIY4Qz1t1fZQoaAZHQHDZ2Ef1YhdoB0v4aAhHQLCdKRgZ0jl1fZQoaAZHQHHBZyhi9ZloB00kAWgIR0CwnTZb2USqdX2UKGgGR0BuRvXumaYvaAdNKAFoCEdAsJ2fybx3FHV9lChoBkdAcU/05EMLGGgHTUQBaAhHQLCdtEjPfKp1fZQoaAZHQG4DU6xPfsNoB0v+aAhHQLCeAuSfUWl1fZQoaAZHQGwJZh8YyftoB0v+aAhHQLCeUg4Otnx1fZQoaAZHQHFQaC6H0shoB01IAWgIR0CwnrpaA4GVdX2UKGgGR0BwhhmapgkUaAdNCgFoCEdAsJ7UGUwBYHV9lChoBkdAcy8w5/9YOmgHS/RoCEdAsJ7YnlXA/XV9lChoBkdAc3s8r7O3UmgHTSwBaAhHQLCe88WbgCR1fZQoaAZHQHFQSiqQzUJoB0vvaAhHQLCe/lY2bXp1fZQoaAZHQHH3ZHuqm0poB00NAWgIR0Cwnx+rdWQwdX2UKGgGR0BxFcn/kvK2aAdNQAFoCEdAsJ9O6NEPUnV9lChoBkdAcA6fO2RaHWgHTScBaAhHQLCftGyX2M91fZQoaAZHQHCy6TOgQH1oB0v0aAhHQLCf4gcLjPx1fZQoaAZHQHIB2Ef1YhdoB006AWgIR0CwoAPfsNUgdX2UKGgGR0BuXTQeFL39aAdNTwFoCEdAsKAdyp71I3V9lChoBkdAbSs8+zMRpWgHTQ0BaAhHQLCgSggow251fZQoaAZHQHF4w7YChexoB0v4aAhHQLCge71qWTp1fZQoaAZHQHBbd92HLzRoB016AWgIR0CwoIbILgGbdX2UKGgGR0BwBx4A0bcXaAdL/WgIR0CwoONHc1wYdX2UKGgGR0BxWWBGx2SuaAdL6WgIR0CwoUJPVNHpdX2UKGgGR0BsdKpR4yGjaAdNIQFoCEdAsKHIlAu7H3V9lChoBkdAcZzh3aBZp2gHTQUBaAhHQLCiAByS3b51fZQoaAZHQHK2lUhmoR9oB00rAWgIR0Cwog7d30PIdX2UKGgGR0Bxx3M7lq8EaAdNIQFoCEdAsKIjAxi5NHV9lChoBkdAcBOX/YJ3PmgHTQkBaAhHQLCipeK8+Rp1fZQoaAZHQHCb0QPI4l1oB00IAWgIR0Cwosik43m3dX2UKGgGR0BuXSHARChOaAdL/WgIR0CwospvcafjdX2UKGgGR0BwvSTpxFRYaAdL5WgIR0Cwosyz5XU6dX2UKGgGR0Bx+GCK77KraAdNQgFoCEdAsKLbZ6D5CXV9lChoBkdAbbg4XoC+12gHS/BoCEdAsKMJIatLc3V9lChoBkdAcFhYCyQgcWgHTYMBaAhHQLCjDFCb+cZ1ZS4="
|
| 49 |
},
|
| 50 |
"ep_success_buffer": {
|
| 51 |
":type:": "<class 'collections.deque'>",
|
ppo-LunarLander-v2/policy.optimizer.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 88362
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1831470247bf64bc8da79077b4e3ed873b826c4746c628fa1d35ca1d8cd7f334
|
| 3 |
size 88362
|
ppo-LunarLander-v2/policy.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 43762
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5486092e53ce30d99a4d9e48da7d7c4a8c65f3bc4222a058c7a0a17613122dc0
|
| 3 |
size 43762
|
replay.mp4
CHANGED
|
Binary files a/replay.mp4 and b/replay.mp4 differ
|
|
|
results.json
CHANGED
|
@@ -1 +1 @@
|
|
| 1 |
-
{"mean_reward":
|
|
|
|
| 1 |
+
{"mean_reward": 251.87058919999998, "std_reward": 17.437971208857622, "is_deterministic": true, "n_eval_episodes": 10, "eval_datetime": "2023-12-28T22:19:17.904279"}
|