Upload PPO LunarLander-v2 trained agent
Browse files- README.md +1 -1
- config.json +1 -1
- lunarLander.zip +2 -2
- lunarLander/data +19 -19
- lunarLander/policy.optimizer.pth +1 -1
- lunarLander/policy.pth +1 -1
- replay.mp4 +0 -0
- results.json +1 -1
README.md
CHANGED
|
@@ -16,7 +16,7 @@ model-index:
|
|
| 16 |
type: LunarLander-v2
|
| 17 |
metrics:
|
| 18 |
- type: mean_reward
|
| 19 |
-
value:
|
| 20 |
name: mean_reward
|
| 21 |
verified: false
|
| 22 |
---
|
|
|
|
| 16 |
type: LunarLander-v2
|
| 17 |
metrics:
|
| 18 |
- type: mean_reward
|
| 19 |
+
value: 249.30 +/- 14.06
|
| 20 |
name: mean_reward
|
| 21 |
verified: false
|
| 22 |
---
|
config.json
CHANGED
|
@@ -1 +1 @@
|
|
| 1 |
-
{"policy_class": {":type:": "<class 'abc.ABCMeta'>", ":serialized:": "gAWVOwAAAAAAAACMIXN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi5wb2xpY2llc5SMEUFjdG9yQ3JpdGljUG9saWN5lJOULg==", "__module__": "stable_baselines3.common.policies", "__doc__": "\n Policy class for actor-critic algorithms (has both policy and value prediction).\n Used by A2C, PPO and the likes.\n\n :param observation_space: Observation space\n :param action_space: Action space\n :param lr_schedule: Learning rate schedule (could be constant)\n :param net_arch: The specification of the policy and value networks.\n :param activation_fn: Activation function\n :param ortho_init: Whether to use or not orthogonal initialization\n :param use_sde: Whether to use State Dependent Exploration or not\n :param log_std_init: Initial value for the log standard deviation\n :param full_std: Whether to use (n_features x n_actions) parameters\n for the std instead of only (n_features,) when using gSDE\n :param use_expln: Use ``expln()`` function instead of ``exp()`` to ensure\n a positive standard deviation (cf paper). It allows to keep variance\n above zero and prevent it from growing too fast. In practice, ``exp()`` is usually enough.\n :param squash_output: Whether to squash the output using a tanh function,\n this allows to ensure boundaries when using gSDE.\n :param features_extractor_class: Features extractor to use.\n :param features_extractor_kwargs: Keyword arguments\n to pass to the features extractor.\n :param share_features_extractor: If True, the features extractor is shared between the policy and value networks.\n :param normalize_images: Whether to normalize images or not,\n dividing by 255.0 (True by default)\n :param optimizer_class: The optimizer to use,\n ``th.optim.Adam`` by default\n :param optimizer_kwargs: Additional keyword arguments,\n excluding the learning rate, to pass to the optimizer\n ", "__init__": "<function ActorCriticPolicy.__init__ at 0x7f951beec040>", "_get_constructor_parameters": "<function ActorCriticPolicy._get_constructor_parameters at 0x7f951beec0d0>", "reset_noise": "<function ActorCriticPolicy.reset_noise at 0x7f951beec160>", "_build_mlp_extractor": "<function ActorCriticPolicy._build_mlp_extractor at 0x7f951beec1f0>", "_build": "<function ActorCriticPolicy._build at 0x7f951beec280>", "forward": "<function ActorCriticPolicy.forward at 0x7f951beec310>", "extract_features": "<function ActorCriticPolicy.extract_features at 0x7f951beec3a0>", "_get_action_dist_from_latent": "<function ActorCriticPolicy._get_action_dist_from_latent at 0x7f951beec430>", "_predict": "<function ActorCriticPolicy._predict at 0x7f951beec4c0>", "evaluate_actions": "<function ActorCriticPolicy.evaluate_actions at 0x7f951beec550>", "get_distribution": "<function ActorCriticPolicy.get_distribution at 0x7f951beec5e0>", "predict_values": "<function ActorCriticPolicy.predict_values at 0x7f951beec670>", "__abstractmethods__": "frozenset()", "_abc_impl": "<_abc._abc_data object at 0x7f951c0e4380>"}, "verbose": 1, "policy_kwargs": {}, "num_timesteps": 162985, "_total_timesteps": 1000000, "_num_timesteps_at_start": 0, "seed": null, "action_noise": null, "start_time": 1688501746450727266, "learning_rate": 0.0003, "tensorboard_log": null, "_last_obs": {":type:": "<class 'numpy.ndarray'>", ":serialized:": "gAWVlQAAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJYgAAAAAAAAAMLEgL6HLBc/LlEDvrX4k75gPgC9M5ToOwAAAAAAAAAAlIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUiYiHlFKUKEsDjAE8lE5OTkr/////Sv////9LAHSUYksBSwiGlIwBQ5R0lFKULg=="}, "_last_episode_starts": {":type:": "<class 'numpy.ndarray'>", ":serialized:": "gAWVdAAAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJYBAAAAAAAAAACUjAVudW1weZSMBWR0eXBllJOUjAJiMZSJiIeUUpQoSwOMAXyUTk5OSv////9K/////0sAdJRiSwGFlIwBQ5R0lFKULg=="}, "_last_original_obs": null, "_episode_num": 0, "use_sde": false, "sde_sample_freq": -1, "_current_progress_remaining": 0.837184, "_stats_window_size": 100, "ep_info_buffer": {":type:": "<class 'collections.deque'>", ":serialized:": "gAWVLwwAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKUKH2UKIwBcpRHQFnUFC9h7VuMAWyUTegDjAF0lEdAYTpUZNwiq3V9lChoBkfASZYre67NCGgHS7ZoCEdAYUKZa3ZwoHV9lChoBkdAWVj9XLeQ+2gHTegDaAhHQGF7P8ZUDMh1fZQoaAZHQGEVUD2alUJoB03oA2gIR0BhpJOUMXrMdX2UKGgGR0BIZelTFVDKaAdN6ANoCEdAYc45WilBQnV9lChoBkfAYvqsOG0u2GgHS99oCEdAYdVKvFFUhnV9lChoBkdAJDS1eBxxUGgHS8NoCEdAYdtT72tdRnV9lChoBkfAYfPJcxCY1GgHS81oCEdAYetAUtZmqnV9lChoBkfASnZsXSBsh2gHS8hoCEdAYfEvNeMQ3HV9lChoBkfAM85aV2Rq5GgHS6toCEdAYfbRtxdY4nV9lChoBkfAQ/sMEzO5a2gHS/poCEdAYf5eJpFkQXV9lChoBkdAY8KBkI5YHWgHTbECaAhHQGIemvW6K+B1fZQoaAZHQEd6lchTwUhoB0vWaAhHQGIlK7I1cdJ1fZQoaAZHwDJnuKGcnVpoB0uuaAhHQGIqZHd43WF1fZQoaAZHQCgyVrylN11oB0u6aAhHQGI5i3G4qgB1fZQoaAZHQF0MmwJPZZloB03oA2gIR0BiYxNdqtYCdX2UKGgGR0AuucPvrnklaAdN6ANoCEdAYoNLqUu+RHV9lChoBkdAYPOB4D9wWGgHTegDaAhHQGKstfgJkXl1fZQoaAZHQE3TCpFTeftoB03oA2gIR0Bi5MbgjyFxdX2UKGgGR0Be4etSydFwaAdNxgJoCEdAYxJg0j1PFnV9lChoBkdAVpg7kn1FpmgHTegDaAhHQGNPhgeA/cF1fZQoaAZHwEOdwDNhVlxoB0vtaAhHQGNags052hZ1fZQoaAZHQFYIxT850bNoB03oA2gIR0BjilRrJr+HdX2UKGgGR0A3qLm6oVEeaAdL02gIR0Bjmro8p1A8dX2UKGgGR0AhJ07KaG5+aAdL3WgIR0BjogJVsDW9dX2UKGgGR0BUgr8FY+0PaAdN6ANoCEdAY8w83dbgTHV9lChoBkdAZ7Jev6j322gHTTcDaAhHQGPwvRiPQv91fZQoaAZHwDigPoV2zOZoB0u/aAhHQGP2xOUMXrN1fZQoaAZHwFAbpeNT989oB0vKaAhHQGP87F0gbId1fZQoaAZHwDmrFvQ4S6FoB0vDaAhHQGQC/OUt7KJ1fZQoaAZHQDCPxZuAI6doB0u4aAhHQGQIwZ4wAVB1fZQoaAZHQGFjigkC3gFoB03oA2gIR0BkMYoy9EkTdX2UKGgGR8Ai2uDjBEa3aAdL02gIR0BkOATh5xBFdX2UKGgGR0BbQajWTX8PaAdN6ANoCEdAZGCglnh86XV9lChoBkdAYAl+5OJtSGgHTegDaAhHQGSJxi5NGmV1fZQoaAZHQFHo6XjU/fRoB03oA2gIR0BktmNT987ZdX2UKGgGR8BdtcrNGEwnaAdNUAFoCEdAZNM2H+Idl3V9lChoBkdAXyyjwhGH6GgHTegDaAhHQGUMI7Njbzt1fZQoaAZHQFjU9lVcUudoB03oA2gIR0BlSgKMNtqIdX2UKGgGR8Bi8vNC7btaaAdNRgJoCEdAZWN5ckdFOXV9lChoBkdAVj9YwIt16mgHTegDaAhHQGWMtrj5sTF1fZQoaAZHQD0tVea8YhtoB03oA2gIR0Blti0WuX/pdX2UKGgGR8BROw/9pAUtaAdN6ANoCEdAZd7apPykK3V9lChoBkdAOqO+ZgG8mWgHTegDaAhHQGYHqL0jC551fZQoaAZHwFQAxnWattBoB02vAWgIR0BmHklAu7HydX2UKGgGR0BqcEujASFoaAdNIwNoCEdAZjixk/bCanV9lChoBkdATc4OQQtjC2gHTegDaAhHQGZhvgvUSZl1fZQoaAZHv6ymALApKBdoB01LAWgIR0BmdOdkJ8fFdX2UKGgGR8A8vnMMZxaQaAdNOQFoCEdAZn7EroW56XV9lChoBkdAWtk+LWI42mgHTegDaAhHQGappj+aScN1fZQoaAZHQGCG1XvH93toB03IAmgIR0Bm1eoegctHdX2UKGgGR0A2foVVPva2aAdNAAFoCEdAZuCwA2hqTXV9lChoBkdAYiUxGDtgKGgHTbEBaAhHQGbyg7o0Q9R1fZQoaAZHQFXh3hGYrrhoB03oA2gIR0BnLsXJo0yhdX2UKGgGR8BURmgSOBDpaAdNawFoCEdAZ01QBPsRhHV9lChoBkfAIPa0QbuMM2gHTSEBaAhHQGdakRBeHBV1fZQoaAZHQEq/HAAQxvhoB03oA2gIR0BnhisXBP9DdX2UKGgGR0BX7ZuuRs/IaAdN6ANoCEdAZ6+quKXOW3V9lChoBkfAKp/t6X0GvGgHTS0BaAhHQGe5RYzSCvp1fZQoaAZHQF0XoWYWtU5oB03oA2gIR0Bn4uMhouf3dX2UKGgGR8BVvPboKUmlaAdNowFoCEdAZ/jcafjCHnV9lChoBkdAYehtMPBi1GgHTY4CaAhHQGgNg9/z8P51fZQoaAZHwDxY495hScdoB01QAWgIR0BoIVFBppN9dX2UKGgGR8BI1afapPykaAdNLgFoCEdAaCr6k690zXV9lChoBkdAWXPdznzQNWgHTegDaAhHQGhUTjWCmMx1fZQoaAZHwEP2GsV+I/JoB0v+aAhHQGhb/zJ6po91fZQoaAZHQDZTIo3Jgb9oB00uAWgIR0BoZWwTufEodX2UKGgGR0BhQsUKzAvdaAdN6ANoCEdAaI7jjJdSl3V9lChoBkdAW6BYNiH6/WgHTegDaAhHQGjA2FN+LFZ1fZQoaAZHQF2SI1tO2y9oB03oA2gIR0Bo+rG7z06HdX2UKGgGR8AhNxFRYRukaAdLymgIR0BpA75XU6PsdX2UKGgGR0BTDWEbo8p1aAdN6ANoCEdAaUEaGYa5w3V9lChoBkdAUjVQxesxPGgHTegDaAhHQGlyB7eEZix1fZQoaAZHwDZksrd30PJoB0voaAhHQGmCvWQOnVJ1fZQoaAZHQFmP5rxiG35oB03oA2gIR0Bpq4VCXyAhdX2UKGgGR0BbxDkyULUkaAdN6ANoCEdAadTKsdT5wnV9lChoBkdATSH5LytmtmgHTegDaAhHQGn9ra24NI91fZQoaAZHQFi9AymALApoB03oA2gIR0BqJ1DUmUnpdX2UKGgGR0Bj2lvbXYlIaAdNdANoCEdAakPmI0qH5HV9lChoBkdAQdnNmlImPmgHTQoBaAhHQGpWPS+g13t1fZQoaAZHQF4UXp4bCJpoB03oA2gIR0Bqf8m4RVZLdX2UKGgGR8A7zD+irT6SaAdNeQFoCEdAaotYT0xubnV9lChoBkdAWsn4j8k2P2gHTegDaAhHQGrAURnOB191fZQoaAZHQFeb48lolD5oB03oA2gIR0Bq+lMXaakRdX2UKGgGR8Bad9SIgvDhaAdNmgFoCEdAawxxDLKV6nV9lChoBkfAaz1LRrrPdGgHTSECaAhHQGsz6wMYuTR1fZQoaAZHQGF68baRISVoB03oA2gIR0BrZob+98JEdX2UKGgGR0BZMqdUbT+eaAdN6ANoCEdAa5CKKpDNQnV9lChoBkdAZERpNbkfcWgHTegDaAhHQGu6J0nw5Np1fZQoaAZHwFCBJ7LMcIZoB00QAWgIR0BrwptrKvFFdX2UKGgGR0BYj1bRnezlaAdNRAJoCEdAa92TIvJzUHV9lChoBkdAZq0A3kxREWgHTU4DaAhHQGv4dXDFZPl1fZQoaAZHwFn0ZxaPjn5oB025AWgIR0BsD9gpjMFEdX2UKGgGR0BglhHPNVzZaAdN6ANoCEdAbDk+UQkHEHV9lChoBkdAOv3mNipeeGgHTegDaAhHQGxifsE7nxJ1fZQoaAZHQGQq+XzDn/1oB03oA2gIR0Bsi+weNkvsdX2UKGgGR0Biu362v0ROaAdN6ANoCEdAbMOPMjeKsXV9lChoBkfAVBS5Gz8gp2gHTUYBaAhHQGzQ+MqBmPJ1fZQoaAZHQFag7eEZiuxoB03oA2gIR0BtDPIGQjlgdX2UKGgGR0Blk/h0hePaaAdNkwFoCEdAbR8cwxnFpHVlLg=="}, "ep_success_buffer": {":type:": "<class 'collections.deque'>", ":serialized:": "gAWVIAAAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKULg=="}, "_n_updates": 636, "observation_space": {":type:": "<class 'gymnasium.spaces.box.Box'>", ":serialized:": "gAWVcAIAAAAAAACMFGd5bW5hc2l1bS5zcGFjZXMuYm94lIwDQm94lJOUKYGUfZQojAVkdHlwZZSMBW51bXB5lGgFk5SMAmY0lImIh5RSlChLA4wBPJROTk5K/////0r/////SwB0lGKMDWJvdW5kZWRfYmVsb3eUjBJudW1weS5jb3JlLm51bWVyaWOUjAtfZnJvbWJ1ZmZlcpSTlCiWCAAAAAAAAAABAQEBAQEBAZRoB4wCYjGUiYiHlFKUKEsDjAF8lE5OTkr/////Sv////9LAHSUYksIhZSMAUOUdJRSlIwNYm91bmRlZF9hYm92ZZRoECiWCAAAAAAAAAABAQEBAQEBAZRoFEsIhZRoGHSUUpSMBl9zaGFwZZRLCIWUjANsb3eUaBAoliAAAAAAAAAAAAC0wgAAtMIAAKDAAACgwNsPScAAAKDAAAAAgAAAAICUaApLCIWUaBh0lFKUjARoaWdolGgQKJYgAAAAAAAAAAAAtEIAALRCAACgQAAAoEDbD0lAAACgQAAAgD8AAIA/lGgKSwiFlGgYdJRSlIwIbG93X3JlcHKUjFtbLTkwLiAgICAgICAgLTkwLiAgICAgICAgIC01LiAgICAgICAgIC01LiAgICAgICAgIC0zLjE0MTU5MjcgIC01LgogIC0wLiAgICAgICAgIC0wLiAgICAgICBdlIwJaGlnaF9yZXBylIxTWzkwLiAgICAgICAgOTAuICAgICAgICAgNS4gICAgICAgICA1LiAgICAgICAgIDMuMTQxNTkyNyAgNS4KICAxLiAgICAgICAgIDEuICAgICAgIF2UjApfbnBfcmFuZG9tlE51Yi4=", "dtype": "float32", "bounded_below": "[ True True True True True True True True]", "bounded_above": "[ True True True True True True True True]", "_shape": [8], "low": "[-90. -90. -5. -5. -3.1415927 -5.\n -0. -0. ]", "high": "[90. 90. 5. 5. 3.1415927 5.\n 1. 1. ]", "low_repr": "[-90. -90. -5. -5. -3.1415927 -5.\n -0. -0. ]", "high_repr": "[90. 90. 5. 5. 3.1415927 5.\n 1. 1. ]", "_np_random": null}, "action_space": {":type:": "<class 'gymnasium.spaces.discrete.Discrete'>", ":serialized:": "gAWV1QAAAAAAAACMGWd5bW5hc2l1bS5zcGFjZXMuZGlzY3JldGWUjAhEaXNjcmV0ZZSTlCmBlH2UKIwBbpSMFW51bXB5LmNvcmUubXVsdGlhcnJheZSMBnNjYWxhcpSTlIwFbnVtcHmUjAVkdHlwZZSTlIwCaTiUiYiHlFKUKEsDjAE8lE5OTkr/////Sv////9LAHSUYkMIBAAAAAAAAACUhpRSlIwFc3RhcnSUaAhoDkMIAAAAAAAAAACUhpRSlIwGX3NoYXBllCloCmgOjApfbnBfcmFuZG9tlE51Yi4=", "n": "4", "start": "0", "_shape": [], "dtype": "int64", "_np_random": null}, "n_envs": 1, "n_steps": 1024, "gamma": 0.999, "gae_lambda": 0.98, "ent_coef": 0.01, "vf_coef": 0.5, "max_grad_norm": 0.5, "batch_size": 64, "n_epochs": 4, "clip_range": {":type:": "<class 'function'>", ":serialized:": "gAWVxQIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwOX21ha2VfZnVuY3Rpb26Uk5QoaACMDV9idWlsdGluX3R5cGWUk5SMCENvZGVUeXBllIWUUpQoSwFLAEsASwFLAUsTQwSIAFMAlE6FlCmMAV+UhZSMSS91c3IvbG9jYWwvbGliL3B5dGhvbjMuMTAvZGlzdC1wYWNrYWdlcy9zdGFibGVfYmFzZWxpbmVzMy9jb21tb24vdXRpbHMucHmUjARmdW5jlEuEQwIEAZSMA3ZhbJSFlCl0lFKUfZQojAtfX3BhY2thZ2VfX5SMGHN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbpSMCF9fbmFtZV9flIwec3RhYmxlX2Jhc2VsaW5lczMuY29tbW9uLnV0aWxzlIwIX19maWxlX1+UjEkvdXNyL2xvY2FsL2xpYi9weXRob24zLjEwL2Rpc3QtcGFja2FnZXMvc3RhYmxlX2Jhc2VsaW5lczMvY29tbW9uL3V0aWxzLnB5lHVOTmgAjBBfbWFrZV9lbXB0eV9jZWxslJOUKVKUhZR0lFKUjBxjbG91ZHBpY2tsZS5jbG91ZHBpY2tsZV9mYXN0lIwSX2Z1bmN0aW9uX3NldHN0YXRllJOUaB99lH2UKGgWaA2MDF9fcXVhbG5hbWVfX5SMGWNvbnN0YW50X2ZuLjxsb2NhbHM+LmZ1bmOUjA9fX2Fubm90YXRpb25zX1+UfZSMDl9fa3dkZWZhdWx0c19flE6MDF9fZGVmYXVsdHNfX5ROjApfX21vZHVsZV9flGgXjAdfX2RvY19flE6MC19fY2xvc3VyZV9flGgAjApfbWFrZV9jZWxslJOURz/JmZmZmZmahZRSlIWUjBdfY2xvdWRwaWNrbGVfc3VibW9kdWxlc5RdlIwLX19nbG9iYWxzX1+UfZR1hpSGUjAu"}, "clip_range_vf": null, "normalize_advantage": true, "target_kl": null, "lr_schedule": {":type:": "<class 'function'>", ":serialized:": "gAWVxQIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwOX21ha2VfZnVuY3Rpb26Uk5QoaACMDV9idWlsdGluX3R5cGWUk5SMCENvZGVUeXBllIWUUpQoSwFLAEsASwFLAUsTQwSIAFMAlE6FlCmMAV+UhZSMSS91c3IvbG9jYWwvbGliL3B5dGhvbjMuMTAvZGlzdC1wYWNrYWdlcy9zdGFibGVfYmFzZWxpbmVzMy9jb21tb24vdXRpbHMucHmUjARmdW5jlEuEQwIEAZSMA3ZhbJSFlCl0lFKUfZQojAtfX3BhY2thZ2VfX5SMGHN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbpSMCF9fbmFtZV9flIwec3RhYmxlX2Jhc2VsaW5lczMuY29tbW9uLnV0aWxzlIwIX19maWxlX1+UjEkvdXNyL2xvY2FsL2xpYi9weXRob24zLjEwL2Rpc3QtcGFja2FnZXMvc3RhYmxlX2Jhc2VsaW5lczMvY29tbW9uL3V0aWxzLnB5lHVOTmgAjBBfbWFrZV9lbXB0eV9jZWxslJOUKVKUhZR0lFKUjBxjbG91ZHBpY2tsZS5jbG91ZHBpY2tsZV9mYXN0lIwSX2Z1bmN0aW9uX3NldHN0YXRllJOUaB99lH2UKGgWaA2MDF9fcXVhbG5hbWVfX5SMGWNvbnN0YW50X2ZuLjxsb2NhbHM+LmZ1bmOUjA9fX2Fubm90YXRpb25zX1+UfZSMDl9fa3dkZWZhdWx0c19flE6MDF9fZGVmYXVsdHNfX5ROjApfX21vZHVsZV9flGgXjAdfX2RvY19flE6MC19fY2xvc3VyZV9flGgAjApfbWFrZV9jZWxslJOURz8zqSowVTJhhZRSlIWUjBdfY2xvdWRwaWNrbGVfc3VibW9kdWxlc5RdlIwLX19nbG9iYWxzX1+UfZR1hpSGUjAu"}, "system_info": {"OS": "Linux-5.15.107+-x86_64-with-glibc2.31 # 1 SMP Sat Apr 29 09:15:28 UTC 2023", "Python": "3.10.12", "Stable-Baselines3": "2.0.0a5", "PyTorch": "2.0.1+cu118", "GPU Enabled": "False", "Numpy": "1.22.4", "Cloudpickle": "2.2.1", "Gymnasium": "0.28.1", "OpenAI Gym": "0.25.2"}}
|
|
|
|
| 1 |
+
{"policy_class": {":type:": "<class 'abc.ABCMeta'>", ":serialized:": "gAWVOwAAAAAAAACMIXN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi5wb2xpY2llc5SMEUFjdG9yQ3JpdGljUG9saWN5lJOULg==", "__module__": "stable_baselines3.common.policies", "__doc__": "\n Policy class for actor-critic algorithms (has both policy and value prediction).\n Used by A2C, PPO and the likes.\n\n :param observation_space: Observation space\n :param action_space: Action space\n :param lr_schedule: Learning rate schedule (could be constant)\n :param net_arch: The specification of the policy and value networks.\n :param activation_fn: Activation function\n :param ortho_init: Whether to use or not orthogonal initialization\n :param use_sde: Whether to use State Dependent Exploration or not\n :param log_std_init: Initial value for the log standard deviation\n :param full_std: Whether to use (n_features x n_actions) parameters\n for the std instead of only (n_features,) when using gSDE\n :param use_expln: Use ``expln()`` function instead of ``exp()`` to ensure\n a positive standard deviation (cf paper). It allows to keep variance\n above zero and prevent it from growing too fast. In practice, ``exp()`` is usually enough.\n :param squash_output: Whether to squash the output using a tanh function,\n this allows to ensure boundaries when using gSDE.\n :param features_extractor_class: Features extractor to use.\n :param features_extractor_kwargs: Keyword arguments\n to pass to the features extractor.\n :param share_features_extractor: If True, the features extractor is shared between the policy and value networks.\n :param normalize_images: Whether to normalize images or not,\n dividing by 255.0 (True by default)\n :param optimizer_class: The optimizer to use,\n ``th.optim.Adam`` by default\n :param optimizer_kwargs: Additional keyword arguments,\n excluding the learning rate, to pass to the optimizer\n ", "__init__": "<function ActorCriticPolicy.__init__ at 0x7f2faf8163b0>", "_get_constructor_parameters": "<function ActorCriticPolicy._get_constructor_parameters at 0x7f2faf816440>", "reset_noise": "<function ActorCriticPolicy.reset_noise at 0x7f2faf8164d0>", "_build_mlp_extractor": "<function ActorCriticPolicy._build_mlp_extractor at 0x7f2faf816560>", "_build": "<function ActorCriticPolicy._build at 0x7f2faf8165f0>", "forward": "<function ActorCriticPolicy.forward at 0x7f2faf816680>", "extract_features": "<function ActorCriticPolicy.extract_features at 0x7f2faf816710>", "_get_action_dist_from_latent": "<function ActorCriticPolicy._get_action_dist_from_latent at 0x7f2faf8167a0>", "_predict": "<function ActorCriticPolicy._predict at 0x7f2faf816830>", "evaluate_actions": "<function ActorCriticPolicy.evaluate_actions at 0x7f2faf8168c0>", "get_distribution": "<function ActorCriticPolicy.get_distribution at 0x7f2faf816950>", "predict_values": "<function ActorCriticPolicy.predict_values at 0x7f2faf8169e0>", "__abstractmethods__": "frozenset()", "_abc_impl": "<_abc._abc_data object at 0x7f2faf811b00>"}, "verbose": 1, "policy_kwargs": {}, "num_timesteps": 1000448, "_total_timesteps": 1000000, "_num_timesteps_at_start": 0, "seed": null, "action_noise": null, "start_time": 1688537446515482402, "learning_rate": 0.0003, "tensorboard_log": null, "_last_obs": {":type:": "<class 'numpy.ndarray'>", ":serialized:": "gAWVlQAAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJYgAAAAAAAAAOaQ2T2rT/E965/2vceImL4tUuu8OVyQugAAAAAAAAAAlIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUiYiHlFKUKEsDjAE8lE5OTkr/////Sv////9LAHSUYksBSwiGlIwBQ5R0lFKULg=="}, "_last_episode_starts": {":type:": "<class 'numpy.ndarray'>", ":serialized:": "gAWVdAAAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJYBAAAAAAAAAACUjAVudW1weZSMBWR0eXBllJOUjAJiMZSJiIeUUpQoSwOMAXyUTk5OSv////9K/////0sAdJRiSwGFlIwBQ5R0lFKULg=="}, "_last_original_obs": null, "_episode_num": 0, "use_sde": false, "sde_sample_freq": -1, "_current_progress_remaining": -0.00044800000000000395, "_stats_window_size": 100, "ep_info_buffer": {":type:": "<class 'collections.deque'>", ":serialized:": "gAWVKgwAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKUKH2UKIwBcpRHQFJa/+KjzqeMAWyUTRsBjAF0lEdAlBZQ6dUbUHV9lChoBkdAb52jFAE+xGgHTVUBaAhHQJQXiK/Efkp1fZQoaAZHQHEbNe2NNrVoB03jAWgIR0CUGlukk8ifdX2UKGgGR0BwJYVXV9WqaAdNxAFoCEdAlBwVdgOSXHV9lChoBkdAcHsNpM6BAmgHTbYBaAhHQJQeqVpsXSB1fZQoaAZHQEBcLOzIFNdoB0v1aAhHQJQfl7BwdbR1fZQoaAZHQG7K0ZFXq7loB01QAWgIR0CUIMvAGjbjdX2UKGgGR0BwD72bobGWaAdNYAJoCEdAlCSThHbypnV9lChoBkdAbqAzHCGetmgHTU0BaAhHQJQmIHlfZ291fZQoaAZHQD6eGnGbTc9oB00RAWgIR0CUKN1aGHpKdX2UKGgGR0A6IsPatcOcaAdNBQFoCEdAlCodD6WPcXV9lChoBkdARXM1jy4FzWgHTQ8BaAhHQJQrfN9ph4N1fZQoaAZHQD1+wNb1RLtoB0v/aAhHQJQssVGkN4J1fZQoaAZHQGvLLB0p3HJoB01VAWgIR0CUMAw35vcadX2UKGgGR0BwV7qFAVwhaAdNXAFoCEdAlDHV5jYqXnV9lChoBkdAaugIk7fYSWgHTYIBaAhHQJQz3k1dgOV1fZQoaAZHP86on8baRIVoB00NAWgIR0CUNtVoHs1LdX2UKGgGR0BQH+inHeabaAdL/2gIR0CUOCP8yeqadX2UKGgGR0BxzQHNX5nEaAdNVwFoCEdAlDmxMBZIQXV9lChoBkdANmd3bEgnt2gHS+ZoCEdAlDqFAu7HyXV9lChoBkdAauTd8iOea2gHTYUBaAhHQJQ9COktVaR1fZQoaAZHQEP66K+BYmtoB00HAWgIR0CUPf7fHggpdX2UKGgGR0ByIOavzOHGaAdNxQFoCEdAlD+rjtG/e3V9lChoBkdAcA7dLQHAymgHTXoCaAhHQJRDE+X7cfx1fZQoaAZHQHH+KG5+YtxoB01dAWgIR0CURF4qgAZLdX2UKGgGR0BD/cRDkU9IaAdNAwFoCEdAlEZS1uzhP3V9lChoBkdANJMIu5BkZ2gHS+FoCEdAlEcmMfigkHV9lChoBkdAcBw4wyqMnGgHTXABaAhHQJRIeErXlKd1fZQoaAZHQHABeA7PppxoB00mAWgIR0CUSowu/UONdX2UKGgGR8AwHCojv/ipaAdL0mgIR0CUS1EV32VWdX2UKGgGR0Bw5kQjD8+BaAdNwwFoCEdAlEz49LYf4nV9lChoBkdAUR62nbZezGgHTSwBaAhHQJROCl2vB8B1fZQoaAZHQEvpim2sq8VoB00mAWgIR0CUUCtCAtnPdX2UKGgGR0BFfXkgfU4JaAdL02gIR0CUUOq/ub7TdX2UKGgGR0BFDfRu0kWzaAdLzWgIR0CUUaea8YhudX2UKGgGR0Bw29nlGPPtaAdNnAFoCEdAlFQ1PWQOnXV9lChoBkdAb3/pKzzErGgHTWsBaAhHQJRVfT4L1Ep1fZQoaAZHQEuxJOnEVFhoB00hAWgIR0CUVokMkQf7dX2UKGgGR0BvckGJN0vHaAdNTQFoCEdAlFjMLncL0HV9lChoBkdAUZo2XLNfPWgHTRMBaAhHQJRZyq814xF1fZQoaAZHQEJ8qkM1CPZoB0vMaAhHQJRahNL127p1fZQoaAZHQECvSXMQmNRoB0vnaAhHQJRbXKFIuoR1fZQoaAZHQG8iVi4J/odoB01dAWgIR0CUXf9/BnBddX2UKGgGR0Bto251/2CeaAdNLwFoCEdAlF8URe1KG3V9lChoBkdAcDJ2oegctGgHTS4BaAhHQJRgOhTOxB51fZQoaAZHQHBxYGdI5HVoB01nAWgIR0CUYv6+nIhhdX2UKGgGR0BvOHVqesgdaAdNbwFoCEdAlGTE/8l5W3V9lChoBkdAbSpvoePq92gHTVwBaAhHQJRmfhCMPz51fZQoaAZHQG10cr7O3UhoB02EAWgIR0CUad4m1IAfdX2UKGgGR0BvSEm0E5hjaAdNOgFoCEdAlGto73fygHV9lChoBkdALdw9RrJr+GgHS/RoCEdAlGyfOUt7KXV9lChoBkdAbTFsFdLQHGgHTWgBaAhHQJRwAFY+0PZ1fZQoaAZHQGH4qMvRJEpoB03oA2gIR0CUdUqWC2+gdX2UKGgGR0BwQb7oB7u2aAdNSwFoCEdAlHhrQC0WuXV9lChoBkdAbaEx0MgEEGgHTYYBaAhHQJR52rHU+cJ1fZQoaAZHQFBAzSkTHsFoB00TAWgIR0CUevnO0LMLdX2UKGgGR0BvilndweeWaAdNbgFoCEdAlH1Mf3evZHV9lChoBkdANqSJ40Mw12gHS/JoCEdAlH4rNW2gF3V9lChoBkdAPEdFF2FFlWgHS+hoCEdAlH8b2pQ1rXV9lChoBkdATRaYTj/+9GgHS9loCEdAlH/ghStNjHV9lChoBkdAb7zpt78ejmgHTUIBaAhHQJSCDktEofF1fZQoaAZHQFGrsByS3b5oB00gAWgIR0CUgxSK3uuzdX2UKGgGR0Bw8XtShrWRaAdNNgFoCEdAlIQvDcdo4HV9lChoBkdAQKNiKBNEgGgHS/xoCEdAlIYX8wYcenV9lChoBkdANoOM+/xlQWgHTQIBaAhHQJSHBe7cwg11fZQoaAZHQDHlic5Ke05oB00LAWgIR0CUh/3fQ8fWdX2UKGgGR0BNZdb5dnkDaAdL+mgIR0CUiOJrtVrAdX2UKGgGR0BwrKVnmJWOaAdNbwFoCEdAlItCfDk2gnV9lChoBkfALz/C66J66mgHTQwBaAhHQJSMQOoYNy51fZQoaAZHQESE4J/oaDRoB0vyaAhHQJSNGyKNyYJ1fZQoaAZHQHK8veP7vXtoB00vAWgIR0CUjzLU1AJLdX2UKGgGR0BKoBppN9H+aAdNEgFoCEdAlJAqWkadc3V9lChoBkdAQZHQ6ZH/cWgHTRoBaAhHQJSRJrLyMDR1fZQoaAZHQEC/jdYW+GpoB0vXaAhHQJSR63H7xd91fZQoaAZHQDO52eQMhHNoB0vuaAhHQJSTvbfxc3V1fZQoaAZHQE3xXRPXTVloB0vmaAhHQJSUj7j1f3N1fZQoaAZHQG+SNATqSoxoB001AWgIR0CUlaw1R+BpdX2UKGgGR0BAj4BeXzDoaAdL8mgIR0CUloh60IC2dX2UKGgGR0BurustCiRGaAdNhgFoCEdAlJjmbb1yvXV9lChoBke/7bh1klNUO2gHS/xoCEdAlJnVNcnmaHV9lChoBkdAcA+WNFSbY2gHTS8BaAhHQJSa7TWoWHl1fZQoaAZHQC2mwNb1RLtoB0v5aAhHQJSb0D8tPHl1fZQoaAZHQHEcDlHSWqtoB01sAWgIR0CUniFMZgogdX2UKGgGR0BwCqP5pJwsaAdNOgFoCEdAlJ9CgCfYjHV9lChoBkdAPPHtjTa0yGgHTQoBaAhHQJSgMvHtF8Z1fZQoaAZHQHAsu8XenAJoB02ZAWgIR0CUo+mP5pJxdX2UKGgGR0BDDaMrEtNBaAdL/GgIR0CUpRPva11GdX2UKGgGR0Bv68gfU4JeaAdNeAFoCEdAlKbnB1s+FHV9lChoBkdAcmPbtqpLmWgHTfABaAhHQJSqvvOQhfV1fZQoaAZHQEc3nyNGViZoB00FAWgIR0CUrB7XxvvSdX2UKGgGR0Aj2912aDwpaAdL9WgIR0CUrVlwcYIjdX2UKGgGR0Bx0XMqz7djaAdNOgFoCEdAlLDWhAWznnV9lChoBkdAb9kHuZ1FIGgHTUoBaAhHQJSysPd2xIJ1fZQoaAZHQHFVNSZSeiBoB001AWgIR0CUtHrdnCfpdX2UKGgGR0Bwu7ww0wajaAdNTwFoCEdAlLeNuk1uSHV9lChoBkdAcWMhPCVKPGgHTWMBaAhHQJS44plSS/11fZQoaAZHP+dwgDA8B+5oB0v2aAhHQJS5vfXPJJZ1fZQoaAZHQHCtMCgbp/xoB01kAWgIR0CUvAldC3PSdX2UKGgGR0Bw6LRsuWa+aAdNPwFoCEdAlL0teIEbHnV9lChoBkdAcLujOs1baGgHTacBaAhHQJS+uXE61b91ZS4="}, "ep_success_buffer": {":type:": "<class 'collections.deque'>", ":serialized:": "gAWVIAAAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKULg=="}, "_n_updates": 3908, "observation_space": {":type:": "<class 'gymnasium.spaces.box.Box'>", ":serialized:": "gAWVcAIAAAAAAACMFGd5bW5hc2l1bS5zcGFjZXMuYm94lIwDQm94lJOUKYGUfZQojAVkdHlwZZSMBW51bXB5lGgFk5SMAmY0lImIh5RSlChLA4wBPJROTk5K/////0r/////SwB0lGKMDWJvdW5kZWRfYmVsb3eUjBJudW1weS5jb3JlLm51bWVyaWOUjAtfZnJvbWJ1ZmZlcpSTlCiWCAAAAAAAAAABAQEBAQEBAZRoB4wCYjGUiYiHlFKUKEsDjAF8lE5OTkr/////Sv////9LAHSUYksIhZSMAUOUdJRSlIwNYm91bmRlZF9hYm92ZZRoECiWCAAAAAAAAAABAQEBAQEBAZRoFEsIhZRoGHSUUpSMBl9zaGFwZZRLCIWUjANsb3eUaBAoliAAAAAAAAAAAAC0wgAAtMIAAKDAAACgwNsPScAAAKDAAAAAgAAAAICUaApLCIWUaBh0lFKUjARoaWdolGgQKJYgAAAAAAAAAAAAtEIAALRCAACgQAAAoEDbD0lAAACgQAAAgD8AAIA/lGgKSwiFlGgYdJRSlIwIbG93X3JlcHKUjFtbLTkwLiAgICAgICAgLTkwLiAgICAgICAgIC01LiAgICAgICAgIC01LiAgICAgICAgIC0zLjE0MTU5MjcgIC01LgogIC0wLiAgICAgICAgIC0wLiAgICAgICBdlIwJaGlnaF9yZXBylIxTWzkwLiAgICAgICAgOTAuICAgICAgICAgNS4gICAgICAgICA1LiAgICAgICAgIDMuMTQxNTkyNyAgNS4KICAxLiAgICAgICAgIDEuICAgICAgIF2UjApfbnBfcmFuZG9tlE51Yi4=", "dtype": "float32", "bounded_below": "[ True True True True True True True True]", "bounded_above": "[ True True True True True True True True]", "_shape": [8], "low": "[-90. -90. -5. -5. -3.1415927 -5.\n -0. -0. ]", "high": "[90. 90. 5. 5. 3.1415927 5.\n 1. 1. ]", "low_repr": "[-90. -90. -5. -5. -3.1415927 -5.\n -0. -0. ]", "high_repr": "[90. 90. 5. 5. 3.1415927 5.\n 1. 1. ]", "_np_random": null}, "action_space": {":type:": "<class 'gymnasium.spaces.discrete.Discrete'>", ":serialized:": "gAWV1QAAAAAAAACMGWd5bW5hc2l1bS5zcGFjZXMuZGlzY3JldGWUjAhEaXNjcmV0ZZSTlCmBlH2UKIwBbpSMFW51bXB5LmNvcmUubXVsdGlhcnJheZSMBnNjYWxhcpSTlIwFbnVtcHmUjAVkdHlwZZSTlIwCaTiUiYiHlFKUKEsDjAE8lE5OTkr/////Sv////9LAHSUYkMIBAAAAAAAAACUhpRSlIwFc3RhcnSUaAhoDkMIAAAAAAAAAACUhpRSlIwGX3NoYXBllCloCmgOjApfbnBfcmFuZG9tlE51Yi4=", "n": "4", "start": "0", "_shape": [], "dtype": "int64", "_np_random": null}, "n_envs": 1, "n_steps": 1024, "gamma": 0.999, "gae_lambda": 0.98, "ent_coef": 0.01, "vf_coef": 0.5, "max_grad_norm": 0.5, "batch_size": 64, "n_epochs": 4, "clip_range": {":type:": "<class 'function'>", ":serialized:": "gAWVxQIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwOX21ha2VfZnVuY3Rpb26Uk5QoaACMDV9idWlsdGluX3R5cGWUk5SMCENvZGVUeXBllIWUUpQoSwFLAEsASwFLAUsTQwSIAFMAlE6FlCmMAV+UhZSMSS91c3IvbG9jYWwvbGliL3B5dGhvbjMuMTAvZGlzdC1wYWNrYWdlcy9zdGFibGVfYmFzZWxpbmVzMy9jb21tb24vdXRpbHMucHmUjARmdW5jlEuEQwIEAZSMA3ZhbJSFlCl0lFKUfZQojAtfX3BhY2thZ2VfX5SMGHN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbpSMCF9fbmFtZV9flIwec3RhYmxlX2Jhc2VsaW5lczMuY29tbW9uLnV0aWxzlIwIX19maWxlX1+UjEkvdXNyL2xvY2FsL2xpYi9weXRob24zLjEwL2Rpc3QtcGFja2FnZXMvc3RhYmxlX2Jhc2VsaW5lczMvY29tbW9uL3V0aWxzLnB5lHVOTmgAjBBfbWFrZV9lbXB0eV9jZWxslJOUKVKUhZR0lFKUjBxjbG91ZHBpY2tsZS5jbG91ZHBpY2tsZV9mYXN0lIwSX2Z1bmN0aW9uX3NldHN0YXRllJOUaB99lH2UKGgWaA2MDF9fcXVhbG5hbWVfX5SMGWNvbnN0YW50X2ZuLjxsb2NhbHM+LmZ1bmOUjA9fX2Fubm90YXRpb25zX1+UfZSMDl9fa3dkZWZhdWx0c19flE6MDF9fZGVmYXVsdHNfX5ROjApfX21vZHVsZV9flGgXjAdfX2RvY19flE6MC19fY2xvc3VyZV9flGgAjApfbWFrZV9jZWxslJOURz/JmZmZmZmahZRSlIWUjBdfY2xvdWRwaWNrbGVfc3VibW9kdWxlc5RdlIwLX19nbG9iYWxzX1+UfZR1hpSGUjAu"}, "clip_range_vf": null, "normalize_advantage": true, "target_kl": null, "lr_schedule": {":type:": "<class 'function'>", ":serialized:": "gAWVxQIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwOX21ha2VfZnVuY3Rpb26Uk5QoaACMDV9idWlsdGluX3R5cGWUk5SMCENvZGVUeXBllIWUUpQoSwFLAEsASwFLAUsTQwSIAFMAlE6FlCmMAV+UhZSMSS91c3IvbG9jYWwvbGliL3B5dGhvbjMuMTAvZGlzdC1wYWNrYWdlcy9zdGFibGVfYmFzZWxpbmVzMy9jb21tb24vdXRpbHMucHmUjARmdW5jlEuEQwIEAZSMA3ZhbJSFlCl0lFKUfZQojAtfX3BhY2thZ2VfX5SMGHN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbpSMCF9fbmFtZV9flIwec3RhYmxlX2Jhc2VsaW5lczMuY29tbW9uLnV0aWxzlIwIX19maWxlX1+UjEkvdXNyL2xvY2FsL2xpYi9weXRob24zLjEwL2Rpc3QtcGFja2FnZXMvc3RhYmxlX2Jhc2VsaW5lczMvY29tbW9uL3V0aWxzLnB5lHVOTmgAjBBfbWFrZV9lbXB0eV9jZWxslJOUKVKUhZR0lFKUjBxjbG91ZHBpY2tsZS5jbG91ZHBpY2tsZV9mYXN0lIwSX2Z1bmN0aW9uX3NldHN0YXRllJOUaB99lH2UKGgWaA2MDF9fcXVhbG5hbWVfX5SMGWNvbnN0YW50X2ZuLjxsb2NhbHM+LmZ1bmOUjA9fX2Fubm90YXRpb25zX1+UfZSMDl9fa3dkZWZhdWx0c19flE6MDF9fZGVmYXVsdHNfX5ROjApfX21vZHVsZV9flGgXjAdfX2RvY19flE6MC19fY2xvc3VyZV9flGgAjApfbWFrZV9jZWxslJOURz8zqSowVTJhhZRSlIWUjBdfY2xvdWRwaWNrbGVfc3VibW9kdWxlc5RdlIwLX19nbG9iYWxzX1+UfZR1hpSGUjAu"}, "system_info": {"OS": "Linux-5.15.107+-x86_64-with-glibc2.31 # 1 SMP Sat Apr 29 09:15:28 UTC 2023", "Python": "3.10.12", "Stable-Baselines3": "2.0.0a5", "PyTorch": "2.0.1+cu118", "GPU Enabled": "False", "Numpy": "1.22.4", "Cloudpickle": "2.2.1", "Gymnasium": "0.28.1", "OpenAI Gym": "0.25.2"}}
|
lunarLander.zip
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7141e0a048f0a0002b023db5f446477d5a5476472860280235c36c69f11b88b1
|
| 3 |
+
size 145554
|
lunarLander/data
CHANGED
|
@@ -4,34 +4,34 @@
|
|
| 4 |
":serialized:": "gAWVOwAAAAAAAACMIXN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi5wb2xpY2llc5SMEUFjdG9yQ3JpdGljUG9saWN5lJOULg==",
|
| 5 |
"__module__": "stable_baselines3.common.policies",
|
| 6 |
"__doc__": "\n Policy class for actor-critic algorithms (has both policy and value prediction).\n Used by A2C, PPO and the likes.\n\n :param observation_space: Observation space\n :param action_space: Action space\n :param lr_schedule: Learning rate schedule (could be constant)\n :param net_arch: The specification of the policy and value networks.\n :param activation_fn: Activation function\n :param ortho_init: Whether to use or not orthogonal initialization\n :param use_sde: Whether to use State Dependent Exploration or not\n :param log_std_init: Initial value for the log standard deviation\n :param full_std: Whether to use (n_features x n_actions) parameters\n for the std instead of only (n_features,) when using gSDE\n :param use_expln: Use ``expln()`` function instead of ``exp()`` to ensure\n a positive standard deviation (cf paper). It allows to keep variance\n above zero and prevent it from growing too fast. In practice, ``exp()`` is usually enough.\n :param squash_output: Whether to squash the output using a tanh function,\n this allows to ensure boundaries when using gSDE.\n :param features_extractor_class: Features extractor to use.\n :param features_extractor_kwargs: Keyword arguments\n to pass to the features extractor.\n :param share_features_extractor: If True, the features extractor is shared between the policy and value networks.\n :param normalize_images: Whether to normalize images or not,\n dividing by 255.0 (True by default)\n :param optimizer_class: The optimizer to use,\n ``th.optim.Adam`` by default\n :param optimizer_kwargs: Additional keyword arguments,\n excluding the learning rate, to pass to the optimizer\n ",
|
| 7 |
-
"__init__": "<function ActorCriticPolicy.__init__ at
|
| 8 |
-
"_get_constructor_parameters": "<function ActorCriticPolicy._get_constructor_parameters at
|
| 9 |
-
"reset_noise": "<function ActorCriticPolicy.reset_noise at
|
| 10 |
-
"_build_mlp_extractor": "<function ActorCriticPolicy._build_mlp_extractor at
|
| 11 |
-
"_build": "<function ActorCriticPolicy._build at
|
| 12 |
-
"forward": "<function ActorCriticPolicy.forward at
|
| 13 |
-
"extract_features": "<function ActorCriticPolicy.extract_features at
|
| 14 |
-
"_get_action_dist_from_latent": "<function ActorCriticPolicy._get_action_dist_from_latent at
|
| 15 |
-
"_predict": "<function ActorCriticPolicy._predict at
|
| 16 |
-
"evaluate_actions": "<function ActorCriticPolicy.evaluate_actions at
|
| 17 |
-
"get_distribution": "<function ActorCriticPolicy.get_distribution at
|
| 18 |
-
"predict_values": "<function ActorCriticPolicy.predict_values at
|
| 19 |
"__abstractmethods__": "frozenset()",
|
| 20 |
-
"_abc_impl": "<_abc._abc_data object at
|
| 21 |
},
|
| 22 |
"verbose": 1,
|
| 23 |
"policy_kwargs": {},
|
| 24 |
-
"num_timesteps":
|
| 25 |
"_total_timesteps": 1000000,
|
| 26 |
"_num_timesteps_at_start": 0,
|
| 27 |
"seed": null,
|
| 28 |
"action_noise": null,
|
| 29 |
-
"start_time":
|
| 30 |
"learning_rate": 0.0003,
|
| 31 |
"tensorboard_log": null,
|
| 32 |
"_last_obs": {
|
| 33 |
":type:": "<class 'numpy.ndarray'>",
|
| 34 |
-
":serialized:": "
|
| 35 |
},
|
| 36 |
"_last_episode_starts": {
|
| 37 |
":type:": "<class 'numpy.ndarray'>",
|
|
@@ -41,17 +41,17 @@
|
|
| 41 |
"_episode_num": 0,
|
| 42 |
"use_sde": false,
|
| 43 |
"sde_sample_freq": -1,
|
| 44 |
-
"_current_progress_remaining": 0.
|
| 45 |
"_stats_window_size": 100,
|
| 46 |
"ep_info_buffer": {
|
| 47 |
":type:": "<class 'collections.deque'>",
|
| 48 |
-
":serialized:": "
|
| 49 |
},
|
| 50 |
"ep_success_buffer": {
|
| 51 |
":type:": "<class 'collections.deque'>",
|
| 52 |
":serialized:": "gAWVIAAAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKULg=="
|
| 53 |
},
|
| 54 |
-
"_n_updates":
|
| 55 |
"observation_space": {
|
| 56 |
":type:": "<class 'gymnasium.spaces.box.Box'>",
|
| 57 |
":serialized:": "gAWVcAIAAAAAAACMFGd5bW5hc2l1bS5zcGFjZXMuYm94lIwDQm94lJOUKYGUfZQojAVkdHlwZZSMBW51bXB5lGgFk5SMAmY0lImIh5RSlChLA4wBPJROTk5K/////0r/////SwB0lGKMDWJvdW5kZWRfYmVsb3eUjBJudW1weS5jb3JlLm51bWVyaWOUjAtfZnJvbWJ1ZmZlcpSTlCiWCAAAAAAAAAABAQEBAQEBAZRoB4wCYjGUiYiHlFKUKEsDjAF8lE5OTkr/////Sv////9LAHSUYksIhZSMAUOUdJRSlIwNYm91bmRlZF9hYm92ZZRoECiWCAAAAAAAAAABAQEBAQEBAZRoFEsIhZRoGHSUUpSMBl9zaGFwZZRLCIWUjANsb3eUaBAoliAAAAAAAAAAAAC0wgAAtMIAAKDAAACgwNsPScAAAKDAAAAAgAAAAICUaApLCIWUaBh0lFKUjARoaWdolGgQKJYgAAAAAAAAAAAAtEIAALRCAACgQAAAoEDbD0lAAACgQAAAgD8AAIA/lGgKSwiFlGgYdJRSlIwIbG93X3JlcHKUjFtbLTkwLiAgICAgICAgLTkwLiAgICAgICAgIC01LiAgICAgICAgIC01LiAgICAgICAgIC0zLjE0MTU5MjcgIC01LgogIC0wLiAgICAgICAgIC0wLiAgICAgICBdlIwJaGlnaF9yZXBylIxTWzkwLiAgICAgICAgOTAuICAgICAgICAgNS4gICAgICAgICA1LiAgICAgICAgIDMuMTQxNTkyNyAgNS4KICAxLiAgICAgICAgIDEuICAgICAgIF2UjApfbnBfcmFuZG9tlE51Yi4=",
|
|
|
|
| 4 |
":serialized:": "gAWVOwAAAAAAAACMIXN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi5wb2xpY2llc5SMEUFjdG9yQ3JpdGljUG9saWN5lJOULg==",
|
| 5 |
"__module__": "stable_baselines3.common.policies",
|
| 6 |
"__doc__": "\n Policy class for actor-critic algorithms (has both policy and value prediction).\n Used by A2C, PPO and the likes.\n\n :param observation_space: Observation space\n :param action_space: Action space\n :param lr_schedule: Learning rate schedule (could be constant)\n :param net_arch: The specification of the policy and value networks.\n :param activation_fn: Activation function\n :param ortho_init: Whether to use or not orthogonal initialization\n :param use_sde: Whether to use State Dependent Exploration or not\n :param log_std_init: Initial value for the log standard deviation\n :param full_std: Whether to use (n_features x n_actions) parameters\n for the std instead of only (n_features,) when using gSDE\n :param use_expln: Use ``expln()`` function instead of ``exp()`` to ensure\n a positive standard deviation (cf paper). It allows to keep variance\n above zero and prevent it from growing too fast. In practice, ``exp()`` is usually enough.\n :param squash_output: Whether to squash the output using a tanh function,\n this allows to ensure boundaries when using gSDE.\n :param features_extractor_class: Features extractor to use.\n :param features_extractor_kwargs: Keyword arguments\n to pass to the features extractor.\n :param share_features_extractor: If True, the features extractor is shared between the policy and value networks.\n :param normalize_images: Whether to normalize images or not,\n dividing by 255.0 (True by default)\n :param optimizer_class: The optimizer to use,\n ``th.optim.Adam`` by default\n :param optimizer_kwargs: Additional keyword arguments,\n excluding the learning rate, to pass to the optimizer\n ",
|
| 7 |
+
"__init__": "<function ActorCriticPolicy.__init__ at 0x7f2faf8163b0>",
|
| 8 |
+
"_get_constructor_parameters": "<function ActorCriticPolicy._get_constructor_parameters at 0x7f2faf816440>",
|
| 9 |
+
"reset_noise": "<function ActorCriticPolicy.reset_noise at 0x7f2faf8164d0>",
|
| 10 |
+
"_build_mlp_extractor": "<function ActorCriticPolicy._build_mlp_extractor at 0x7f2faf816560>",
|
| 11 |
+
"_build": "<function ActorCriticPolicy._build at 0x7f2faf8165f0>",
|
| 12 |
+
"forward": "<function ActorCriticPolicy.forward at 0x7f2faf816680>",
|
| 13 |
+
"extract_features": "<function ActorCriticPolicy.extract_features at 0x7f2faf816710>",
|
| 14 |
+
"_get_action_dist_from_latent": "<function ActorCriticPolicy._get_action_dist_from_latent at 0x7f2faf8167a0>",
|
| 15 |
+
"_predict": "<function ActorCriticPolicy._predict at 0x7f2faf816830>",
|
| 16 |
+
"evaluate_actions": "<function ActorCriticPolicy.evaluate_actions at 0x7f2faf8168c0>",
|
| 17 |
+
"get_distribution": "<function ActorCriticPolicy.get_distribution at 0x7f2faf816950>",
|
| 18 |
+
"predict_values": "<function ActorCriticPolicy.predict_values at 0x7f2faf8169e0>",
|
| 19 |
"__abstractmethods__": "frozenset()",
|
| 20 |
+
"_abc_impl": "<_abc._abc_data object at 0x7f2faf811b00>"
|
| 21 |
},
|
| 22 |
"verbose": 1,
|
| 23 |
"policy_kwargs": {},
|
| 24 |
+
"num_timesteps": 1000448,
|
| 25 |
"_total_timesteps": 1000000,
|
| 26 |
"_num_timesteps_at_start": 0,
|
| 27 |
"seed": null,
|
| 28 |
"action_noise": null,
|
| 29 |
+
"start_time": 1688537446515482402,
|
| 30 |
"learning_rate": 0.0003,
|
| 31 |
"tensorboard_log": null,
|
| 32 |
"_last_obs": {
|
| 33 |
":type:": "<class 'numpy.ndarray'>",
|
| 34 |
+
":serialized:": "gAWVlQAAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJYgAAAAAAAAAOaQ2T2rT/E965/2vceImL4tUuu8OVyQugAAAAAAAAAAlIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUiYiHlFKUKEsDjAE8lE5OTkr/////Sv////9LAHSUYksBSwiGlIwBQ5R0lFKULg=="
|
| 35 |
},
|
| 36 |
"_last_episode_starts": {
|
| 37 |
":type:": "<class 'numpy.ndarray'>",
|
|
|
|
| 41 |
"_episode_num": 0,
|
| 42 |
"use_sde": false,
|
| 43 |
"sde_sample_freq": -1,
|
| 44 |
+
"_current_progress_remaining": -0.00044800000000000395,
|
| 45 |
"_stats_window_size": 100,
|
| 46 |
"ep_info_buffer": {
|
| 47 |
":type:": "<class 'collections.deque'>",
|
| 48 |
+
":serialized:": "gAWVKgwAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKUKH2UKIwBcpRHQFJa/+KjzqeMAWyUTRsBjAF0lEdAlBZQ6dUbUHV9lChoBkdAb52jFAE+xGgHTVUBaAhHQJQXiK/Efkp1fZQoaAZHQHEbNe2NNrVoB03jAWgIR0CUGlukk8ifdX2UKGgGR0BwJYVXV9WqaAdNxAFoCEdAlBwVdgOSXHV9lChoBkdAcHsNpM6BAmgHTbYBaAhHQJQeqVpsXSB1fZQoaAZHQEBcLOzIFNdoB0v1aAhHQJQfl7BwdbR1fZQoaAZHQG7K0ZFXq7loB01QAWgIR0CUIMvAGjbjdX2UKGgGR0BwD72bobGWaAdNYAJoCEdAlCSThHbypnV9lChoBkdAbqAzHCGetmgHTU0BaAhHQJQmIHlfZ291fZQoaAZHQD6eGnGbTc9oB00RAWgIR0CUKN1aGHpKdX2UKGgGR0A6IsPatcOcaAdNBQFoCEdAlCodD6WPcXV9lChoBkdARXM1jy4FzWgHTQ8BaAhHQJQrfN9ph4N1fZQoaAZHQD1+wNb1RLtoB0v/aAhHQJQssVGkN4J1fZQoaAZHQGvLLB0p3HJoB01VAWgIR0CUMAw35vcadX2UKGgGR0BwV7qFAVwhaAdNXAFoCEdAlDHV5jYqXnV9lChoBkdAaugIk7fYSWgHTYIBaAhHQJQz3k1dgOV1fZQoaAZHP86on8baRIVoB00NAWgIR0CUNtVoHs1LdX2UKGgGR0BQH+inHeabaAdL/2gIR0CUOCP8yeqadX2UKGgGR0BxzQHNX5nEaAdNVwFoCEdAlDmxMBZIQXV9lChoBkdANmd3bEgnt2gHS+ZoCEdAlDqFAu7HyXV9lChoBkdAauTd8iOea2gHTYUBaAhHQJQ9COktVaR1fZQoaAZHQEP66K+BYmtoB00HAWgIR0CUPf7fHggpdX2UKGgGR0ByIOavzOHGaAdNxQFoCEdAlD+rjtG/e3V9lChoBkdAcA7dLQHAymgHTXoCaAhHQJRDE+X7cfx1fZQoaAZHQHH+KG5+YtxoB01dAWgIR0CURF4qgAZLdX2UKGgGR0BD/cRDkU9IaAdNAwFoCEdAlEZS1uzhP3V9lChoBkdANJMIu5BkZ2gHS+FoCEdAlEcmMfigkHV9lChoBkdAcBw4wyqMnGgHTXABaAhHQJRIeErXlKd1fZQoaAZHQHABeA7PppxoB00mAWgIR0CUSowu/UONdX2UKGgGR8AwHCojv/ipaAdL0mgIR0CUS1EV32VWdX2UKGgGR0Bw5kQjD8+BaAdNwwFoCEdAlEz49LYf4nV9lChoBkdAUR62nbZezGgHTSwBaAhHQJROCl2vB8B1fZQoaAZHQEvpim2sq8VoB00mAWgIR0CUUCtCAtnPdX2UKGgGR0BFfXkgfU4JaAdL02gIR0CUUOq/ub7TdX2UKGgGR0BFDfRu0kWzaAdLzWgIR0CUUaea8YhudX2UKGgGR0Bw29nlGPPtaAdNnAFoCEdAlFQ1PWQOnXV9lChoBkdAb3/pKzzErGgHTWsBaAhHQJRVfT4L1Ep1fZQoaAZHQEuxJOnEVFhoB00hAWgIR0CUVokMkQf7dX2UKGgGR0BvckGJN0vHaAdNTQFoCEdAlFjMLncL0HV9lChoBkdAUZo2XLNfPWgHTRMBaAhHQJRZyq814xF1fZQoaAZHQEJ8qkM1CPZoB0vMaAhHQJRahNL127p1fZQoaAZHQECvSXMQmNRoB0vnaAhHQJRbXKFIuoR1fZQoaAZHQG8iVi4J/odoB01dAWgIR0CUXf9/BnBddX2UKGgGR0Bto251/2CeaAdNLwFoCEdAlF8URe1KG3V9lChoBkdAcDJ2oegctGgHTS4BaAhHQJRgOhTOxB51fZQoaAZHQHBxYGdI5HVoB01nAWgIR0CUYv6+nIhhdX2UKGgGR0BvOHVqesgdaAdNbwFoCEdAlGTE/8l5W3V9lChoBkdAbSpvoePq92gHTVwBaAhHQJRmfhCMPz51fZQoaAZHQG10cr7O3UhoB02EAWgIR0CUad4m1IAfdX2UKGgGR0BvSEm0E5hjaAdNOgFoCEdAlGto73fygHV9lChoBkdALdw9RrJr+GgHS/RoCEdAlGyfOUt7KXV9lChoBkdAbTFsFdLQHGgHTWgBaAhHQJRwAFY+0PZ1fZQoaAZHQGH4qMvRJEpoB03oA2gIR0CUdUqWC2+gdX2UKGgGR0BwQb7oB7u2aAdNSwFoCEdAlHhrQC0WuXV9lChoBkdAbaEx0MgEEGgHTYYBaAhHQJR52rHU+cJ1fZQoaAZHQFBAzSkTHsFoB00TAWgIR0CUevnO0LMLdX2UKGgGR0BvilndweeWaAdNbgFoCEdAlH1Mf3evZHV9lChoBkdANqSJ40Mw12gHS/JoCEdAlH4rNW2gF3V9lChoBkdAPEdFF2FFlWgHS+hoCEdAlH8b2pQ1rXV9lChoBkdATRaYTj/+9GgHS9loCEdAlH/ghStNjHV9lChoBkdAb7zpt78ejmgHTUIBaAhHQJSCDktEofF1fZQoaAZHQFGrsByS3b5oB00gAWgIR0CUgxSK3uuzdX2UKGgGR0Bw8XtShrWRaAdNNgFoCEdAlIQvDcdo4HV9lChoBkdAQKNiKBNEgGgHS/xoCEdAlIYX8wYcenV9lChoBkdANoOM+/xlQWgHTQIBaAhHQJSHBe7cwg11fZQoaAZHQDHlic5Ke05oB00LAWgIR0CUh/3fQ8fWdX2UKGgGR0BNZdb5dnkDaAdL+mgIR0CUiOJrtVrAdX2UKGgGR0BwrKVnmJWOaAdNbwFoCEdAlItCfDk2gnV9lChoBkfALz/C66J66mgHTQwBaAhHQJSMQOoYNy51fZQoaAZHQESE4J/oaDRoB0vyaAhHQJSNGyKNyYJ1fZQoaAZHQHK8veP7vXtoB00vAWgIR0CUjzLU1AJLdX2UKGgGR0BKoBppN9H+aAdNEgFoCEdAlJAqWkadc3V9lChoBkdAQZHQ6ZH/cWgHTRoBaAhHQJSRJrLyMDR1fZQoaAZHQEC/jdYW+GpoB0vXaAhHQJSR63H7xd91fZQoaAZHQDO52eQMhHNoB0vuaAhHQJSTvbfxc3V1fZQoaAZHQE3xXRPXTVloB0vmaAhHQJSUj7j1f3N1fZQoaAZHQG+SNATqSoxoB001AWgIR0CUlaw1R+BpdX2UKGgGR0BAj4BeXzDoaAdL8mgIR0CUloh60IC2dX2UKGgGR0BurustCiRGaAdNhgFoCEdAlJjmbb1yvXV9lChoBke/7bh1klNUO2gHS/xoCEdAlJnVNcnmaHV9lChoBkdAcA+WNFSbY2gHTS8BaAhHQJSa7TWoWHl1fZQoaAZHQC2mwNb1RLtoB0v5aAhHQJSb0D8tPHl1fZQoaAZHQHEcDlHSWqtoB01sAWgIR0CUniFMZgogdX2UKGgGR0BwCqP5pJwsaAdNOgFoCEdAlJ9CgCfYjHV9lChoBkdAPPHtjTa0yGgHTQoBaAhHQJSgMvHtF8Z1fZQoaAZHQHAsu8XenAJoB02ZAWgIR0CUo+mP5pJxdX2UKGgGR0BDDaMrEtNBaAdL/GgIR0CUpRPva11GdX2UKGgGR0Bv68gfU4JeaAdNeAFoCEdAlKbnB1s+FHV9lChoBkdAcmPbtqpLmWgHTfABaAhHQJSqvvOQhfV1fZQoaAZHQEc3nyNGViZoB00FAWgIR0CUrB7XxvvSdX2UKGgGR0Aj2912aDwpaAdL9WgIR0CUrVlwcYIjdX2UKGgGR0Bx0XMqz7djaAdNOgFoCEdAlLDWhAWznnV9lChoBkdAb9kHuZ1FIGgHTUoBaAhHQJSysPd2xIJ1fZQoaAZHQHFVNSZSeiBoB001AWgIR0CUtHrdnCfpdX2UKGgGR0Bwu7ww0wajaAdNTwFoCEdAlLeNuk1uSHV9lChoBkdAcWMhPCVKPGgHTWMBaAhHQJS44plSS/11fZQoaAZHP+dwgDA8B+5oB0v2aAhHQJS5vfXPJJZ1fZQoaAZHQHCtMCgbp/xoB01kAWgIR0CUvAldC3PSdX2UKGgGR0Bw6LRsuWa+aAdNPwFoCEdAlL0teIEbHnV9lChoBkdAcLujOs1baGgHTacBaAhHQJS+uXE61b91ZS4="
|
| 49 |
},
|
| 50 |
"ep_success_buffer": {
|
| 51 |
":type:": "<class 'collections.deque'>",
|
| 52 |
":serialized:": "gAWVIAAAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKULg=="
|
| 53 |
},
|
| 54 |
+
"_n_updates": 3908,
|
| 55 |
"observation_space": {
|
| 56 |
":type:": "<class 'gymnasium.spaces.box.Box'>",
|
| 57 |
":serialized:": "gAWVcAIAAAAAAACMFGd5bW5hc2l1bS5zcGFjZXMuYm94lIwDQm94lJOUKYGUfZQojAVkdHlwZZSMBW51bXB5lGgFk5SMAmY0lImIh5RSlChLA4wBPJROTk5K/////0r/////SwB0lGKMDWJvdW5kZWRfYmVsb3eUjBJudW1weS5jb3JlLm51bWVyaWOUjAtfZnJvbWJ1ZmZlcpSTlCiWCAAAAAAAAAABAQEBAQEBAZRoB4wCYjGUiYiHlFKUKEsDjAF8lE5OTkr/////Sv////9LAHSUYksIhZSMAUOUdJRSlIwNYm91bmRlZF9hYm92ZZRoECiWCAAAAAAAAAABAQEBAQEBAZRoFEsIhZRoGHSUUpSMBl9zaGFwZZRLCIWUjANsb3eUaBAoliAAAAAAAAAAAAC0wgAAtMIAAKDAAACgwNsPScAAAKDAAAAAgAAAAICUaApLCIWUaBh0lFKUjARoaWdolGgQKJYgAAAAAAAAAAAAtEIAALRCAACgQAAAoEDbD0lAAACgQAAAgD8AAIA/lGgKSwiFlGgYdJRSlIwIbG93X3JlcHKUjFtbLTkwLiAgICAgICAgLTkwLiAgICAgICAgIC01LiAgICAgICAgIC01LiAgICAgICAgIC0zLjE0MTU5MjcgIC01LgogIC0wLiAgICAgICAgIC0wLiAgICAgICBdlIwJaGlnaF9yZXBylIxTWzkwLiAgICAgICAgOTAuICAgICAgICAgNS4gICAgICAgICA1LiAgICAgICAgIDMuMTQxNTkyNyAgNS4KICAxLiAgICAgICAgIDEuICAgICAgIF2UjApfbnBfcmFuZG9tlE51Yi4=",
|
lunarLander/policy.optimizer.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 87545
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:fdc7fd324c9ab9e15eccbcba27d3bd3abe2843231cb08cd1d34208b503ee7e09
|
| 3 |
size 87545
|
lunarLander/policy.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 43201
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6d74e8e0afe568fcfeb47abe60d6944af18ca75457efd8efb8df71c8c299f40e
|
| 3 |
size 43201
|
replay.mp4
CHANGED
|
Binary files a/replay.mp4 and b/replay.mp4 differ
|
|
|
results.json
CHANGED
|
@@ -1 +1 @@
|
|
| 1 |
-
{"mean_reward":
|
|
|
|
| 1 |
+
{"mean_reward": 249.29787700000003, "std_reward": 14.060509870680184, "is_deterministic": true, "n_eval_episodes": 10, "eval_datetime": "2023-07-05T06:36:26.947051"}
|