Upload PPO LunarLander-v2 trained agent 2nd
Browse files- README.md +1 -1
- config.json +1 -1
- ppo-LunarLander-v2.zip +2 -2
- ppo-LunarLander-v2/data +23 -23
- ppo-LunarLander-v2/policy.optimizer.pth +1 -1
- ppo-LunarLander-v2/policy.pth +1 -1
- replay.mp4 +0 -0
- results.json +1 -1
README.md
CHANGED
|
@@ -16,7 +16,7 @@ model-index:
|
|
| 16 |
type: LunarLander-v2
|
| 17 |
metrics:
|
| 18 |
- type: mean_reward
|
| 19 |
-
value:
|
| 20 |
name: mean_reward
|
| 21 |
verified: false
|
| 22 |
---
|
|
|
|
| 16 |
type: LunarLander-v2
|
| 17 |
metrics:
|
| 18 |
- type: mean_reward
|
| 19 |
+
value: 278.45 +/- 22.08
|
| 20 |
name: mean_reward
|
| 21 |
verified: false
|
| 22 |
---
|
config.json
CHANGED
|
@@ -1 +1 @@
|
|
| 1 |
-
{"policy_class": {":type:": "<class 'abc.ABCMeta'>", ":serialized:": "gAWVOwAAAAAAAACMIXN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi5wb2xpY2llc5SMEUFjdG9yQ3JpdGljUG9saWN5lJOULg==", "__module__": "stable_baselines3.common.policies", "__doc__": "\n Policy class for actor-critic algorithms (has both policy and value prediction).\n Used by A2C, PPO and the likes.\n\n :param observation_space: Observation space\n :param action_space: Action space\n :param lr_schedule: Learning rate schedule (could be constant)\n :param net_arch: The specification of the policy and value networks.\n :param activation_fn: Activation function\n :param ortho_init: Whether to use or not orthogonal initialization\n :param use_sde: Whether to use State Dependent Exploration or not\n :param log_std_init: Initial value for the log standard deviation\n :param full_std: Whether to use (n_features x n_actions) parameters\n for the std instead of only (n_features,) when using gSDE\n :param use_expln: Use ``expln()`` function instead of ``exp()`` to ensure\n a positive standard deviation (cf paper). It allows to keep variance\n above zero and prevent it from growing too fast. In practice, ``exp()`` is usually enough.\n :param squash_output: Whether to squash the output using a tanh function,\n this allows to ensure boundaries when using gSDE.\n :param features_extractor_class: Features extractor to use.\n :param features_extractor_kwargs: Keyword arguments\n to pass to the features extractor.\n :param share_features_extractor: If True, the features extractor is shared between the policy and value networks.\n :param normalize_images: Whether to normalize images or not,\n dividing by 255.0 (True by default)\n :param optimizer_class: The optimizer to use,\n ``th.optim.Adam`` by default\n :param optimizer_kwargs: Additional keyword arguments,\n excluding the learning rate, to pass to the optimizer\n ", "__init__": "<function ActorCriticPolicy.__init__ at 0x7fa27a4329e0>", "_get_constructor_parameters": "<function ActorCriticPolicy._get_constructor_parameters at 0x7fa27a432a70>", "reset_noise": "<function ActorCriticPolicy.reset_noise at 0x7fa27a432b00>", "_build_mlp_extractor": "<function ActorCriticPolicy._build_mlp_extractor at 0x7fa27a432b90>", "_build": "<function ActorCriticPolicy._build at 0x7fa27a432c20>", "forward": "<function ActorCriticPolicy.forward at 0x7fa27a432cb0>", "extract_features": "<function ActorCriticPolicy.extract_features at 0x7fa27a432d40>", "_get_action_dist_from_latent": "<function ActorCriticPolicy._get_action_dist_from_latent at 0x7fa27a432dd0>", "_predict": "<function ActorCriticPolicy._predict at 0x7fa27a432e60>", "evaluate_actions": "<function ActorCriticPolicy.evaluate_actions at 0x7fa27a432ef0>", "get_distribution": "<function ActorCriticPolicy.get_distribution at 0x7fa27a432f80>", "predict_values": "<function ActorCriticPolicy.predict_values at 0x7fa27a433010>", "__abstractmethods__": "frozenset()", "_abc_impl": "<_abc._abc_data object at 0x7fa27a3de140>"}, "verbose": 1, "policy_kwargs": {}, "num_timesteps": 1015808, "_total_timesteps": 1000000, "_num_timesteps_at_start": 0, "seed": null, "action_noise": null, "start_time": 1706627912711162291, "learning_rate": 0.0003, "tensorboard_log": null, "_last_obs": {":type:": "<class 'numpy.ndarray'>", ":serialized:": "gAWVdQIAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJYAAgAAAAAAAABUUry7uqY9qTuJvYZkVb6fbzO9O2d9vAAAAAAAAAAAzV5YvYWzuLmJPzu6dZKctcoDgrrY9lk5AACAPwAAgD+z4ce97BHEOEUlizt5OZo42sMauzfoLLoAAIA/AACAP024Lb1II6O61UBOujPqDbWIzUS4+rZqOQAAgD8AAIA/piqhvVz/YLpy30a44hK2Muia6bozpGM3AACAPwAAgD9Nrg09KXgOuj4IXruawAU3sdKeOqiYVDoAAIA/AACAP5phrrspQCK6jY+QO7AxNjh8uwK6uwXmuAAAgD8AAIA/TeVhvSn0fLoyw0w75vfjNR1uGLtl22u6AACAPwAAgD8z1I+8ZK7ZPY4VDL59lSu+YAmnvJryfbwAAAAAAAAAAAAUT76wSEE/cMf1vKKalr6ttli9UpVPPQAAAAAAAAAAzczhvUgFjrrW4II5icV+NDdcDrsM3Ze4AACAPwAAgD8AVMS8XMs3uihkF7iaFBEyTP6iuhQ5LzcAAIA/AACAP4D+Zb0UBKO6akl8u1UTwLaQSnC60CuROgAAgD8AAIA/M3lXvXt5Ij+2mFY8IH+TvqG8Dr1gAN28AAAAAAAAAAAAfbA8j/JouqsMzbt0k7y1g06Fu2gcLDUAAIA/AACAP1orhL32VGa6d6WIO2ZfDDdpFU47x4WcugAAgD8AAIA/lIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUiYiHlFKUKEsDjAE8lE5OTkr/////Sv////9LAHSUYksQSwiGlIwBQ5R0lFKULg=="}, "_last_episode_starts": {":type:": "<class 'numpy.ndarray'>", ":serialized:": "gAWVgwAAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJYQAAAAAAAAAAAAAAAAAAAAAAAAAAAAAACUjAVudW1weZSMBWR0eXBllJOUjAJiMZSJiIeUUpQoSwOMAXyUTk5OSv////9K/////0sAdJRiSxCFlIwBQ5R0lFKULg=="}, "_last_original_obs": null, "_episode_num": 0, "use_sde": false, "sde_sample_freq": -1, "_current_progress_remaining": -0.015808000000000044, "_stats_window_size": 100, "ep_info_buffer": {":type:": "<class 'collections.deque'>", ":serialized:": "gAWVLgwAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKUKH2UKIwBcpRHQE1PbmEGqxWMAWyUTS4BjAF0lEdAlUjv0ulGgHV9lChoBkdAZJaYR/ViF2gHTegDaAhHQJVKfDsMRYl1fZQoaAZHQGNb/TTfBN5oB03oA2gIR0CVXJbRnezldX2UKGgGR0BnQUQK8cuKaAdN6ANoCEdAlV7YZVGTcXV9lChoBkdAXCD+YMOPNmgHTegDaAhHQJVii+cpb2V1fZQoaAZHQGapJC0F8ohoB03oA2gIR0CVa1vXbuc+dX2UKGgGR0BlHZc1O0swaAdN6ANoCEdAlWuuyquKXXV9lChoBkdAZBmQiiZfD2gHTegDaAhHQJVxKepXIU91fZQoaAZHQGBT0wztTk1oB03oA2gIR0CVcSzp5eJIdX2UKGgGR0ApC2vStvGZaAdL7WgIR0CVef7PIGQkdX2UKGgGR0BgcUdNnGsFaAdN6ANoCEdAlX6DG1hLG3V9lChoBkdAYL9AqNIbwWgHTegDaAhHQJV/GDK5kLB1fZQoaAZHQGGGg3974SJoB03oA2gIR0CVhqWxQizLdX2UKGgGR0BczYEOiFj/aAdN6ANoCEdAlY3Bcu8K5XV9lChoBkdAYiFluFYdQ2gHTegDaAhHQJWVtbA1vVF1fZQoaAZHQGDc3YL9deJoB03oA2gIR0CVl3xXGOuJdX2UKGgGR0BgC7JCBwuNaAdN6ANoCEdAlZjdXDFZPnV9lChoBkdAZctqs2eg+WgHTegDaAhHQJWZFzRx95R1fZQoaAZHQGSdDU3GXHBoB03oA2gIR0CVmpOI68xsdX2UKGgGR0Bi0cuQIUrTaAdN6ANoCEdAlZw1jqfOEHV9lChoBkdAYVSX0oScsmgHTegDaAhHQJWwWy9mHxl1fZQoaAZHQGRX/I8yN4toB03oA2gIR0CVss0UGmk4dX2UKGgGR0Bj3coH9m6HaAdN6ANoCEdAlboI5ggHNXV9lChoBkdAQX1Nvfj0c2gHS8ZoCEdAlb2PGQ0XQHV9lChoBkdAX1QmMOwxFmgHTegDaAhHQJW/vaoMrmR1fZQoaAZHQGDNH752yLRoB03oA2gIR0CVv8RMvh60dX2UKGgGR0BnIwJXyRSxaAdN6ANoCEdAlcmQxvegtnV9lChoBkdAY6F0btJFs2gHTegDaAhHQJXNf0Zm7J51fZQoaAZHQGHpA31jAi5oB03oA2gIR0CVzgNipeeGdX2UKGgGR0Bk5dCeEqUeaAdN6ANoCEdAldThJVbRnnV9lChoBkdAZZS+K0lZ5mgHTegDaAhHQJXcLCIk7fZ1fZQoaAZHQG25g6ltTDRoB01aAWgIR0CV3i0qYqoZdX2UKGgGR0BE0aUzKs+3aAdL2mgIR0CV4lFxn3+NdX2UKGgGR0BkzLRWtEG8aAdN6ANoCEdAleKHSa3I/HV9lChoBkdAY+eh+OOsDGgHTegDaAhHQJXj+BnSOR11fZQoaAZHQGHRBWxQizNoB03oA2gIR0CV5WWwNb1RdX2UKGgGR0BegUMLF4s3aAdN6ANoCEdAleWik9ECvHV9lChoBkdAYO+UYbbUPWgHTegDaAhHQJXnJHiFTNt1fZQoaAZHQGFqZKWcBltoB03oA2gIR0CV6JyIYWLxdX2UKGgGR0BjAHgP3BYWaAdN6ANoCEdAlfzdKVY6n3V9lChoBkdAUEYSZjQRgGgHS8toCEdAlgBjDbah6HV9lChoBkdALb8Ti83+/GgHS8RoCEdAlgS30kGA1HV9lChoBkdAYe/mOEM9bGgHTegDaAhHQJYHCXokiUx1fZQoaAZHQFq9zhxYJVtoB03oA2gIR0CWCnid8RcvdX2UKGgGR0BiOabpeNT+aAdN6ANoCEdAlgwZ0jkdWHV9lChoBkdAYeoug6EJ0GgHTegDaAhHQJYMHNTtLL91fZQoaAZHQD6qKTB68g9oB0vkaAhHQJYM5sLv1Dl1fZQoaAZHQGQczJQtSQ5oB03oA2gIR0CWFyPw/gR9dX2UKGgGR0BfGz9OymhuaAdN6ANoCEdAlheb9VFQVXV9lChoBkdAQOMLjPv8ZWgHS9hoCEdAliVY/3WWhXV9lChoBkdAZIw5eZ5Rj2gHTegDaAhHQJYnIjNY8uB1fZQoaAZHQGDVcynDR+loB03oA2gIR0CWKPVhCtzTdX2UKGgGR0BgqSGxlg+haAdN6ANoCEdAliyOhsZYP3V9lChoBkdAY7vAGB4D92gHTegDaAhHQJYsv+sHSnd1fZQoaAZHQGfoB99c8kloB03oA2gIR0CWLzDA8B+4dX2UKGgGR0Bn/LayrxRVaAdN6ANoCEdAli9hdyDIzXV9lChoBkdAYwIBNEgGKWgHTegDaAhHQJYwpsUIsy11fZQoaAZHQDVJYHPeHi5oB0vGaAhHQJYxMSdvsJJ1fZQoaAZHQEn3YQrc0tRoB0vSaAhHQJYzFY5ksjF1fZQoaAZHQGQ3NZFG5MFoB03oA2gIR0CWM+vb48EFdX2UKGgGR0A2qUb1h9b5aAdL1WgIR0CWSL5gPVd5dX2UKGgGR0BHfRd6cAinaAdL+GgIR0CWSpqfe1rqdX2UKGgGR0Bk4SzeGfwraAdN6ANoCEdAlksLiQ1aXHV9lChoBkdAZ6BAbhm5D2gHTegDaAhHQJZNNZjhDPZ1fZQoaAZHQDDcpWmxdIJoB0vFaAhHQJZNkcJdB0J1fZQoaAZHQFfyamXPZ7JoB03oA2gIR0CWUPsQNCqqdX2UKGgGR0Biksny/bj+aAdN6ANoCEdAllMBwIdELHV9lChoBkdAYp/qrzXjEWgHTegDaAhHQJZTBenhsIp1fZQoaAZHQGdZTk6tDD1oB03oA2gIR0CWU++so2GZdX2UKGgGR8A65m/WUbDNaAdL1WgIR0CWWMBtUGVzdX2UKGgGR0BA7j/dZaFFaAdL12gIR0CWXXFkxyn2dX2UKGgGR0BlV3O6d1+zaAdN6ANoCEdAll216/qPfnV9lChoBkdANsXJxNqQBGgHS7poCEdAlmjj3/Pw/nV9lChoBkdAYy6exwAEMmgHTegDaAhHQJZqQUL2HtZ1fZQoaAZHQGI1Zf+jua5oB03oA2gIR0CWcgqfvnbJdX2UKGgGR0BkkZr30wrUaAdN6ANoCEdAlnI8i0OVgXV9lChoBkfAPxhdD6WPcWgHS8poCEdAlnULsv7FbXV9lChoBkdAYA8BczImxGgHTegDaAhHQJZ3R+DvmYB1fZQoaAZHQGcmd87ZFodoB03oA2gIR0CWd/mkFfRedX2UKGgGR0Ae2LbYbsF/aAdL/2gIR0CWeZTYukDZdX2UKGgGR0BhP+HxjJ+2aAdN6ANoCEdAlnp0+X7cf3V9lChoBkdAYI5p3X7LuGgHTegDaAhHQJaTOofjjrB1fZQoaAZHQGQoaZx7zCloB03oA2gIR0CWlTB+F10UdX2UKGgGR0BgHtD8cdYGaAdN6ANoCEdAlpWkMb3oLXV9lChoBkdATsXtrsSkCWgHTQ8BaAhHQJaWCyNXHR11fZQoaAZHQD7ZijL0SRNoB0vtaAhHQJaWkGeMAFR1fZQoaAZHQF9Z0CRwIdFoB03oA2gIR0CWl217Y02tdX2UKGgGR0BLeF5nlGPQaAdL22gIR0CWl7tYjjaPdX2UKGgGR0Bf6/5k9U0faAdN6ANoCEdAlpnvGVAzHnV9lChoBkdAYAk/KQq7RWgHTegDaAhHQJabPbcoH9p1fZQoaAZHQGFeohY/3WZoB03oA2gIR0CWm+ha1TisdX2UKGgGR0BDcD2SMcZMaAdL3mgIR0CWnP/T9bX6dX2UKGgGR0A8ik9lmOENaAdL6WgIR0CWn4YeT3ZgdX2UKGgGR0BitpR4yGi6aAdN6ANoCEdAlqCzmCAc1nV9lChoBkdAYs+OWBz3iGgHTegDaAhHQJalAz41xbV1fZQoaAZHQEg5PRiPQv9oB0vraAhHQJanCu+yquN1fZQoaAZHQGWQ4hllK9RoB03oA2gIR0CWvbanJkoXdX2UKGgGR0BiRQFgUlAvaAdN6ANoCEdAlr30fT1CgXV9lChoBkdAZuca2F36h2gHTegDaAhHQJbDac/dIoV1fZQoaAZHQGOTgdGRV6xoB03oA2gIR0CWxhdweeWfdWUu"}, "ep_success_buffer": {":type:": "<class 'collections.deque'>", ":serialized:": "gAWVIAAAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKULg=="}, "_n_updates": 248, "observation_space": {":type:": "<class 'gymnasium.spaces.box.Box'>", ":serialized:": "gAWVGAMAAAAAAACMFGd5bW5hc2l1bS5zcGFjZXMuYm94lIwDQm94lJOUKYGUfZQojAVkdHlwZZSMBW51bXB5lGgFk5SMAmY0lImIh5RSlChLA4wBPJROTk5K/////0r/////SwB0lGKMDWJvdW5kZWRfYmVsb3eUjBJudW1weS5jb3JlLm51bWVyaWOUjAtfZnJvbWJ1ZmZlcpSTlCiWCAAAAAAAAAABAQEBAQEBAZRoB4wCYjGUiYiHlFKUKEsDjAF8lE5OTkr/////Sv////9LAHSUYksIhZSMAUOUdJRSlIwNYm91bmRlZF9hYm92ZZRoECiWCAAAAAAAAAABAQEBAQEBAZRoFEsIhZRoGHSUUpSMBl9zaGFwZZRLCIWUjANsb3eUaBAoliAAAAAAAAAAAAC0wgAAtMIAAKDAAACgwNsPScAAAKDAAAAAgAAAAICUaApLCIWUaBh0lFKUjARoaWdolGgQKJYgAAAAAAAAAAAAtEIAALRCAACgQAAAoEDbD0lAAACgQAAAgD8AAIA/lGgKSwiFlGgYdJRSlIwIbG93X3JlcHKUjFtbLTkwLiAgICAgICAgLTkwLiAgICAgICAgIC01LiAgICAgICAgIC01LiAgICAgICAgIC0zLjE0MTU5MjcgIC01LgogIC0wLiAgICAgICAgIC0wLiAgICAgICBdlIwJaGlnaF9yZXBylIxTWzkwLiAgICAgICAgOTAuICAgICAgICAgNS4gICAgICAgICA1LiAgICAgICAgIDMuMTQxNTkyNyAgNS4KICAxLiAgICAgICAgIDEuICAgICAgIF2UjApfbnBfcmFuZG9tlIwUbnVtcHkucmFuZG9tLl9waWNrbGWUjBBfX2dlbmVyYXRvcl9jdG9ylJOUjAVQQ0c2NJSFlFKUfZQojA1iaXRfZ2VuZXJhdG9ylIwFUENHNjSUjAVzdGF0ZZR9lChoOooQqQd6iPtn5FTSoo7NriKtEYwDaW5jlIoQHUKmRMujt2jbUTpq74WcE3WMCmhhc191aW50MzKUSwCMCHVpbnRlZ2VylEsAdWJ1Yi4=", "dtype": "float32", "bounded_below": "[ True True True True True True True True]", "bounded_above": "[ True True True True True True True True]", "_shape": [8], "low": "[-90. -90. -5. -5. -3.1415927 -5.\n -0. -0. ]", "high": "[90. 90. 5. 5. 3.1415927 5.\n 1. 1. ]", "low_repr": "[-90. -90. -5. -5. -3.1415927 -5.\n -0. -0. ]", "high_repr": "[90. 90. 5. 5. 3.1415927 5.\n 1. 1. ]", "_np_random": "Generator(PCG64)"}, "action_space": {":type:": "<class 'gymnasium.spaces.discrete.Discrete'>", ":serialized:": "gAWV1QAAAAAAAACMGWd5bW5hc2l1bS5zcGFjZXMuZGlzY3JldGWUjAhEaXNjcmV0ZZSTlCmBlH2UKIwBbpSMFW51bXB5LmNvcmUubXVsdGlhcnJheZSMBnNjYWxhcpSTlIwFbnVtcHmUjAVkdHlwZZSTlIwCaTiUiYiHlFKUKEsDjAE8lE5OTkr/////Sv////9LAHSUYkMIBAAAAAAAAACUhpRSlIwFc3RhcnSUaAhoDkMIAAAAAAAAAACUhpRSlIwGX3NoYXBllCloCmgOjApfbnBfcmFuZG9tlE51Yi4=", "n": "4", "start": "0", "_shape": [], "dtype": "int64", "_np_random": null}, "n_envs": 16, "n_steps": 1024, "gamma": 0.999, "gae_lambda": 0.98, "ent_coef": 0.01, "vf_coef": 0.5, "max_grad_norm": 0.5, "batch_size": 64, "n_epochs": 4, "clip_range": {":type:": "<class 'function'>", ":serialized:": "gAWVxQIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwOX21ha2VfZnVuY3Rpb26Uk5QoaACMDV9idWlsdGluX3R5cGWUk5SMCENvZGVUeXBllIWUUpQoSwFLAEsASwFLAUsTQwSIAFMAlE6FlCmMAV+UhZSMSS91c3IvbG9jYWwvbGliL3B5dGhvbjMuMTAvZGlzdC1wYWNrYWdlcy9zdGFibGVfYmFzZWxpbmVzMy9jb21tb24vdXRpbHMucHmUjARmdW5jlEuEQwIEAZSMA3ZhbJSFlCl0lFKUfZQojAtfX3BhY2thZ2VfX5SMGHN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbpSMCF9fbmFtZV9flIwec3RhYmxlX2Jhc2VsaW5lczMuY29tbW9uLnV0aWxzlIwIX19maWxlX1+UjEkvdXNyL2xvY2FsL2xpYi9weXRob24zLjEwL2Rpc3QtcGFja2FnZXMvc3RhYmxlX2Jhc2VsaW5lczMvY29tbW9uL3V0aWxzLnB5lHVOTmgAjBBfbWFrZV9lbXB0eV9jZWxslJOUKVKUhZR0lFKUjBxjbG91ZHBpY2tsZS5jbG91ZHBpY2tsZV9mYXN0lIwSX2Z1bmN0aW9uX3NldHN0YXRllJOUaB99lH2UKGgWaA2MDF9fcXVhbG5hbWVfX5SMGWNvbnN0YW50X2ZuLjxsb2NhbHM+LmZ1bmOUjA9fX2Fubm90YXRpb25zX1+UfZSMDl9fa3dkZWZhdWx0c19flE6MDF9fZGVmYXVsdHNfX5ROjApfX21vZHVsZV9flGgXjAdfX2RvY19flE6MC19fY2xvc3VyZV9flGgAjApfbWFrZV9jZWxslJOURz/JmZmZmZmahZRSlIWUjBdfY2xvdWRwaWNrbGVfc3VibW9kdWxlc5RdlIwLX19nbG9iYWxzX1+UfZR1hpSGUjAu"}, "clip_range_vf": null, "normalize_advantage": true, "target_kl": null, "lr_schedule": {":type:": "<class 'function'>", ":serialized:": "gAWVxQIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwOX21ha2VfZnVuY3Rpb26Uk5QoaACMDV9idWlsdGluX3R5cGWUk5SMCENvZGVUeXBllIWUUpQoSwFLAEsASwFLAUsTQwSIAFMAlE6FlCmMAV+UhZSMSS91c3IvbG9jYWwvbGliL3B5dGhvbjMuMTAvZGlzdC1wYWNrYWdlcy9zdGFibGVfYmFzZWxpbmVzMy9jb21tb24vdXRpbHMucHmUjARmdW5jlEuEQwIEAZSMA3ZhbJSFlCl0lFKUfZQojAtfX3BhY2thZ2VfX5SMGHN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbpSMCF9fbmFtZV9flIwec3RhYmxlX2Jhc2VsaW5lczMuY29tbW9uLnV0aWxzlIwIX19maWxlX1+UjEkvdXNyL2xvY2FsL2xpYi9weXRob24zLjEwL2Rpc3QtcGFja2FnZXMvc3RhYmxlX2Jhc2VsaW5lczMvY29tbW9uL3V0aWxzLnB5lHVOTmgAjBBfbWFrZV9lbXB0eV9jZWxslJOUKVKUhZR0lFKUjBxjbG91ZHBpY2tsZS5jbG91ZHBpY2tsZV9mYXN0lIwSX2Z1bmN0aW9uX3NldHN0YXRllJOUaB99lH2UKGgWaA2MDF9fcXVhbG5hbWVfX5SMGWNvbnN0YW50X2ZuLjxsb2NhbHM+LmZ1bmOUjA9fX2Fubm90YXRpb25zX1+UfZSMDl9fa3dkZWZhdWx0c19flE6MDF9fZGVmYXVsdHNfX5ROjApfX21vZHVsZV9flGgXjAdfX2RvY19flE6MC19fY2xvc3VyZV9flGgAjApfbWFrZV9jZWxslJOURz8zqSowVTJhhZRSlIWUjBdfY2xvdWRwaWNrbGVfc3VibW9kdWxlc5RdlIwLX19nbG9iYWxzX1+UfZR1hpSGUjAu"}, "system_info": {"OS": "Linux-6.1.58+-x86_64-with-glibc2.35 # 1 SMP PREEMPT_DYNAMIC Sat Nov 18 15:31:17 UTC 2023", "Python": "3.10.12", "Stable-Baselines3": "2.0.0a5", "PyTorch": "2.1.0+cu121", "GPU Enabled": "True", "Numpy": "1.23.5", "Cloudpickle": "2.2.1", "Gymnasium": "0.28.1", "OpenAI Gym": "0.25.2"}}
|
|
|
|
| 1 |
+
{"policy_class": {":type:": "<class 'abc.ABCMeta'>", ":serialized:": "gAWVOwAAAAAAAACMIXN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi5wb2xpY2llc5SMEUFjdG9yQ3JpdGljUG9saWN5lJOULg==", "__module__": "stable_baselines3.common.policies", "__doc__": "\n Policy class for actor-critic algorithms (has both policy and value prediction).\n Used by A2C, PPO and the likes.\n\n :param observation_space: Observation space\n :param action_space: Action space\n :param lr_schedule: Learning rate schedule (could be constant)\n :param net_arch: The specification of the policy and value networks.\n :param activation_fn: Activation function\n :param ortho_init: Whether to use or not orthogonal initialization\n :param use_sde: Whether to use State Dependent Exploration or not\n :param log_std_init: Initial value for the log standard deviation\n :param full_std: Whether to use (n_features x n_actions) parameters\n for the std instead of only (n_features,) when using gSDE\n :param use_expln: Use ``expln()`` function instead of ``exp()`` to ensure\n a positive standard deviation (cf paper). It allows to keep variance\n above zero and prevent it from growing too fast. In practice, ``exp()`` is usually enough.\n :param squash_output: Whether to squash the output using a tanh function,\n this allows to ensure boundaries when using gSDE.\n :param features_extractor_class: Features extractor to use.\n :param features_extractor_kwargs: Keyword arguments\n to pass to the features extractor.\n :param share_features_extractor: If True, the features extractor is shared between the policy and value networks.\n :param normalize_images: Whether to normalize images or not,\n dividing by 255.0 (True by default)\n :param optimizer_class: The optimizer to use,\n ``th.optim.Adam`` by default\n :param optimizer_kwargs: Additional keyword arguments,\n excluding the learning rate, to pass to the optimizer\n ", "__init__": "<function ActorCriticPolicy.__init__ at 0x7dd661f5dbd0>", "_get_constructor_parameters": "<function ActorCriticPolicy._get_constructor_parameters at 0x7dd661f5dc60>", "reset_noise": "<function ActorCriticPolicy.reset_noise at 0x7dd661f5dcf0>", "_build_mlp_extractor": "<function ActorCriticPolicy._build_mlp_extractor at 0x7dd661f5dd80>", "_build": "<function ActorCriticPolicy._build at 0x7dd661f5de10>", "forward": "<function ActorCriticPolicy.forward at 0x7dd661f5dea0>", "extract_features": "<function ActorCriticPolicy.extract_features at 0x7dd661f5df30>", "_get_action_dist_from_latent": "<function ActorCriticPolicy._get_action_dist_from_latent at 0x7dd661f5dfc0>", "_predict": "<function ActorCriticPolicy._predict at 0x7dd661f5e050>", "evaluate_actions": "<function ActorCriticPolicy.evaluate_actions at 0x7dd661f5e0e0>", "get_distribution": "<function ActorCriticPolicy.get_distribution at 0x7dd661f5e170>", "predict_values": "<function ActorCriticPolicy.predict_values at 0x7dd661f5e200>", "__abstractmethods__": "frozenset()", "_abc_impl": "<_abc._abc_data object at 0x7dd661efe480>"}, "verbose": 1, "policy_kwargs": {}, "num_timesteps": 2015232, "_total_timesteps": 2000000, "_num_timesteps_at_start": 0, "seed": null, "action_noise": null, "start_time": 1706812349951161726, "learning_rate": 0.0003, "tensorboard_log": null, "_last_obs": {":type:": "<class 'numpy.ndarray'>", ":serialized:": "gAWVdQIAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJYAAgAAAAAAAGCxZj6JT2A/OIVmOerQ0L6Vyo0+cQoSvgAAAAAAAAAAAEKrPBTwm7r2Si87NvVzPPQXGrojGFY9AACAPwAAgD/NGLa89uxbuuMwVLgKy7Szo+yVOpNndjcAAIA/AACAPxoArr1x30c+WmXvPO9Opb4zfzO97DWmvAAAAAAAAAAA5jvWPcscbj+SeA09T/n2vn3PcD4SCqW9AAAAAAAAAACmaEQ+jojAvOa8ALtCrHQ5rNQpvlYuKjoAAIA/AACAP00PHL1ePrc/LQZNvoRRIr5kB4k8mX6kvQAAAAAAAAAAZuZOuk/Csz8TvqO93hyVvqKBcTpRXJQ8AAAAAAAAAADmmzy97te/P8txfL4x0Ru8NhjPPLhMdrwAAAAAAAAAAOb5ob11Zqw/MBSivnLA1b7B4sa9ZTQ0vgAAAAAAAAAAAMiovT1tFLvKxgI8pkaSPN/MFzwGlHy9AACAPwAAAACAmb295kuVP2YKsb6OMBK/Gt1EvnDHSL4AAAAAAAAAAIDBUD123Ru8s/b4O5+eUTxBZJI96iIxvQAAgD8AAIA/s6+wvcVwKj6G6js+FLWNvkEuUD0a6w49AAAAAAAAAAAzU8M62B23P7DLTT2TyYc+dyHeuiZuOLwAAAAAAAAAACCLdb7f0EE/6XQovYOa576LTJC+5vcbPgAAAAAAAAAAlIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUiYiHlFKUKEsDjAE8lE5OTkr/////Sv////9LAHSUYksQSwiGlIwBQ5R0lFKULg=="}, "_last_episode_starts": {":type:": "<class 'numpy.ndarray'>", ":serialized:": "gAWVgwAAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJYQAAAAAAAAAAAAAAAAAAABAAAAAAAAAACUjAVudW1weZSMBWR0eXBllJOUjAJiMZSJiIeUUpQoSwOMAXyUTk5OSv////9K/////0sAdJRiSxCFlIwBQ5R0lFKULg=="}, "_last_original_obs": null, "_episode_num": 0, "use_sde": false, "sde_sample_freq": -1, "_current_progress_remaining": -0.007616000000000067, "_stats_window_size": 100, "ep_info_buffer": {":type:": "<class 'collections.deque'>", ":serialized:": "gAWV6wsAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKUKH2UKIwBcpRHQHJ+fReC04SMAWyUS8yMAXSUR0CdfzQeFL39dX2UKGgGR0BxQrvphWo4aAdL9WgIR0Cdf0hZQpF1dX2UKGgGR0Bw7NhZyMkyaAdLzGgIR0Cdf2g88s+WdX2UKGgGR0Bysu1MM7U5aAdL62gIR0Cdf3dUsFt9dX2UKGgGR0BwGjT6SDAaaAdL5mgIR0Cdf4HJcPe6dX2UKGgGR0BxtpX+2mYTaAdL7WgIR0Cdf6eFtbcHdX2UKGgGR0ByZHk7wKBvaAdLy2gIR0CdgFqFyq+8dX2UKGgGR0BwuRG8VYZEaAdL8GgIR0CdgIkIomXxdX2UKGgGR0Bza46bONYKaAdL+WgIR0CdgQgGbCrMdX2UKGgGR0Bvg5oEjgQ6aAdLy2gIR0Cdgdm5lOGkdX2UKGgGR0ByHDnjhky2aAdL1WgIR0CdgecQRPGidX2UKGgGR0Bza63VkMCtaAdL6mgIR0CdgmAoXsPbdX2UKGgGR0BwBXbsWweOaAdL52gIR0CdgoiYsunNdX2UKGgGR0BxinWoWHk+aAdL62gIR0CdhHHBUJfIdX2UKGgGR0ByWQ1vVEuyaAdLyWgIR0CdhVmOlwcYdX2UKGgGR0B0Qu7QLNOeaAdL9mgIR0CdhZifxtpFdX2UKGgGR0BxEn0Gu9vkaAdL1WgIR0CdhjvStvGZdX2UKGgGR0BusSZ2IO6NaAdL22gIR0CdhmSqU/wBdX2UKGgGR0Bwe5dfLLZBaAdL6WgIR0CdhqaXrt3OdX2UKGgGR0BwngUZeiSJaAdL5WgIR0CdhrJQtSQ6dX2UKGgGR0BwWNkiD/VBaAdL6WgIR0CdhzbCaZx8dX2UKGgGR0BvLPyLAHmjaAdL02gIR0Cdh4ohY/3WdX2UKGgGR0Bz0tdE9dNWaAdL0mgIR0Cdh6sUqQRxdX2UKGgGR0BxxZsHjZL7aAdL3GgIR0CdiFAxzq8ldX2UKGgGR0BzYuSgXdj5aAdLzGgIR0CdiI0Q9RrKdX2UKGgGR0Bwjo76pHZsaAdL1mgIR0CdiRJnxri3dX2UKGgGR0BwzZWKdhAoaAdL62gIR0CdiTJFspG4dX2UKGgGR0Byb4150KZ2aAdL22gIR0CdiU+MZP2xdX2UKGgGR0BSadzKcNH6aAdLpmgIR0Cdm7mOEM9bdX2UKGgGR0BzKi1mapgkaAdL8mgIR0CdnBnMt9QXdX2UKGgGR0Bxp4o0ALiNaAdL02gIR0CdnCXokiUxdX2UKGgGR0Bx/8hUzbeuaAdLwmgIR0CdnGr8R+SbdX2UKGgGR0Bxt70PH1e0aAdL52gIR0CdnTccENe/dX2UKGgGR0BxcA7hegL7aAdL8WgIR0CdnV6be/HpdX2UKGgGR0BuIog/1QIlaAdNEAFoCEdAnZ2MmBvrGHV9lChoBkdAcLAXP7el9GgHS9hoCEdAnZ2md7OVxHV9lChoBkdAclkj7yhBaGgHS+VoCEdAnZ28OkLx7XV9lChoBkdAUqF9b5dnkGgHS59oCEdAnZ3gY51eSnV9lChoBkdAcIjyFPBSDWgHS+poCEdAnZ4ksSTQmnV9lChoBkdAYvuE8q4H5mgHTegDaAhHQJ2eKrilzlt1fZQoaAZHQG9/kuxrzoVoB0vcaAhHQJ2eUbm2b5N1fZQoaAZHQHKXp2pyZKFoB0viaAhHQJ2eqCrcTJ11fZQoaAZHQHISU3S8an9oB0vlaAhHQJ2fKKAJ9iN1fZQoaAZHQG5uo4EOiFloB00VAWgIR0CdoHxG2CumdX2UKGgGR0BuhDt/nW8RaAdL12gIR0CdoRUGmk30dX2UKGgGR0Bx8Frcj7hvaAdL72gIR0CdoaF+uvECdX2UKGgGR0BxItU4rBj4aAdL5WgIR0CdobL2HtWudX2UKGgGR0BwvBddE9dNaAdNAgFoCEdAnaHAqy4WlHV9lChoBkdAcTqWxyGSIWgHS81oCEdAnaJXjyWiUXV9lChoBkdAcWUFx4ptrWgHS9doCEdAnaKBcqvvB3V9lChoBkdAcEPlUp/gBWgHS+NoCEdAnaKHnlnyu3V9lChoBkdAcuFV1wHZ9WgHS99oCEdAnaKY3irDInV9lChoBkdAcD1RJEpiJGgHS9RoCEdAnaKrCBPKuHV9lChoBkdAcklCAMDwIGgHS8loCEdAnaLjdHlOoHV9lChoBkdAcT4Z4fOlf2gHS9doCEdAnaL/4EfT1HV9lChoBkdAUvh86V+qi2gHS6VoCEdAnaMQTVUdaXV9lChoBkdAcaEZ6lchT2gHTQcBaAhHQJ2jM1JlJ6J1fZQoaAZHQHH4+AAhje9oB0vaaAhHQJ2jiEEkjX51fZQoaAZHQHM3fHggow5oB00SAWgIR0CdpCCU5dWydX2UKGgGR0BttV9fCyhSaAdL3GgIR0CdpU25xzaLdX2UKGgGR0Bwn2msNlRQaAdL2WgIR0CdpcKjSG8FdX2UKGgGR0By6w6bONYKaAdL22gIR0CdpnAxBVuKdX2UKGgGR0Bx2TnJT2nLaAdL42gIR0CdprfDk2gndX2UKGgGR0BxXJ+NLlFMaAdLxWgIR0CdpsBxgiNbdX2UKGgGR0Btzy9sabWmaAdNAgFoCEdAnaddlyzXz3V9lChoBkdAb8fiTdLxqmgHS+BoCEdAnaeDnRsuWnV9lChoBkdAcg8fra/RFGgHS81oCEdAnafAiu+yq3V9lChoBkdAcK+o8ZDRdGgHS+loCEdAnafZP69CeHV9lChoBkdAbUCrkKeCkGgHTQEBaAhHQJ2oPq1PWQR1fZQoaAZHQHLfGQfZElVoB0vraAhHQJ2oShRIjGF1fZQoaAZHQHFaB86V+qloB00NAWgIR0CdqFwDeTFEdX2UKGgGR0Bttgprk8zRaAdL92gIR0CdqHmjj7yhdX2UKGgGR0ByQIMkQf6oaAdNCQFoCEdAnai3AM2FWXV9lChoBkdAcWiiHqNZNmgHS+poCEdAnajIT4+KTHV9lChoBkdAc3IXBxgiNmgHS+JoCEdAnak2jO9nLHV9lChoBkdAchY4ku6ErWgHS9poCEdAnao4mCyyEHV9lChoBkdAcQgdhy8zymgHS+BoCEdAnat9/FzdUXV9lChoBkdAb8Bp6hQFcWgHS/BoCEdAnaw3KOktVnV9lChoBkdAcuHIkJKJ22gHS+5oCEdAnazfmLcbi3V9lChoBkdAcFYacZtNz2gHS8poCEdAnazgGKQ7tHV9lChoBkdAcvafigkC3mgHS91oCEdAnaznck+otXV9lChoBkdAcNOxfv4M4WgHTTEBaAhHQJ2s+aLGaQV1fZQoaAZHQHH1yhSLqD9oB0vjaAhHQJ2s9zltCRh1fZQoaAZHQHEQ8qBmPHVoB0v0aAhHQJ2tJgCwKSh1fZQoaAZHQHJ1zp9qk/NoB0vpaAhHQJ2trJ7sv7F1fZQoaAZHQHOYnFxXGOxoB0v4aAhHQJ2t81/DtPZ1fZQoaAZHQHJskeEIw/RoB0viaAhHQJ2t872criF1fZQoaAZHQHJQBomG/N9oB0vwaAhHQJ2t+qPwNLF1fZQoaAZHQHDKbpqynk1oB0viaAhHQJ2uBRxcVxl1fZQoaAZHQHAPczZYgaFoB0vyaAhHQJ2uzQ0GeMB1fZQoaAZHQHH7oRRMvh9oB0viaAhHQJ2vioBJZnt1fZQoaAZHQHD51vuPV/doB0vsaAhHQJ2xxDc/MW51fZQoaAZHQFHjcnmaH9FoB0u5aAhHQJ2x3JaJQ+F1fZQoaAZHQHDkFKsdT5xoB0vLaAhHQJ2ygHD76551fZQoaAZHQHLNe2NNrTJoB0vVaAhHQJ2zONXHR1J1fZQoaAZHQHKNeY+jdpJoB0v6aAhHQJ2zRXYDklx1fZQoaAZHQHG4sEmplz5oB0vmaAhHQJ2za1Vo6CF1fZQoaAZHQHAmJ4B3iaRoB0vnaAhHQJ2zdSR8twt1fZQoaAZHQHFe2QXAM2FoB0vFaAhHQJ2z6UQkHD91fZQoaAZHQHFK0+otL+RoB0v2aAhHQJ2z/09QoCx1ZS4="}, "ep_success_buffer": {":type:": "<class 'collections.deque'>", ":serialized:": "gAWVIAAAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKULg=="}, "_n_updates": 492, "observation_space": {":type:": "<class 'gymnasium.spaces.box.Box'>", ":serialized:": "gAWVcAIAAAAAAACMFGd5bW5hc2l1bS5zcGFjZXMuYm94lIwDQm94lJOUKYGUfZQojAVkdHlwZZSMBW51bXB5lGgFk5SMAmY0lImIh5RSlChLA4wBPJROTk5K/////0r/////SwB0lGKMDWJvdW5kZWRfYmVsb3eUjBJudW1weS5jb3JlLm51bWVyaWOUjAtfZnJvbWJ1ZmZlcpSTlCiWCAAAAAAAAAABAQEBAQEBAZRoB4wCYjGUiYiHlFKUKEsDjAF8lE5OTkr/////Sv////9LAHSUYksIhZSMAUOUdJRSlIwNYm91bmRlZF9hYm92ZZRoECiWCAAAAAAAAAABAQEBAQEBAZRoFEsIhZRoGHSUUpSMBl9zaGFwZZRLCIWUjANsb3eUaBAoliAAAAAAAAAAAAC0wgAAtMIAAKDAAACgwNsPScAAAKDAAAAAgAAAAICUaApLCIWUaBh0lFKUjARoaWdolGgQKJYgAAAAAAAAAAAAtEIAALRCAACgQAAAoEDbD0lAAACgQAAAgD8AAIA/lGgKSwiFlGgYdJRSlIwIbG93X3JlcHKUjFtbLTkwLiAgICAgICAgLTkwLiAgICAgICAgIC01LiAgICAgICAgIC01LiAgICAgICAgIC0zLjE0MTU5MjcgIC01LgogIC0wLiAgICAgICAgIC0wLiAgICAgICBdlIwJaGlnaF9yZXBylIxTWzkwLiAgICAgICAgOTAuICAgICAgICAgNS4gICAgICAgICA1LiAgICAgICAgIDMuMTQxNTkyNyAgNS4KICAxLiAgICAgICAgIDEuICAgICAgIF2UjApfbnBfcmFuZG9tlE51Yi4=", "dtype": "float32", "bounded_below": "[ True True True True True True True True]", "bounded_above": "[ True True True True True True True True]", "_shape": [8], "low": "[-90. -90. -5. -5. -3.1415927 -5.\n -0. -0. ]", "high": "[90. 90. 5. 5. 3.1415927 5.\n 1. 1. ]", "low_repr": "[-90. -90. -5. -5. -3.1415927 -5.\n -0. -0. ]", "high_repr": "[90. 90. 5. 5. 3.1415927 5.\n 1. 1. ]", "_np_random": null}, "action_space": {":type:": "<class 'gymnasium.spaces.discrete.Discrete'>", ":serialized:": "gAWV1QAAAAAAAACMGWd5bW5hc2l1bS5zcGFjZXMuZGlzY3JldGWUjAhEaXNjcmV0ZZSTlCmBlH2UKIwBbpSMFW51bXB5LmNvcmUubXVsdGlhcnJheZSMBnNjYWxhcpSTlIwFbnVtcHmUjAVkdHlwZZSTlIwCaTiUiYiHlFKUKEsDjAE8lE5OTkr/////Sv////9LAHSUYkMIBAAAAAAAAACUhpRSlIwFc3RhcnSUaAhoDkMIAAAAAAAAAACUhpRSlIwGX3NoYXBllCloCmgOjApfbnBfcmFuZG9tlE51Yi4=", "n": "4", "start": "0", "_shape": [], "dtype": "int64", "_np_random": null}, "n_envs": 16, "n_steps": 1024, "gamma": 0.999, "gae_lambda": 0.98, "ent_coef": 0.01, "vf_coef": 0.5, "max_grad_norm": 0.5, "batch_size": 64, "n_epochs": 4, "clip_range": {":type:": "<class 'function'>", ":serialized:": "gAWVxQIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwOX21ha2VfZnVuY3Rpb26Uk5QoaACMDV9idWlsdGluX3R5cGWUk5SMCENvZGVUeXBllIWUUpQoSwFLAEsASwFLAUsTQwSIAFMAlE6FlCmMAV+UhZSMSS91c3IvbG9jYWwvbGliL3B5dGhvbjMuMTAvZGlzdC1wYWNrYWdlcy9zdGFibGVfYmFzZWxpbmVzMy9jb21tb24vdXRpbHMucHmUjARmdW5jlEuEQwIEAZSMA3ZhbJSFlCl0lFKUfZQojAtfX3BhY2thZ2VfX5SMGHN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbpSMCF9fbmFtZV9flIwec3RhYmxlX2Jhc2VsaW5lczMuY29tbW9uLnV0aWxzlIwIX19maWxlX1+UjEkvdXNyL2xvY2FsL2xpYi9weXRob24zLjEwL2Rpc3QtcGFja2FnZXMvc3RhYmxlX2Jhc2VsaW5lczMvY29tbW9uL3V0aWxzLnB5lHVOTmgAjBBfbWFrZV9lbXB0eV9jZWxslJOUKVKUhZR0lFKUjBxjbG91ZHBpY2tsZS5jbG91ZHBpY2tsZV9mYXN0lIwSX2Z1bmN0aW9uX3NldHN0YXRllJOUaB99lH2UKGgWaA2MDF9fcXVhbG5hbWVfX5SMGWNvbnN0YW50X2ZuLjxsb2NhbHM+LmZ1bmOUjA9fX2Fubm90YXRpb25zX1+UfZSMDl9fa3dkZWZhdWx0c19flE6MDF9fZGVmYXVsdHNfX5ROjApfX21vZHVsZV9flGgXjAdfX2RvY19flE6MC19fY2xvc3VyZV9flGgAjApfbWFrZV9jZWxslJOURz/JmZmZmZmahZRSlIWUjBdfY2xvdWRwaWNrbGVfc3VibW9kdWxlc5RdlIwLX19nbG9iYWxzX1+UfZR1hpSGUjAu"}, "clip_range_vf": null, "normalize_advantage": true, "target_kl": null, "lr_schedule": {":type:": "<class 'function'>", ":serialized:": "gAWVxQIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwOX21ha2VfZnVuY3Rpb26Uk5QoaACMDV9idWlsdGluX3R5cGWUk5SMCENvZGVUeXBllIWUUpQoSwFLAEsASwFLAUsTQwSIAFMAlE6FlCmMAV+UhZSMSS91c3IvbG9jYWwvbGliL3B5dGhvbjMuMTAvZGlzdC1wYWNrYWdlcy9zdGFibGVfYmFzZWxpbmVzMy9jb21tb24vdXRpbHMucHmUjARmdW5jlEuEQwIEAZSMA3ZhbJSFlCl0lFKUfZQojAtfX3BhY2thZ2VfX5SMGHN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbpSMCF9fbmFtZV9flIwec3RhYmxlX2Jhc2VsaW5lczMuY29tbW9uLnV0aWxzlIwIX19maWxlX1+UjEkvdXNyL2xvY2FsL2xpYi9weXRob24zLjEwL2Rpc3QtcGFja2FnZXMvc3RhYmxlX2Jhc2VsaW5lczMvY29tbW9uL3V0aWxzLnB5lHVOTmgAjBBfbWFrZV9lbXB0eV9jZWxslJOUKVKUhZR0lFKUjBxjbG91ZHBpY2tsZS5jbG91ZHBpY2tsZV9mYXN0lIwSX2Z1bmN0aW9uX3NldHN0YXRllJOUaB99lH2UKGgWaA2MDF9fcXVhbG5hbWVfX5SMGWNvbnN0YW50X2ZuLjxsb2NhbHM+LmZ1bmOUjA9fX2Fubm90YXRpb25zX1+UfZSMDl9fa3dkZWZhdWx0c19flE6MDF9fZGVmYXVsdHNfX5ROjApfX21vZHVsZV9flGgXjAdfX2RvY19flE6MC19fY2xvc3VyZV9flGgAjApfbWFrZV9jZWxslJOURz8zqSowVTJhhZRSlIWUjBdfY2xvdWRwaWNrbGVfc3VibW9kdWxlc5RdlIwLX19nbG9iYWxzX1+UfZR1hpSGUjAu"}, "system_info": {"OS": "Linux-6.1.58+-x86_64-with-glibc2.35 # 1 SMP PREEMPT_DYNAMIC Sat Nov 18 15:31:17 UTC 2023", "Python": "3.10.12", "Stable-Baselines3": "2.0.0a5", "PyTorch": "2.1.0+cu121", "GPU Enabled": "True", "Numpy": "1.23.5", "Cloudpickle": "2.2.1", "Gymnasium": "0.28.1", "OpenAI Gym": "0.25.2"}}
|
ppo-LunarLander-v2.zip
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d10a02eba5e5ff71522a689db5bbd6350627a12b84abc0fa720634931048dd40
|
| 3 |
+
size 147952
|
ppo-LunarLander-v2/data
CHANGED
|
@@ -4,57 +4,57 @@
|
|
| 4 |
":serialized:": "gAWVOwAAAAAAAACMIXN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi5wb2xpY2llc5SMEUFjdG9yQ3JpdGljUG9saWN5lJOULg==",
|
| 5 |
"__module__": "stable_baselines3.common.policies",
|
| 6 |
"__doc__": "\n Policy class for actor-critic algorithms (has both policy and value prediction).\n Used by A2C, PPO and the likes.\n\n :param observation_space: Observation space\n :param action_space: Action space\n :param lr_schedule: Learning rate schedule (could be constant)\n :param net_arch: The specification of the policy and value networks.\n :param activation_fn: Activation function\n :param ortho_init: Whether to use or not orthogonal initialization\n :param use_sde: Whether to use State Dependent Exploration or not\n :param log_std_init: Initial value for the log standard deviation\n :param full_std: Whether to use (n_features x n_actions) parameters\n for the std instead of only (n_features,) when using gSDE\n :param use_expln: Use ``expln()`` function instead of ``exp()`` to ensure\n a positive standard deviation (cf paper). It allows to keep variance\n above zero and prevent it from growing too fast. In practice, ``exp()`` is usually enough.\n :param squash_output: Whether to squash the output using a tanh function,\n this allows to ensure boundaries when using gSDE.\n :param features_extractor_class: Features extractor to use.\n :param features_extractor_kwargs: Keyword arguments\n to pass to the features extractor.\n :param share_features_extractor: If True, the features extractor is shared between the policy and value networks.\n :param normalize_images: Whether to normalize images or not,\n dividing by 255.0 (True by default)\n :param optimizer_class: The optimizer to use,\n ``th.optim.Adam`` by default\n :param optimizer_kwargs: Additional keyword arguments,\n excluding the learning rate, to pass to the optimizer\n ",
|
| 7 |
-
"__init__": "<function ActorCriticPolicy.__init__ at
|
| 8 |
-
"_get_constructor_parameters": "<function ActorCriticPolicy._get_constructor_parameters at
|
| 9 |
-
"reset_noise": "<function ActorCriticPolicy.reset_noise at
|
| 10 |
-
"_build_mlp_extractor": "<function ActorCriticPolicy._build_mlp_extractor at
|
| 11 |
-
"_build": "<function ActorCriticPolicy._build at
|
| 12 |
-
"forward": "<function ActorCriticPolicy.forward at
|
| 13 |
-
"extract_features": "<function ActorCriticPolicy.extract_features at
|
| 14 |
-
"_get_action_dist_from_latent": "<function ActorCriticPolicy._get_action_dist_from_latent at
|
| 15 |
-
"_predict": "<function ActorCriticPolicy._predict at
|
| 16 |
-
"evaluate_actions": "<function ActorCriticPolicy.evaluate_actions at
|
| 17 |
-
"get_distribution": "<function ActorCriticPolicy.get_distribution at
|
| 18 |
-
"predict_values": "<function ActorCriticPolicy.predict_values at
|
| 19 |
"__abstractmethods__": "frozenset()",
|
| 20 |
-
"_abc_impl": "<_abc._abc_data object at
|
| 21 |
},
|
| 22 |
"verbose": 1,
|
| 23 |
"policy_kwargs": {},
|
| 24 |
-
"num_timesteps":
|
| 25 |
-
"_total_timesteps":
|
| 26 |
"_num_timesteps_at_start": 0,
|
| 27 |
"seed": null,
|
| 28 |
"action_noise": null,
|
| 29 |
-
"start_time":
|
| 30 |
"learning_rate": 0.0003,
|
| 31 |
"tensorboard_log": null,
|
| 32 |
"_last_obs": {
|
| 33 |
":type:": "<class 'numpy.ndarray'>",
|
| 34 |
-
":serialized:": "
|
| 35 |
},
|
| 36 |
"_last_episode_starts": {
|
| 37 |
":type:": "<class 'numpy.ndarray'>",
|
| 38 |
-
":serialized:": "
|
| 39 |
},
|
| 40 |
"_last_original_obs": null,
|
| 41 |
"_episode_num": 0,
|
| 42 |
"use_sde": false,
|
| 43 |
"sde_sample_freq": -1,
|
| 44 |
-
"_current_progress_remaining": -0.
|
| 45 |
"_stats_window_size": 100,
|
| 46 |
"ep_info_buffer": {
|
| 47 |
":type:": "<class 'collections.deque'>",
|
| 48 |
-
":serialized:": "
|
| 49 |
},
|
| 50 |
"ep_success_buffer": {
|
| 51 |
":type:": "<class 'collections.deque'>",
|
| 52 |
":serialized:": "gAWVIAAAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKULg=="
|
| 53 |
},
|
| 54 |
-
"_n_updates":
|
| 55 |
"observation_space": {
|
| 56 |
":type:": "<class 'gymnasium.spaces.box.Box'>",
|
| 57 |
-
":serialized:": "
|
| 58 |
"dtype": "float32",
|
| 59 |
"bounded_below": "[ True True True True True True True True]",
|
| 60 |
"bounded_above": "[ True True True True True True True True]",
|
|
@@ -65,7 +65,7 @@
|
|
| 65 |
"high": "[90. 90. 5. 5. 3.1415927 5.\n 1. 1. ]",
|
| 66 |
"low_repr": "[-90. -90. -5. -5. -3.1415927 -5.\n -0. -0. ]",
|
| 67 |
"high_repr": "[90. 90. 5. 5. 3.1415927 5.\n 1. 1. ]",
|
| 68 |
-
"_np_random":
|
| 69 |
},
|
| 70 |
"action_space": {
|
| 71 |
":type:": "<class 'gymnasium.spaces.discrete.Discrete'>",
|
|
|
|
| 4 |
":serialized:": "gAWVOwAAAAAAAACMIXN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi5wb2xpY2llc5SMEUFjdG9yQ3JpdGljUG9saWN5lJOULg==",
|
| 5 |
"__module__": "stable_baselines3.common.policies",
|
| 6 |
"__doc__": "\n Policy class for actor-critic algorithms (has both policy and value prediction).\n Used by A2C, PPO and the likes.\n\n :param observation_space: Observation space\n :param action_space: Action space\n :param lr_schedule: Learning rate schedule (could be constant)\n :param net_arch: The specification of the policy and value networks.\n :param activation_fn: Activation function\n :param ortho_init: Whether to use or not orthogonal initialization\n :param use_sde: Whether to use State Dependent Exploration or not\n :param log_std_init: Initial value for the log standard deviation\n :param full_std: Whether to use (n_features x n_actions) parameters\n for the std instead of only (n_features,) when using gSDE\n :param use_expln: Use ``expln()`` function instead of ``exp()`` to ensure\n a positive standard deviation (cf paper). It allows to keep variance\n above zero and prevent it from growing too fast. In practice, ``exp()`` is usually enough.\n :param squash_output: Whether to squash the output using a tanh function,\n this allows to ensure boundaries when using gSDE.\n :param features_extractor_class: Features extractor to use.\n :param features_extractor_kwargs: Keyword arguments\n to pass to the features extractor.\n :param share_features_extractor: If True, the features extractor is shared between the policy and value networks.\n :param normalize_images: Whether to normalize images or not,\n dividing by 255.0 (True by default)\n :param optimizer_class: The optimizer to use,\n ``th.optim.Adam`` by default\n :param optimizer_kwargs: Additional keyword arguments,\n excluding the learning rate, to pass to the optimizer\n ",
|
| 7 |
+
"__init__": "<function ActorCriticPolicy.__init__ at 0x7dd661f5dbd0>",
|
| 8 |
+
"_get_constructor_parameters": "<function ActorCriticPolicy._get_constructor_parameters at 0x7dd661f5dc60>",
|
| 9 |
+
"reset_noise": "<function ActorCriticPolicy.reset_noise at 0x7dd661f5dcf0>",
|
| 10 |
+
"_build_mlp_extractor": "<function ActorCriticPolicy._build_mlp_extractor at 0x7dd661f5dd80>",
|
| 11 |
+
"_build": "<function ActorCriticPolicy._build at 0x7dd661f5de10>",
|
| 12 |
+
"forward": "<function ActorCriticPolicy.forward at 0x7dd661f5dea0>",
|
| 13 |
+
"extract_features": "<function ActorCriticPolicy.extract_features at 0x7dd661f5df30>",
|
| 14 |
+
"_get_action_dist_from_latent": "<function ActorCriticPolicy._get_action_dist_from_latent at 0x7dd661f5dfc0>",
|
| 15 |
+
"_predict": "<function ActorCriticPolicy._predict at 0x7dd661f5e050>",
|
| 16 |
+
"evaluate_actions": "<function ActorCriticPolicy.evaluate_actions at 0x7dd661f5e0e0>",
|
| 17 |
+
"get_distribution": "<function ActorCriticPolicy.get_distribution at 0x7dd661f5e170>",
|
| 18 |
+
"predict_values": "<function ActorCriticPolicy.predict_values at 0x7dd661f5e200>",
|
| 19 |
"__abstractmethods__": "frozenset()",
|
| 20 |
+
"_abc_impl": "<_abc._abc_data object at 0x7dd661efe480>"
|
| 21 |
},
|
| 22 |
"verbose": 1,
|
| 23 |
"policy_kwargs": {},
|
| 24 |
+
"num_timesteps": 2015232,
|
| 25 |
+
"_total_timesteps": 2000000,
|
| 26 |
"_num_timesteps_at_start": 0,
|
| 27 |
"seed": null,
|
| 28 |
"action_noise": null,
|
| 29 |
+
"start_time": 1706812349951161726,
|
| 30 |
"learning_rate": 0.0003,
|
| 31 |
"tensorboard_log": null,
|
| 32 |
"_last_obs": {
|
| 33 |
":type:": "<class 'numpy.ndarray'>",
|
| 34 |
+
":serialized:": "gAWVdQIAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJYAAgAAAAAAAGCxZj6JT2A/OIVmOerQ0L6Vyo0+cQoSvgAAAAAAAAAAAEKrPBTwm7r2Si87NvVzPPQXGrojGFY9AACAPwAAgD/NGLa89uxbuuMwVLgKy7Szo+yVOpNndjcAAIA/AACAPxoArr1x30c+WmXvPO9Opb4zfzO97DWmvAAAAAAAAAAA5jvWPcscbj+SeA09T/n2vn3PcD4SCqW9AAAAAAAAAACmaEQ+jojAvOa8ALtCrHQ5rNQpvlYuKjoAAIA/AACAP00PHL1ePrc/LQZNvoRRIr5kB4k8mX6kvQAAAAAAAAAAZuZOuk/Csz8TvqO93hyVvqKBcTpRXJQ8AAAAAAAAAADmmzy97te/P8txfL4x0Ru8NhjPPLhMdrwAAAAAAAAAAOb5ob11Zqw/MBSivnLA1b7B4sa9ZTQ0vgAAAAAAAAAAAMiovT1tFLvKxgI8pkaSPN/MFzwGlHy9AACAPwAAAACAmb295kuVP2YKsb6OMBK/Gt1EvnDHSL4AAAAAAAAAAIDBUD123Ru8s/b4O5+eUTxBZJI96iIxvQAAgD8AAIA/s6+wvcVwKj6G6js+FLWNvkEuUD0a6w49AAAAAAAAAAAzU8M62B23P7DLTT2TyYc+dyHeuiZuOLwAAAAAAAAAACCLdb7f0EE/6XQovYOa576LTJC+5vcbPgAAAAAAAAAAlIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUiYiHlFKUKEsDjAE8lE5OTkr/////Sv////9LAHSUYksQSwiGlIwBQ5R0lFKULg=="
|
| 35 |
},
|
| 36 |
"_last_episode_starts": {
|
| 37 |
":type:": "<class 'numpy.ndarray'>",
|
| 38 |
+
":serialized:": "gAWVgwAAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJYQAAAAAAAAAAAAAAAAAAABAAAAAAAAAACUjAVudW1weZSMBWR0eXBllJOUjAJiMZSJiIeUUpQoSwOMAXyUTk5OSv////9K/////0sAdJRiSxCFlIwBQ5R0lFKULg=="
|
| 39 |
},
|
| 40 |
"_last_original_obs": null,
|
| 41 |
"_episode_num": 0,
|
| 42 |
"use_sde": false,
|
| 43 |
"sde_sample_freq": -1,
|
| 44 |
+
"_current_progress_remaining": -0.007616000000000067,
|
| 45 |
"_stats_window_size": 100,
|
| 46 |
"ep_info_buffer": {
|
| 47 |
":type:": "<class 'collections.deque'>",
|
| 48 |
+
":serialized:": "gAWV6wsAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKUKH2UKIwBcpRHQHJ+fReC04SMAWyUS8yMAXSUR0CdfzQeFL39dX2UKGgGR0BxQrvphWo4aAdL9WgIR0Cdf0hZQpF1dX2UKGgGR0Bw7NhZyMkyaAdLzGgIR0Cdf2g88s+WdX2UKGgGR0Bysu1MM7U5aAdL62gIR0Cdf3dUsFt9dX2UKGgGR0BwGjT6SDAaaAdL5mgIR0Cdf4HJcPe6dX2UKGgGR0BxtpX+2mYTaAdL7WgIR0Cdf6eFtbcHdX2UKGgGR0ByZHk7wKBvaAdLy2gIR0CdgFqFyq+8dX2UKGgGR0BwuRG8VYZEaAdL8GgIR0CdgIkIomXxdX2UKGgGR0Bza46bONYKaAdL+WgIR0CdgQgGbCrMdX2UKGgGR0Bvg5oEjgQ6aAdLy2gIR0Cdgdm5lOGkdX2UKGgGR0ByHDnjhky2aAdL1WgIR0CdgecQRPGidX2UKGgGR0Bza63VkMCtaAdL6mgIR0CdgmAoXsPbdX2UKGgGR0BwBXbsWweOaAdL52gIR0CdgoiYsunNdX2UKGgGR0BxinWoWHk+aAdL62gIR0CdhHHBUJfIdX2UKGgGR0ByWQ1vVEuyaAdLyWgIR0CdhVmOlwcYdX2UKGgGR0B0Qu7QLNOeaAdL9mgIR0CdhZifxtpFdX2UKGgGR0BxEn0Gu9vkaAdL1WgIR0CdhjvStvGZdX2UKGgGR0BusSZ2IO6NaAdL22gIR0CdhmSqU/wBdX2UKGgGR0Bwe5dfLLZBaAdL6WgIR0CdhqaXrt3OdX2UKGgGR0BwngUZeiSJaAdL5WgIR0CdhrJQtSQ6dX2UKGgGR0BwWNkiD/VBaAdL6WgIR0CdhzbCaZx8dX2UKGgGR0BvLPyLAHmjaAdL02gIR0Cdh4ohY/3WdX2UKGgGR0Bz0tdE9dNWaAdL0mgIR0Cdh6sUqQRxdX2UKGgGR0BxxZsHjZL7aAdL3GgIR0CdiFAxzq8ldX2UKGgGR0BzYuSgXdj5aAdLzGgIR0CdiI0Q9RrKdX2UKGgGR0Bwjo76pHZsaAdL1mgIR0CdiRJnxri3dX2UKGgGR0BwzZWKdhAoaAdL62gIR0CdiTJFspG4dX2UKGgGR0Byb4150KZ2aAdL22gIR0CdiU+MZP2xdX2UKGgGR0BSadzKcNH6aAdLpmgIR0Cdm7mOEM9bdX2UKGgGR0BzKi1mapgkaAdL8mgIR0CdnBnMt9QXdX2UKGgGR0Bxp4o0ALiNaAdL02gIR0CdnCXokiUxdX2UKGgGR0Bx/8hUzbeuaAdLwmgIR0CdnGr8R+SbdX2UKGgGR0Bxt70PH1e0aAdL52gIR0CdnTccENe/dX2UKGgGR0BxcA7hegL7aAdL8WgIR0CdnV6be/HpdX2UKGgGR0BuIog/1QIlaAdNEAFoCEdAnZ2MmBvrGHV9lChoBkdAcLAXP7el9GgHS9hoCEdAnZ2md7OVxHV9lChoBkdAclkj7yhBaGgHS+VoCEdAnZ28OkLx7XV9lChoBkdAUqF9b5dnkGgHS59oCEdAnZ3gY51eSnV9lChoBkdAcIjyFPBSDWgHS+poCEdAnZ4ksSTQmnV9lChoBkdAYvuE8q4H5mgHTegDaAhHQJ2eKrilzlt1fZQoaAZHQG9/kuxrzoVoB0vcaAhHQJ2eUbm2b5N1fZQoaAZHQHKXp2pyZKFoB0viaAhHQJ2eqCrcTJ11fZQoaAZHQHISU3S8an9oB0vlaAhHQJ2fKKAJ9iN1fZQoaAZHQG5uo4EOiFloB00VAWgIR0CdoHxG2CumdX2UKGgGR0BuhDt/nW8RaAdL12gIR0CdoRUGmk30dX2UKGgGR0Bx8Frcj7hvaAdL72gIR0CdoaF+uvECdX2UKGgGR0BxItU4rBj4aAdL5WgIR0CdobL2HtWudX2UKGgGR0BwvBddE9dNaAdNAgFoCEdAnaHAqy4WlHV9lChoBkdAcTqWxyGSIWgHS81oCEdAnaJXjyWiUXV9lChoBkdAcWUFx4ptrWgHS9doCEdAnaKBcqvvB3V9lChoBkdAcEPlUp/gBWgHS+NoCEdAnaKHnlnyu3V9lChoBkdAcuFV1wHZ9WgHS99oCEdAnaKY3irDInV9lChoBkdAcD1RJEpiJGgHS9RoCEdAnaKrCBPKuHV9lChoBkdAcklCAMDwIGgHS8loCEdAnaLjdHlOoHV9lChoBkdAcT4Z4fOlf2gHS9doCEdAnaL/4EfT1HV9lChoBkdAUvh86V+qi2gHS6VoCEdAnaMQTVUdaXV9lChoBkdAcaEZ6lchT2gHTQcBaAhHQJ2jM1JlJ6J1fZQoaAZHQHH4+AAhje9oB0vaaAhHQJ2jiEEkjX51fZQoaAZHQHM3fHggow5oB00SAWgIR0CdpCCU5dWydX2UKGgGR0BttV9fCyhSaAdL3GgIR0CdpU25xzaLdX2UKGgGR0Bwn2msNlRQaAdL2WgIR0CdpcKjSG8FdX2UKGgGR0By6w6bONYKaAdL22gIR0CdpnAxBVuKdX2UKGgGR0Bx2TnJT2nLaAdL42gIR0CdprfDk2gndX2UKGgGR0BxXJ+NLlFMaAdLxWgIR0CdpsBxgiNbdX2UKGgGR0Btzy9sabWmaAdNAgFoCEdAnaddlyzXz3V9lChoBkdAb8fiTdLxqmgHS+BoCEdAnaeDnRsuWnV9lChoBkdAcg8fra/RFGgHS81oCEdAnafAiu+yq3V9lChoBkdAcK+o8ZDRdGgHS+loCEdAnafZP69CeHV9lChoBkdAbUCrkKeCkGgHTQEBaAhHQJ2oPq1PWQR1fZQoaAZHQHLfGQfZElVoB0vraAhHQJ2oShRIjGF1fZQoaAZHQHFaB86V+qloB00NAWgIR0CdqFwDeTFEdX2UKGgGR0Bttgprk8zRaAdL92gIR0CdqHmjj7yhdX2UKGgGR0ByQIMkQf6oaAdNCQFoCEdAnai3AM2FWXV9lChoBkdAcWiiHqNZNmgHS+poCEdAnajIT4+KTHV9lChoBkdAc3IXBxgiNmgHS+JoCEdAnak2jO9nLHV9lChoBkdAchY4ku6ErWgHS9poCEdAnao4mCyyEHV9lChoBkdAcQgdhy8zymgHS+BoCEdAnat9/FzdUXV9lChoBkdAb8Bp6hQFcWgHS/BoCEdAnaw3KOktVnV9lChoBkdAcuHIkJKJ22gHS+5oCEdAnazfmLcbi3V9lChoBkdAcFYacZtNz2gHS8poCEdAnazgGKQ7tHV9lChoBkdAcvafigkC3mgHS91oCEdAnaznck+otXV9lChoBkdAcNOxfv4M4WgHTTEBaAhHQJ2s+aLGaQV1fZQoaAZHQHH1yhSLqD9oB0vjaAhHQJ2s9zltCRh1fZQoaAZHQHEQ8qBmPHVoB0v0aAhHQJ2tJgCwKSh1fZQoaAZHQHJ1zp9qk/NoB0vpaAhHQJ2trJ7sv7F1fZQoaAZHQHOYnFxXGOxoB0v4aAhHQJ2t81/DtPZ1fZQoaAZHQHJskeEIw/RoB0viaAhHQJ2t872criF1fZQoaAZHQHJQBomG/N9oB0vwaAhHQJ2t+qPwNLF1fZQoaAZHQHDKbpqynk1oB0viaAhHQJ2uBRxcVxl1fZQoaAZHQHAPczZYgaFoB0vyaAhHQJ2uzQ0GeMB1fZQoaAZHQHH7oRRMvh9oB0viaAhHQJ2vioBJZnt1fZQoaAZHQHD51vuPV/doB0vsaAhHQJ2xxDc/MW51fZQoaAZHQFHjcnmaH9FoB0u5aAhHQJ2x3JaJQ+F1fZQoaAZHQHDkFKsdT5xoB0vLaAhHQJ2ygHD76551fZQoaAZHQHLNe2NNrTJoB0vVaAhHQJ2zONXHR1J1fZQoaAZHQHKNeY+jdpJoB0v6aAhHQJ2zRXYDklx1fZQoaAZHQHG4sEmplz5oB0vmaAhHQJ2za1Vo6CF1fZQoaAZHQHAmJ4B3iaRoB0vnaAhHQJ2zdSR8twt1fZQoaAZHQHFe2QXAM2FoB0vFaAhHQJ2z6UQkHD91fZQoaAZHQHFK0+otL+RoB0v2aAhHQJ2z/09QoCx1ZS4="
|
| 49 |
},
|
| 50 |
"ep_success_buffer": {
|
| 51 |
":type:": "<class 'collections.deque'>",
|
| 52 |
":serialized:": "gAWVIAAAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKULg=="
|
| 53 |
},
|
| 54 |
+
"_n_updates": 492,
|
| 55 |
"observation_space": {
|
| 56 |
":type:": "<class 'gymnasium.spaces.box.Box'>",
|
| 57 |
+
":serialized:": "gAWVcAIAAAAAAACMFGd5bW5hc2l1bS5zcGFjZXMuYm94lIwDQm94lJOUKYGUfZQojAVkdHlwZZSMBW51bXB5lGgFk5SMAmY0lImIh5RSlChLA4wBPJROTk5K/////0r/////SwB0lGKMDWJvdW5kZWRfYmVsb3eUjBJudW1weS5jb3JlLm51bWVyaWOUjAtfZnJvbWJ1ZmZlcpSTlCiWCAAAAAAAAAABAQEBAQEBAZRoB4wCYjGUiYiHlFKUKEsDjAF8lE5OTkr/////Sv////9LAHSUYksIhZSMAUOUdJRSlIwNYm91bmRlZF9hYm92ZZRoECiWCAAAAAAAAAABAQEBAQEBAZRoFEsIhZRoGHSUUpSMBl9zaGFwZZRLCIWUjANsb3eUaBAoliAAAAAAAAAAAAC0wgAAtMIAAKDAAACgwNsPScAAAKDAAAAAgAAAAICUaApLCIWUaBh0lFKUjARoaWdolGgQKJYgAAAAAAAAAAAAtEIAALRCAACgQAAAoEDbD0lAAACgQAAAgD8AAIA/lGgKSwiFlGgYdJRSlIwIbG93X3JlcHKUjFtbLTkwLiAgICAgICAgLTkwLiAgICAgICAgIC01LiAgICAgICAgIC01LiAgICAgICAgIC0zLjE0MTU5MjcgIC01LgogIC0wLiAgICAgICAgIC0wLiAgICAgICBdlIwJaGlnaF9yZXBylIxTWzkwLiAgICAgICAgOTAuICAgICAgICAgNS4gICAgICAgICA1LiAgICAgICAgIDMuMTQxNTkyNyAgNS4KICAxLiAgICAgICAgIDEuICAgICAgIF2UjApfbnBfcmFuZG9tlE51Yi4=",
|
| 58 |
"dtype": "float32",
|
| 59 |
"bounded_below": "[ True True True True True True True True]",
|
| 60 |
"bounded_above": "[ True True True True True True True True]",
|
|
|
|
| 65 |
"high": "[90. 90. 5. 5. 3.1415927 5.\n 1. 1. ]",
|
| 66 |
"low_repr": "[-90. -90. -5. -5. -3.1415927 -5.\n -0. -0. ]",
|
| 67 |
"high_repr": "[90. 90. 5. 5. 3.1415927 5.\n 1. 1. ]",
|
| 68 |
+
"_np_random": null
|
| 69 |
},
|
| 70 |
"action_space": {
|
| 71 |
":type:": "<class 'gymnasium.spaces.discrete.Discrete'>",
|
ppo-LunarLander-v2/policy.optimizer.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 88362
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e843defc19bb451add04f7231351b867dc927fae00d3efe783b45eb37522a7fd
|
| 3 |
size 88362
|
ppo-LunarLander-v2/policy.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 43762
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:066da068df85c747da4c5ef772cf31a7594985badceee45b7a4bbc0be7392d61
|
| 3 |
size 43762
|
replay.mp4
CHANGED
|
Binary files a/replay.mp4 and b/replay.mp4 differ
|
|
|
results.json
CHANGED
|
@@ -1 +1 @@
|
|
| 1 |
-
{"mean_reward":
|
|
|
|
| 1 |
+
{"mean_reward": 278.4466041, "std_reward": 22.083014518370803, "is_deterministic": true, "n_eval_episodes": 10, "eval_datetime": "2024-02-01T19:16:49.629312"}
|