Upload PPO LunarLander-v2 LunarLander-v2_PPO_ne16_ns1024_b64_e4_cpu_TotalStep24000K.zip
Browse files- LunarLander-v2_PPO_ne16_ns1024_b64_e4_cpu_TotalStep24000K.zip +2 -2
- LunarLander-v2_PPO_ne16_ns1024_b64_e4_cpu_TotalStep24000K/data +16 -16
- LunarLander-v2_PPO_ne16_ns1024_b64_e4_cpu_TotalStep24000K/policy.optimizer.pth +1 -1
- LunarLander-v2_PPO_ne16_ns1024_b64_e4_cpu_TotalStep24000K/policy.pth +1 -1
- README.md +1 -1
- config.json +1 -1
- replay.mp4 +2 -2
- results.json +1 -1
LunarLander-v2_PPO_ne16_ns1024_b64_e4_cpu_TotalStep24000K.zip
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:fe93965636abe8b0a946d5805158b5f4fffa13ee927dbcc6701820a77aecd615
|
| 3 |
+
size 147520
|
LunarLander-v2_PPO_ne16_ns1024_b64_e4_cpu_TotalStep24000K/data
CHANGED
|
@@ -4,20 +4,20 @@
|
|
| 4 |
":serialized:": "gAWVOwAAAAAAAACMIXN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi5wb2xpY2llc5SMEUFjdG9yQ3JpdGljUG9saWN5lJOULg==",
|
| 5 |
"__module__": "stable_baselines3.common.policies",
|
| 6 |
"__doc__": "\n Policy class for actor-critic algorithms (has both policy and value prediction).\n Used by A2C, PPO and the likes.\n\n :param observation_space: Observation space\n :param action_space: Action space\n :param lr_schedule: Learning rate schedule (could be constant)\n :param net_arch: The specification of the policy and value networks.\n :param activation_fn: Activation function\n :param ortho_init: Whether to use or not orthogonal initialization\n :param use_sde: Whether to use State Dependent Exploration or not\n :param log_std_init: Initial value for the log standard deviation\n :param full_std: Whether to use (n_features x n_actions) parameters\n for the std instead of only (n_features,) when using gSDE\n :param use_expln: Use ``expln()`` function instead of ``exp()`` to ensure\n a positive standard deviation (cf paper). It allows to keep variance\n above zero and prevent it from growing too fast. In practice, ``exp()`` is usually enough.\n :param squash_output: Whether to squash the output using a tanh function,\n this allows to ensure boundaries when using gSDE.\n :param features_extractor_class: Features extractor to use.\n :param features_extractor_kwargs: Keyword arguments\n to pass to the features extractor.\n :param share_features_extractor: If True, the features extractor is shared between the policy and value networks.\n :param normalize_images: Whether to normalize images or not,\n dividing by 255.0 (True by default)\n :param optimizer_class: The optimizer to use,\n ``th.optim.Adam`` by default\n :param optimizer_kwargs: Additional keyword arguments,\n excluding the learning rate, to pass to the optimizer\n ",
|
| 7 |
-
"__init__": "<function ActorCriticPolicy.__init__ at
|
| 8 |
-
"_get_constructor_parameters": "<function ActorCriticPolicy._get_constructor_parameters at
|
| 9 |
-
"reset_noise": "<function ActorCriticPolicy.reset_noise at
|
| 10 |
-
"_build_mlp_extractor": "<function ActorCriticPolicy._build_mlp_extractor at
|
| 11 |
-
"_build": "<function ActorCriticPolicy._build at
|
| 12 |
-
"forward": "<function ActorCriticPolicy.forward at
|
| 13 |
-
"extract_features": "<function ActorCriticPolicy.extract_features at
|
| 14 |
-
"_get_action_dist_from_latent": "<function ActorCriticPolicy._get_action_dist_from_latent at
|
| 15 |
-
"_predict": "<function ActorCriticPolicy._predict at
|
| 16 |
-
"evaluate_actions": "<function ActorCriticPolicy.evaluate_actions at
|
| 17 |
-
"get_distribution": "<function ActorCriticPolicy.get_distribution at
|
| 18 |
-
"predict_values": "<function ActorCriticPolicy.predict_values at
|
| 19 |
"__abstractmethods__": "frozenset()",
|
| 20 |
-
"_abc_impl": "<_abc_data object at
|
| 21 |
},
|
| 22 |
"verbose": 0,
|
| 23 |
"policy_kwargs": {},
|
|
@@ -26,12 +26,12 @@
|
|
| 26 |
"_num_timesteps_at_start": 0,
|
| 27 |
"seed": null,
|
| 28 |
"action_noise": null,
|
| 29 |
-
"start_time":
|
| 30 |
"learning_rate": 0.0003,
|
| 31 |
"tensorboard_log": null,
|
| 32 |
"_last_obs": {
|
| 33 |
":type:": "<class 'numpy.ndarray'>",
|
| 34 |
-
":serialized:": "
|
| 35 |
},
|
| 36 |
"_last_episode_starts": {
|
| 37 |
":type:": "<class 'numpy.ndarray'>",
|
|
@@ -45,7 +45,7 @@
|
|
| 45 |
"_stats_window_size": 100,
|
| 46 |
"ep_info_buffer": {
|
| 47 |
":type:": "<class 'collections.deque'>",
|
| 48 |
-
":serialized:": "
|
| 49 |
},
|
| 50 |
"ep_success_buffer": {
|
| 51 |
":type:": "<class 'collections.deque'>",
|
|
|
|
| 4 |
":serialized:": "gAWVOwAAAAAAAACMIXN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi5wb2xpY2llc5SMEUFjdG9yQ3JpdGljUG9saWN5lJOULg==",
|
| 5 |
"__module__": "stable_baselines3.common.policies",
|
| 6 |
"__doc__": "\n Policy class for actor-critic algorithms (has both policy and value prediction).\n Used by A2C, PPO and the likes.\n\n :param observation_space: Observation space\n :param action_space: Action space\n :param lr_schedule: Learning rate schedule (could be constant)\n :param net_arch: The specification of the policy and value networks.\n :param activation_fn: Activation function\n :param ortho_init: Whether to use or not orthogonal initialization\n :param use_sde: Whether to use State Dependent Exploration or not\n :param log_std_init: Initial value for the log standard deviation\n :param full_std: Whether to use (n_features x n_actions) parameters\n for the std instead of only (n_features,) when using gSDE\n :param use_expln: Use ``expln()`` function instead of ``exp()`` to ensure\n a positive standard deviation (cf paper). It allows to keep variance\n above zero and prevent it from growing too fast. In practice, ``exp()`` is usually enough.\n :param squash_output: Whether to squash the output using a tanh function,\n this allows to ensure boundaries when using gSDE.\n :param features_extractor_class: Features extractor to use.\n :param features_extractor_kwargs: Keyword arguments\n to pass to the features extractor.\n :param share_features_extractor: If True, the features extractor is shared between the policy and value networks.\n :param normalize_images: Whether to normalize images or not,\n dividing by 255.0 (True by default)\n :param optimizer_class: The optimizer to use,\n ``th.optim.Adam`` by default\n :param optimizer_kwargs: Additional keyword arguments,\n excluding the learning rate, to pass to the optimizer\n ",
|
| 7 |
+
"__init__": "<function ActorCriticPolicy.__init__ at 0x7f2803f51160>",
|
| 8 |
+
"_get_constructor_parameters": "<function ActorCriticPolicy._get_constructor_parameters at 0x7f2803f511f0>",
|
| 9 |
+
"reset_noise": "<function ActorCriticPolicy.reset_noise at 0x7f2803f51280>",
|
| 10 |
+
"_build_mlp_extractor": "<function ActorCriticPolicy._build_mlp_extractor at 0x7f2803f51310>",
|
| 11 |
+
"_build": "<function ActorCriticPolicy._build at 0x7f2803f513a0>",
|
| 12 |
+
"forward": "<function ActorCriticPolicy.forward at 0x7f2803f51430>",
|
| 13 |
+
"extract_features": "<function ActorCriticPolicy.extract_features at 0x7f2803f514c0>",
|
| 14 |
+
"_get_action_dist_from_latent": "<function ActorCriticPolicy._get_action_dist_from_latent at 0x7f2803f51550>",
|
| 15 |
+
"_predict": "<function ActorCriticPolicy._predict at 0x7f2803f515e0>",
|
| 16 |
+
"evaluate_actions": "<function ActorCriticPolicy.evaluate_actions at 0x7f2803f51670>",
|
| 17 |
+
"get_distribution": "<function ActorCriticPolicy.get_distribution at 0x7f2803f51700>",
|
| 18 |
+
"predict_values": "<function ActorCriticPolicy.predict_values at 0x7f2803f51790>",
|
| 19 |
"__abstractmethods__": "frozenset()",
|
| 20 |
+
"_abc_impl": "<_abc_data object at 0x7f2803fb2d20>"
|
| 21 |
},
|
| 22 |
"verbose": 0,
|
| 23 |
"policy_kwargs": {},
|
|
|
|
| 26 |
"_num_timesteps_at_start": 0,
|
| 27 |
"seed": null,
|
| 28 |
"action_noise": null,
|
| 29 |
+
"start_time": 1744329544138845470,
|
| 30 |
"learning_rate": 0.0003,
|
| 31 |
"tensorboard_log": null,
|
| 32 |
"_last_obs": {
|
| 33 |
":type:": "<class 'numpy.ndarray'>",
|
| 34 |
+
":serialized:": "gAWVdQIAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJYAAgAAAAAAAGbmubsUjLq6IpNLsiQTaq7Ewps5Co+eMgAAgD8AAIA/ABoaPKSAY7sgHeu7PFFAPN2bl7ySMik9AACAPwAAgD9m1lG7jw0SvM6S8DyHyq48ZE93vQY8kD0AAIA/AACAP2aGADoUvIq6tVORtXfMXLDblhY6Q560NAAAgD8AAIA/ZqTTPFIXybsfK0u+ziGcPFiGI72HRoM9AACAPwAAgD/Nsl69Nji8P5F2Br/dGBQ+jJcFvOdUsr0AAAAAAAAAADPZ/TzsSo27Dg6Avv2lZj0DggI9MEobvgAAgD8AAIA/8qeAvgFlLT+Yn4k9pbwwv0KqDb8qD1w+AAAAAAAAAABAOd49TXECPnMJzr6h5Q+/wXF8vcKGrr4AAAAAAAAAAJpjIDx7LJW6PpoLNbyefS/2k/y6oB1itAAAgD8AAIA/fdyAPkupPz/EXoC9xCQ0v+xvFT/1dES+AAAAAAAAAACjQW6+zYBgP86ZOL7p7Pu+xmsSv8qXHL4AAAAAAAAAALPGHL1xKR27AlhquJvJijyYnhk8ixxwvQAAgD8AAIA/mvv0PLbQQrw743u+mtCAPeNBWD3KodM8AACAPwAAgD9W1GC+jU8jPwbkGj5fEki/2+PXvmiZhz4AAAAAAAAAAM04O70LMZI/umNKvo8EVr8bHx++yepJvgAAAAAAAAAAlIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUiYiHlFKUKEsDjAE8lE5OTkr/////Sv////9LAHSUYksQSwiGlIwBQ5R0lFKULg=="
|
| 35 |
},
|
| 36 |
"_last_episode_starts": {
|
| 37 |
":type:": "<class 'numpy.ndarray'>",
|
|
|
|
| 45 |
"_stats_window_size": 100,
|
| 46 |
"ep_info_buffer": {
|
| 47 |
":type:": "<class 'collections.deque'>",
|
| 48 |
+
":serialized:": "gAWV4QsAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKUKH2UKIwBcpRHQHL0rmEGqxWMAWyUS6WMAXSUR0EIfePAVoHtdX2UKGgGR0Bz/de2NNrTaAdLqGgIR0EIff5VUVBVdX2UKGgGR0ByxdMDfWMCaAdLlWgIR0EIff5SmALBdX2UKGgGR0Bxrk35vcagaAdLpmgIR0EIff6iqp97dX2UKGgGR0By4U98qnWKaAdLomgIR0EIff7U8vEkdX2UKGgGR0BvYmY6XBxhaAdLmGgIR0EIff7cPWhAdX2UKGgGR0Bzh0Cp3os7aAdLu2gIR0EIff/t/nW8dX2UKGgGR0Bx8vXDm8ujaAdLk2gIR0EIfgABSLqEdX2UKGgGR0BoaDkdV/+baAdN6ANoCEdBCH4BBmseXHV9lChoBkdAckut+CsfaGgHS6doCEdBCH4BF6E8JXV9lChoBkdAcbRy3CsOomgHS51oCEdBCH4Bs1Mue3V9lChoBkdAcTuYFJQLu2gHS6RoCEdBCH4B8V32VXV9lChoBkdAcQbMUypJgGgHS55oCEdBCH4B/1DjR3V9lChoBkdAcTrN4qwyI2gHS61oCEdBCH4CurhisnV9lChoBkdAcWLsLORkmWgHS5hoCEdBCH4De5YozHV9lChoBkdAcrpf/3nIQ2gHS6loCEdBCH4D9fZ26nV9lChoBkdAcxat78ejmGgHS8doCEdBCH4EOlEZznV9lChoBkdAcqz86mwaBWgHS5hoCEdBCH4EqX0GvHV9lChoBkdAb41azNUwSWgHS5loCEdBCH4E6Qmu1XV9lChoBkdAcmR0xubZvmgHS5hoCEdBCH4E5eNT+HV9lChoBkdAdD2y1/lQuWgHS8hoCEdBCH4GQiLVF3V9lChoBkdAcmyKqGUOeGgHS6NoCEdBCH4GXw7T2HV9lChoBkdAc3q8XenAI2gHS85oCEdBCH4GgCZF5XV9lChoBkdAdCelJYkmhWgHS69oCEdBCH4GxylvZXV9lChoBkdAcliRDkU9IWgHS6JoCEdBCH4HM0WM0nV9lChoBkdAcVrJnQID5mgHS6BoCEdBCH4HMCHRC3V9lChoBkdAcguqBEroXGgHS45oCEdBCH4HKSTyKHV9lChoBkdAceohaTwDvGgHS4RoCEdBCH4Ht3wCsHV9lChoBkdAb3H2AXl8xGgHS5toCEdBCH4H1LYf4nV9lChoBkdAcg3uhsZYP2gHS7FoCEdBCH4IhZjhDXV9lChoBkdAcMIgte2NN2gHS5FoCEdBCH4Ix9RaYHV9lChoBkdAcbspvxYq5WgHS5VoCEdBCH4JzR0EHXV9lChoBkdAc3IgqVhTfmgHS6toCEdBCH4KfB+F13V9lChoBkdAcY/fGMn7YWgHS6loCEdBCH4LEhaC+XV9lChoBkdAcOUqubI91WgHS6poCEdBCH4LT0nPV3V9lChoBkdAc9lXmeUY9GgHS7JoCEdBCH4LkELYw3V9lChoBkdActCD5CWu5mgHS59oCEdBCH4ML/bTMXV9lChoBkdAcfIzByjpLWgHS5loCEdBCH4MeJiy6nV9lChoBkdAceI5v99+gGgHS41oCEdBCH4MfY+SsHV9lChoBkdAcmaqDsdDIGgHS7BoCEdBCH4M/H/953V9lChoBkdAcEOjTa0x/WgHS49oCEdBCH4NX6MzdnV9lChoBkdAcjY+r2g3+GgHS7doCEdBCH4NiC17Y3V9lChoBkdAc+TzshPj42gHS6doCEdBCH4Nr3RG+nV9lChoBkdAc8bdi2DxsmgHS7RoCEdBCH4OO+23KHV9lChoBkdAcPnnezlcQmgHS5hoCEdBCH4OxK15SnV9lChoBkdAc6UK0D2alWgHS69oCEdBCH4O3uXu3XV9lChoBkdAcpB23rleW2gHS6poCEdBCH4P8xKxs3V9lChoBkdAceyvJzT4L2gHS5doCEdBCH4QE9ZA6nV9lChoBkdAc0nKb8WKuWgHS5xoCEdBCH4REjNY83V9lChoBkdAcbRkOZssQWgHS5RoCEdBCH4SGd5IH3V9lChoBkdAcibgOBlMAWgHS7VoCEdBCH4S6K3uu3V9lChoBkdAcIN0J4SpSGgHS5ZoCEdBCH4TLxG2C3V9lChoBkdAcrLCo0hvBWgHS7poCEdBCH4Ta68QI3V9lChoBkdActeN4JNTLmgHS6toCEdBCH4UHXAdn3V9lChoBkdAcsig4wRGt2gHS6JoCEdBCH4USH/LknV9lChoBkdAcuzEidJ8OWgHS6doCEdBCH4Uw95hSnV9lChoBkdAc9R0elsP8WgHS6RoCEdBCH4U7R0EHXV9lChoBkdAcJZX6InBtWgHS5loCEdBCH4U88V58nV9lChoBkdAczG08vEjxGgHS41oCEdBCH4VExZdOnV9lChoBkdAcyJW3Sa3JGgHS6loCEdBCH4VBZr57HV9lChoBkdAdIxIvalDW2gHS9RoCEdBCH4VbYwqRXV9lChoBkdAcq9yKNyYHGgHS6VoCEdBCH4V9wYLs3V9lChoBkdAckU1r6+FlGgHS61oCEdBCH4XCqOtGXV9lChoBkdAc6njDbah6GgHS69oCEdBCH4XNU6xPnV9lChoBkdAb1g+UyHmBGgHS6FoCEdBCH4XdwR5DHV9lChoBkdAcHOpTuOS4mgHS59oCEdBCH4YOjASF3V9lChoBkdAcdgrPt2LYWgHS49oCEdBCH4Y3EwWWXV9lChoBkdAcFTvYODraGgHS65oCEdBCH4ZxlpXZHV9lChoBkdAcgKB2OhkAmgHS7RoCEdBCH4aYfr8i3V9lChoBkdAcVf6GQCCBmgHS5loCEdBCH4bG8274HV9lChoBkdAcdCM5wOvuGgHS61oCEdBCH4bNXaJynV9lChoBkdAcnTPjGT9sWgHS6JoCEdBCH4bj4agmXV9lChoBkdAcnyVLzwtrmgHS7loCEdBCH4bmNNrTHV9lChoBkdAcgPlHSWqtGgHS6loCEdBCH4blOmBOHV9lChoBkdAcqTdeIEbHmgHS55oCEdBCH4b5AGB4HV9lChoBkdAcC5XfZVXFWgHS5doCEdBCH4cNdC3PXV9lChoBkdAco1BGQSzxGgHS7loCEdBCH4cYyuZC3V9lChoBkdAdHjYvnKW9mgHS7xoCEdBCH4cg9JSSHV9lChoBkdAcRLKxLTQV2gHS4xoCEdBCH4c8OwxFnV9lChoBkdAdEQL8Jlar2gHS6BoCEdBCH4eC6+WW3V9lChoBkdAckzmKIi1RmgHS6loCEdBCH4e0gMc63V9lChoBkdAcw5uTA31jGgHS45oCEdBCH4gMpEx7HV9lChoBkdAcnVNe+mFamgHS6NoCEdBCH4gRlxwQ3V9lChoBkdAc5X9VFQVK2gHS7hoCEdBCH4gjMFEA3V9lChoBkdAcYrrYoRZlmgHS5toCEdBCH4hRwn6VXV9lChoBkdAccnQD3dsSGgHS4xoCEdBCH4hYg5imXV9lChoBkdAc2xIZIg/1WgHS6NoCEdBCH4iOGmDUXV9lChoBkdAcGpeeWfK6mgHS59oCEdBCH4ihUNrkHV9lChoBkdAcnR2PDHfdmgHS6FoCEdBCH4iqFK02XV9lChoBkdAccM11nuiOGgHS6ZoCEdBCH4i2KjzqnV9lChoBkdAcNrgsbvPT2gHS59oCEdBCH4i2Cwr2HV9lChoBkdAcZTFG5MDfWgHS5poCEdBCH4jOnGbTnV9lChoBkdAcsp9gWrOq2gHS6poCEdBCH4jpouf3HV9lChoBkdAcz1TewcHW2gHS69oCEdBCH4jqOJcgXV9lChoBkdAch2TBqKxcGgHS7loCEdBCH4kqH/LknV9lChoBkdAcxlEhaC+UWgHS5VoCEdBCH4k2B8QZnV9lChoBkdAc0rBreqJdmgHS6toCEdBCH4lDifg8HV9lChoBkdAcyEVhTfixWgHS5FoCEdBCH4ltsnAqXV9lChoBkdAcVvnCO3lS2gHS59oCEdBCH4mqYsunXV9lChoBkdAcMSB5HEuQWgHS7JoCEdBCH4nNld1MnVlLg=="
|
| 49 |
},
|
| 50 |
"ep_success_buffer": {
|
| 51 |
":type:": "<class 'collections.deque'>",
|
LunarLander-v2_PPO_ne16_ns1024_b64_e4_cpu_TotalStep24000K/policy.optimizer.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 87978
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ec68a1786ab89097fe96a73178a00113ffe521803d4dc126b433a51ade638c9e
|
| 3 |
size 87978
|
LunarLander-v2_PPO_ne16_ns1024_b64_e4_cpu_TotalStep24000K/policy.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 43634
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5821a6b7035b55482295616cf258ac1d7dea941b229b48baed9ebd20d8687444
|
| 3 |
size 43634
|
README.md
CHANGED
|
@@ -16,7 +16,7 @@ model-index:
|
|
| 16 |
type: LunarLander-v2
|
| 17 |
metrics:
|
| 18 |
- type: mean_reward
|
| 19 |
-
value: 296.
|
| 20 |
name: mean_reward
|
| 21 |
verified: false
|
| 22 |
---
|
|
|
|
| 16 |
type: LunarLander-v2
|
| 17 |
metrics:
|
| 18 |
- type: mean_reward
|
| 19 |
+
value: 296.81 +/- 20.48
|
| 20 |
name: mean_reward
|
| 21 |
verified: false
|
| 22 |
---
|
config.json
CHANGED
|
@@ -1 +1 @@
|
|
| 1 |
-
{"policy_class": {":type:": "<class 'abc.ABCMeta'>", ":serialized:": "gAWVOwAAAAAAAACMIXN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi5wb2xpY2llc5SMEUFjdG9yQ3JpdGljUG9saWN5lJOULg==", "__module__": "stable_baselines3.common.policies", "__doc__": "\n Policy class for actor-critic algorithms (has both policy and value prediction).\n Used by A2C, PPO and the likes.\n\n :param observation_space: Observation space\n :param action_space: Action space\n :param lr_schedule: Learning rate schedule (could be constant)\n :param net_arch: The specification of the policy and value networks.\n :param activation_fn: Activation function\n :param ortho_init: Whether to use or not orthogonal initialization\n :param use_sde: Whether to use State Dependent Exploration or not\n :param log_std_init: Initial value for the log standard deviation\n :param full_std: Whether to use (n_features x n_actions) parameters\n for the std instead of only (n_features,) when using gSDE\n :param use_expln: Use ``expln()`` function instead of ``exp()`` to ensure\n a positive standard deviation (cf paper). It allows to keep variance\n above zero and prevent it from growing too fast. In practice, ``exp()`` is usually enough.\n :param squash_output: Whether to squash the output using a tanh function,\n this allows to ensure boundaries when using gSDE.\n :param features_extractor_class: Features extractor to use.\n :param features_extractor_kwargs: Keyword arguments\n to pass to the features extractor.\n :param share_features_extractor: If True, the features extractor is shared between the policy and value networks.\n :param normalize_images: Whether to normalize images or not,\n dividing by 255.0 (True by default)\n :param optimizer_class: The optimizer to use,\n ``th.optim.Adam`` by default\n :param optimizer_kwargs: Additional keyword arguments,\n excluding the learning rate, to pass to the optimizer\n ", "__init__": "<function ActorCriticPolicy.__init__ at 0x7ff7606941f0>", "_get_constructor_parameters": "<function ActorCriticPolicy._get_constructor_parameters at 0x7ff760694280>", "reset_noise": "<function ActorCriticPolicy.reset_noise at 0x7ff760694310>", "_build_mlp_extractor": "<function ActorCriticPolicy._build_mlp_extractor at 0x7ff7606943a0>", "_build": "<function ActorCriticPolicy._build at 0x7ff760694430>", "forward": "<function ActorCriticPolicy.forward at 0x7ff7606944c0>", "extract_features": "<function ActorCriticPolicy.extract_features at 0x7ff760694550>", "_get_action_dist_from_latent": "<function ActorCriticPolicy._get_action_dist_from_latent at 0x7ff7606945e0>", "_predict": "<function ActorCriticPolicy._predict at 0x7ff760694670>", "evaluate_actions": "<function ActorCriticPolicy.evaluate_actions at 0x7ff760694700>", "get_distribution": "<function ActorCriticPolicy.get_distribution at 0x7ff760694790>", "predict_values": "<function ActorCriticPolicy.predict_values at 0x7ff760694820>", "__abstractmethods__": "frozenset()", "_abc_impl": "<_abc_data object at 0x7ff7606f3d20>"}, "verbose": 0, "policy_kwargs": {}, "num_timesteps": 25001984, "_total_timesteps": 25000000, "_num_timesteps_at_start": 0, "seed": null, "action_noise": null, "start_time": 1744326422856315556, "learning_rate": 0.0003, "tensorboard_log": null, "_last_obs": {":type:": "<class 'numpy.ndarray'>", ":serialized:": "gAWVdQIAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJYAAgAAAAAAAGZeKrvh8Iq6Z2EwtsLjH7G8Q406qmBZNQAAgD8AAIA/ug4tvkYEiD90IsC+NSUJv+p90L6UOsC+AAAAAAAAAACAgzE9Uy8WP+Zjizwl8me/irLtPR03Y70AAAAAAAAAAAA8kzzI6qo7kCkMvsGknr7GOQC9ve/svQAAAAAAAAAAZgqTPCaTqj+gdmg+/roTv6TrpzyDiCg+AAAAAAAAAACago48w9VcumWlf7aUYImxo0sAOwLdljUAAIA/AACAPwpSd76LWFk/52I/viIbAL8CcxK/GR2tvQAAAAAAAAAAM21XPIIsrj9uWHQ+cMn/vqbS6ztQUvc9AAAAAAAAAADNpH68SMmZuq2yQDtkQiw2DO3+Og79XLoAAIA/AACAPyZWtT2A6a0+u4/2veUBP7/NziU+07oQvgAAAAAAAAAAAO4DvD2/arvhq4c92zyCPArOrLxVUWA9AACAPwAAgD/N4O69h5fIPu7o8z0ZeTK/lsCQvn8oLz4AAAAAAAAAAMDqjL3ASK0/fV45v8Hjxr7Z1c87CtkavgAAAAAAAAAAmt8svI+xBLxyf6Q8Q2MiPQa8VT3myQO+AACAPwAAgD+zCWO+Y0Q2P/0x+7yvTBm/AgYVv6oOvj0AAAAAAAAAAH0we75bpTo/InaOvAxQAr+DBAW/tKk6PAAAAAAAAAAAlIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUiYiHlFKUKEsDjAE8lE5OTkr/////Sv////9LAHSUYksQSwiGlIwBQ5R0lFKULg=="}, "_last_episode_starts": {":type:": "<class 'numpy.ndarray'>", ":serialized:": "gAWVgwAAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJYQAAAAAAAAAAAAAAAAAAAAAAAAAAAAAACUjAVudW1weZSMBWR0eXBllJOUjAJiMZSJiIeUUpQoSwOMAXyUTk5OSv////9K/////0sAdJRiSxCFlIwBQ5R0lFKULg=="}, "_last_original_obs": null, "_episode_num": 0, "use_sde": false, "sde_sample_freq": -1, "_current_progress_remaining": -7.935999999997279e-05, "_stats_window_size": 100, "ep_info_buffer": {":type:": "<class 'collections.deque'>", ":serialized:": "gAWV4QsAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKUKH2UKIwBcpRHQHK+ZXhfjS6MAWyUTYgBjAF0lEdBCBUMV+iJwnV9lChoBkdAcC9T37DVIGgHS6toCEdBCBUMqCnP3XV9lChoBkdAc0ewqy4WlGgHS7FoCEdBCBUM2fK6nXV9lChoBkdActOGx2SuAGgHS8BoCEdBCBUNe+IuXnV9lChoBkdAcy/lbeMyamgHS6xoCEdBCBUNrYmLL3V9lChoBkdAcnjUt7KJVWgHS5ZoCEdBCBUN7wx33nV9lChoBkdAdFUt03fhuWgHS9BoCEdBCBUkZTqB3HV9lChoBkdAc7qKm8/Uv2gHS7NoCEdBCBUkp7E5yXV9lChoBkdAcn9rd30PH2gHS7FoCEdBCBUk6ZKFqXV9lChoBkdAci0h/y5I6WgHS7NoCEdBCBUlS7I1cnV9lChoBkdAc7/HjIaLoGgHS6RoCEdBCBUlYy/KyXV9lChoBkdAcYdY8uBczWgHS5BoCEdBCBUlaKxcFHV9lChoBkdAcicSmIj4YmgHS65oCEdBCBUmap3otHV9lChoBkdAc4MqJ/G2kWgHS7xoCEdBCBUmcb1h9nV9lChoBkdAcXTw9aEBbWgHS5hoCEdBCBUms580DXV9lChoBkdAcrFg1m8M/mgHS7xoCEdBCBUnIR02cnV9lChoBkdAcmeKSgXdkGgHS61oCEdBCBUnw9ic5XV9lChoBkdAc51OZssQNGgHS6hoCEdBCBUnzKs+3nV9lChoBkdAcUMkVeruIGgHS6NoCEdBCBUocSVW0nV9lChoBkdAdHsu6VdHD2gHS9toCEdBCBUooDmr83V9lChoBkdAcxwhiLEUCmgHS7VoCEdBCBUo3LaEjHV9lChoBkdAc08scyWRimgHS79oCEdBCBUpmZThpHV9lChoBkdAclEyNXHR1GgHS5ZoCEdBCBUplPnB+HV9lChoBkdAcptA8SwnpmgHS5BoCEdBCBUqQjnmrHV9lChoBkdAcDiNsFdLQGgHS6toCEdBCBUqhIg/1XV9lChoBkdAcbuDmKZUk2gHS5RoCEdBCBUqcGJN03V9lChoBkdAcXhnvlU6xWgHS6BoCEdBCBUqv0TURXV9lChoBkdAchuJEH+qBGgHS7ZoCEdBCBUrGyon8nV9lChoBkdAcBgAS39aU2gHS5FoCEdBCBUrUIzFdnV9lChoBkdAccF6+36RAGgHS45oCEdBCBUsd7IDHXV9lChoBkdAc+Gx2B8QZmgHS7RoCEdBCBUserwOOXV9lChoBkdAczTlpXZGrmgHS7ZoCEdBCBUtRKzzE3V9lChoBkdAdA08a4tpVWgHS8JoCEdBCBUtQ7xNI3V9lChoBkdAcdnyDqW1MWgHS6RoCEdBCBUtTFhod3V9lChoBkdAcRRkOI68x2gHS6FoCEdBCBUt0Vk+YHV9lChoBkdAc0YMERrad2gHS6ZoCEdBCBUuXJV81HV9lChoBkdAcl5qaw2VFGgHS6toCEdBCBUuStDD0nV9lChoBkdAcUlrzGxUvWgHS45oCEdBCBUuVGNJe3V9lChoBkdAcqnr1dxAB2gHS6FoCEdBCBUu7r9l3HV9lChoBkdAciY/io86m2gHS4RoCEdBCBUvGbiIcnV9lChoBkdAcsbZzxPO6mgHS4xoCEdBCBUvI3Lmp3V9lChoBkdAcWTUNrj5sWgHS5loCEdBCBUwbT4L1HV9lChoBkdAcZcQyylenmgHS7ZoCEdBCBUwiEYfn3V9lChoBkdAc/MrjHXEqGgHS75oCEdBCBUwqXfIjnV9lChoBkdAcY/TS9du52gHS4hoCEdBCBUx9IRRM3V9lChoBkdAdBEtXxOLzmgHS8toCEdBCBUyDfixV3V9lChoBkdAcwQya/h2n2gHS6toCEdBCBUyctdzGXV9lChoBkdAcgjW6K+BYmgHS7doCEdBCBUy3ZK3/nV9lChoBkdAcjdbBXS0B2gHS5RoCEdBCBUy/6XSjXV9lChoBkdAcfMXVbzK92gHS6hoCEdBCBUzI+Y+jnV9lChoBkdAczmPf8/D+GgHS7BoCEdBCBUzdbPhQ3V9lChoBkdAckrfCAMDwGgHS5FoCEdBCBU0NG7SRnV9lChoBkdAc2kHAAQxvmgHS6ZoCEdBCBU0TYukDnV9lChoBkdAc51u1F6RhmgHS6toCEdBCBU0b/VAiXV9lChoBkdAdCWCgK4QSWgHS7VoCEdBCBU0wh+vyXV9lChoBkdAcsBcy31BdGgHS6FoCEdBCBU086gdwXV9lChoBkdAdA1jfvWpZWgHS8poCEdBCBU2VHOKO3V9lChoBkdAchEbGFSKnGgHS6VoCEdBCBU2bkhib3V9lChoBkdAccQ7EHdGiGgHS7BoCEdBCBU2uR9w33V9lChoBkdAc/MzoEB8yGgHS7ZoCEdBCBU3IFC9iHV9lChoBkdAc4HWXkYGdWgHS5toCEdBCBU3aeqaPXV9lChoBkdAcK1MQVbiZWgHS6xoCEdBCBU4GryUcHV9lChoBkdAcrnTgEU0vWgHS6JoCEdBCBU4GtnwonV9lChoBkdAcV7uSOinHmgHS5loCEdBCBU4NB0IT3V9lChoBkdAcytz+FUQ1GgHS5toCEdBCBU4ZR8+inV9lChoBkdAcc4zzmOlwmgHS6JoCEdBCBU4sXrMT3V9lChoBkdAdBeNqQA+6mgHS7hoCEdBCBU5uFpPAXV9lChoBkdAcO9eEqUeMmgHS6RoCEdBCBU529g4O3V9lChoBkdAcnW2LpA2RGgHS7xoCEdBCBU6ojD8+HV9lChoBkdAcle3ueBg/mgHS71oCEdBCBU7bzPKMnV9lChoBkdAcuHjSofjj2gHS8NoCEdBCBU7eMQ2/HV9lChoBkdAcNOupjtojGgHS55oCEdBCBU73AymAXV9lChoBkdAcQqINEw352gHS7BoCEdBCBU8+tyPuHV9lChoBkdAcK/dxyXD32gHS45oCEdBCBU9PgaWHHV9lChoBkdAc80XQ+lj3GgHS8JoCEdBCBU9ZQUHp3V9lChoBkdAcWqcZccENmgHS6hoCEdBCBU+cSAYpHV9lChoBkdAc7SDJU5uImgHS75oCEdBCBU+dEuxr3V9lChoBkdAchKgOSW7e2gHS7VoCEdBCBU+55s0pHV9lChoBkdAc/luUUwi7mgHS7FoCEdBCBU/Ez9CNXV9lChoBkdAc0LQhfShJ2gHS6xoCEdBCBU/Tb+LnHV9lChoBkdAclQ2t+1Bt2gHS+9oCEdBCBVAJbdJrnV9lChoBkdAcWlgte2NN2gHS7loCEdBCBVBDH0btXV9lChoBkdAcp3zQNTcZmgHS8JoCEdBCBVBjo8p1HV9lChoBkdAcxK+AVfu1GgHS41oCEdBCBVBn3L3bnV9lChoBkdAc917/4qPO2gHS61oCEdBCBVCoQBgeHV9lChoBkdAc7fQGOdXk2gHS7RoCEdBCBVC4nc+JXV9lChoBkdAc0C0VJtix2gHS81oCEdBCBVDBgeA/nV9lChoBkdAcCtNucc2i2gHS5toCEdBCBVDutKZlXV9lChoBkdAc+xEV32VV2gHS6toCEdBCBVEjPKMenV9lChoBkdAdA3GQSzw+mgHS89oCEdBCBVFqyprDnV9lChoBkdAcOcKNQ0oB2gHS5ZoCEdBCBVGcr3CbnV9lChoBkdAc28YbKifx2gHS7toCEdBCBVGvgFX73V9lChoBkdAcrXWac7Qs2gHS7ZoCEdBCBVG8u+RHXV9lChoBkdAdDGHp8neBWgHS81oCEdBCBVHEh/y5XV9lChoBkdAdLXEit7rs2gHS9RoCEdBCBVHWinHenV9lChoBkdAdBhQWvbGm2gHS8loCEdBCBVHfPC2t3V9lChoBkdAcM+1mapgkWgHS5poCEdBCBVH+aWonHV9lChoBkdAdDOlkYoAn2gHS7poCEdBCBVIuUpuuXV9lChoBkdAcSlnSOR1YGgHS7BoCEdBCBVI4Ucn3XV9lChoBkdAcebXWvr4WWgHS5toCEdBCBVI5w6ySnVlLg=="}, "ep_success_buffer": {":type:": "<class 'collections.deque'>", ":serialized:": "gAWVIAAAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKULg=="}, "_n_updates": 79400, "observation_space": {":type:": "<class 'gymnasium.spaces.box.Box'>", ":serialized:": "gAWVcAIAAAAAAACMFGd5bW5hc2l1bS5zcGFjZXMuYm94lIwDQm94lJOUKYGUfZQojAVkdHlwZZSMBW51bXB5lGgFk5SMAmY0lImIh5RSlChLA4wBPJROTk5K/////0r/////SwB0lGKMDWJvdW5kZWRfYmVsb3eUjBJudW1weS5jb3JlLm51bWVyaWOUjAtfZnJvbWJ1ZmZlcpSTlCiWCAAAAAAAAAABAQEBAQEBAZRoB4wCYjGUiYiHlFKUKEsDjAF8lE5OTkr/////Sv////9LAHSUYksIhZSMAUOUdJRSlIwNYm91bmRlZF9hYm92ZZRoECiWCAAAAAAAAAABAQEBAQEBAZRoFEsIhZRoGHSUUpSMBl9zaGFwZZRLCIWUjANsb3eUaBAoliAAAAAAAAAAAAC0wgAAtMIAAKDAAACgwNsPScAAAKDAAAAAgAAAAICUaApLCIWUaBh0lFKUjARoaWdolGgQKJYgAAAAAAAAAAAAtEIAALRCAACgQAAAoEDbD0lAAACgQAAAgD8AAIA/lGgKSwiFlGgYdJRSlIwIbG93X3JlcHKUjFtbLTkwLiAgICAgICAgLTkwLiAgICAgICAgIC01LiAgICAgICAgIC01LiAgICAgICAgIC0zLjE0MTU5MjcgIC01LgogIC0wLiAgICAgICAgIC0wLiAgICAgICBdlIwJaGlnaF9yZXBylIxTWzkwLiAgICAgICAgOTAuICAgICAgICAgNS4gICAgICAgICA1LiAgICAgICAgIDMuMTQxNTkyNyAgNS4KICAxLiAgICAgICAgIDEuICAgICAgIF2UjApfbnBfcmFuZG9tlE51Yi4=", "dtype": "float32", "bounded_below": "[ True True True True True True True True]", "bounded_above": "[ True True True True True True True True]", "_shape": [8], "low": "[-90. -90. -5. -5. -3.1415927 -5.\n -0. -0. ]", "high": "[90. 90. 5. 5. 3.1415927 5.\n 1. 1. ]", "low_repr": "[-90. -90. -5. -5. -3.1415927 -5.\n -0. -0. ]", "high_repr": "[90. 90. 5. 5. 3.1415927 5.\n 1. 1. ]", "_np_random": null}, "action_space": {":type:": "<class 'gymnasium.spaces.discrete.Discrete'>", ":serialized:": "gAWV1QAAAAAAAACMGWd5bW5hc2l1bS5zcGFjZXMuZGlzY3JldGWUjAhEaXNjcmV0ZZSTlCmBlH2UKIwBbpSMFW51bXB5LmNvcmUubXVsdGlhcnJheZSMBnNjYWxhcpSTlIwFbnVtcHmUjAVkdHlwZZSTlIwCaTiUiYiHlFKUKEsDjAE8lE5OTkr/////Sv////9LAHSUYkMIBAAAAAAAAACUhpRSlIwFc3RhcnSUaAhoDkMIAAAAAAAAAACUhpRSlIwGX3NoYXBllCloCmgOjApfbnBfcmFuZG9tlE51Yi4=", "n": "4", "start": "0", "_shape": [], "dtype": "int64", "_np_random": null}, "n_envs": 16, "n_steps": 1024, "gamma": 0.999, "gae_lambda": 0.98, "ent_coef": 0.01, "vf_coef": 0.5, "max_grad_norm": 0.5, "batch_size": 64, "n_epochs": 4, "clip_range": {":type:": "<class 'function'>", ":serialized:": "gAWV5QIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwOX21ha2VfZnVuY3Rpb26Uk5QoaACMDV9idWlsdGluX3R5cGWUk5SMCENvZGVUeXBllIWUUpQoSwFLAEsASwFLAUsTQwSIAFMAlE6FlCmMAV+UhZSMZS9ob21lL3NlbmxpLy5jb25kYS9lbnZzL3NlbmxpX1B5dGhvbjMuOC9saWIvcHl0aG9uMy44L3NpdGUtcGFja2FnZXMvc3RhYmxlX2Jhc2VsaW5lczMvY29tbW9uL3V0aWxzLnB5lIwEZnVuY5RLhEMCAAGUjAN2YWyUhZQpdJRSlH2UKIwLX19wYWNrYWdlX1+UjBhzdGFibGVfYmFzZWxpbmVzMy5jb21tb26UjAhfX25hbWVfX5SMHnN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi51dGlsc5SMCF9fZmlsZV9flIxlL2hvbWUvc2VubGkvLmNvbmRhL2VudnMvc2VubGlfUHl0aG9uMy44L2xpYi9weXRob24zLjgvc2l0ZS1wYWNrYWdlcy9zdGFibGVfYmFzZWxpbmVzMy9jb21tb24vdXRpbHMucHmUdU5OaACMEF9tYWtlX2VtcHR5X2NlbGyUk5QpUpSFlHSUUpRoAIwSX2Z1bmN0aW9uX3NldHN0YXRllJOUaB99lH2UKGgWjARmdW5jlIwMX19xdWFsbmFtZV9flIwZY29uc3RhbnRfZm4uPGxvY2Fscz4uZnVuY5SMD19fYW5ub3RhdGlvbnNfX5R9lIwOX19rd2RlZmF1bHRzX1+UTowMX19kZWZhdWx0c19flE6MCl9fbW9kdWxlX1+UaBeMB19fZG9jX1+UTowLX19jbG9zdXJlX1+UaACMCl9tYWtlX2NlbGyUk5RHP8mZmZmZmZqFlFKUhZSMF19jbG91ZHBpY2tsZV9zdWJtb2R1bGVzlF2UjAtfX2dsb2JhbHNfX5R9lHWGlIZSMC4="}, "clip_range_vf": null, "normalize_advantage": true, "target_kl": null, "lr_schedule": {":type:": "<class 'function'>", ":serialized:": "gAWV5QIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwOX21ha2VfZnVuY3Rpb26Uk5QoaACMDV9idWlsdGluX3R5cGWUk5SMCENvZGVUeXBllIWUUpQoSwFLAEsASwFLAUsTQwSIAFMAlE6FlCmMAV+UhZSMZS9ob21lL3NlbmxpLy5jb25kYS9lbnZzL3NlbmxpX1B5dGhvbjMuOC9saWIvcHl0aG9uMy44L3NpdGUtcGFja2FnZXMvc3RhYmxlX2Jhc2VsaW5lczMvY29tbW9uL3V0aWxzLnB5lIwEZnVuY5RLhEMCAAGUjAN2YWyUhZQpdJRSlH2UKIwLX19wYWNrYWdlX1+UjBhzdGFibGVfYmFzZWxpbmVzMy5jb21tb26UjAhfX25hbWVfX5SMHnN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi51dGlsc5SMCF9fZmlsZV9flIxlL2hvbWUvc2VubGkvLmNvbmRhL2VudnMvc2VubGlfUHl0aG9uMy44L2xpYi9weXRob24zLjgvc2l0ZS1wYWNrYWdlcy9zdGFibGVfYmFzZWxpbmVzMy9jb21tb24vdXRpbHMucHmUdU5OaACMEF9tYWtlX2VtcHR5X2NlbGyUk5QpUpSFlHSUUpRoAIwSX2Z1bmN0aW9uX3NldHN0YXRllJOUaB99lH2UKGgWjARmdW5jlIwMX19xdWFsbmFtZV9flIwZY29uc3RhbnRfZm4uPGxvY2Fscz4uZnVuY5SMD19fYW5ub3RhdGlvbnNfX5R9lIwOX19rd2RlZmF1bHRzX1+UTowMX19kZWZhdWx0c19flE6MCl9fbW9kdWxlX1+UaBeMB19fZG9jX1+UTowLX19jbG9zdXJlX1+UaACMCl9tYWtlX2NlbGyUk5RHPzOpKjBVMmGFlFKUhZSMF19jbG91ZHBpY2tsZV9zdWJtb2R1bGVzlF2UjAtfX2dsb2JhbHNfX5R9lHWGlIZSMC4="}, "system_info": {"OS": "Linux-5.15.0-134-generic-x86_64-with-glibc2.17 # 145-Ubuntu SMP Wed Feb 12 20:08:39 UTC 2025", "Python": "3.8.20", "Stable-Baselines3": "2.0.0a5", "PyTorch": "2.4.1+cu121", "GPU Enabled": "True", "Numpy": "1.24.4", "Cloudpickle": "3.1.1", "Gymnasium": "0.28.1", "OpenAI Gym": "0.26.2"}}
|
|
|
|
| 1 |
+
{"policy_class": {":type:": "<class 'abc.ABCMeta'>", ":serialized:": "gAWVOwAAAAAAAACMIXN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi5wb2xpY2llc5SMEUFjdG9yQ3JpdGljUG9saWN5lJOULg==", "__module__": "stable_baselines3.common.policies", "__doc__": "\n Policy class for actor-critic algorithms (has both policy and value prediction).\n Used by A2C, PPO and the likes.\n\n :param observation_space: Observation space\n :param action_space: Action space\n :param lr_schedule: Learning rate schedule (could be constant)\n :param net_arch: The specification of the policy and value networks.\n :param activation_fn: Activation function\n :param ortho_init: Whether to use or not orthogonal initialization\n :param use_sde: Whether to use State Dependent Exploration or not\n :param log_std_init: Initial value for the log standard deviation\n :param full_std: Whether to use (n_features x n_actions) parameters\n for the std instead of only (n_features,) when using gSDE\n :param use_expln: Use ``expln()`` function instead of ``exp()`` to ensure\n a positive standard deviation (cf paper). It allows to keep variance\n above zero and prevent it from growing too fast. In practice, ``exp()`` is usually enough.\n :param squash_output: Whether to squash the output using a tanh function,\n this allows to ensure boundaries when using gSDE.\n :param features_extractor_class: Features extractor to use.\n :param features_extractor_kwargs: Keyword arguments\n to pass to the features extractor.\n :param share_features_extractor: If True, the features extractor is shared between the policy and value networks.\n :param normalize_images: Whether to normalize images or not,\n dividing by 255.0 (True by default)\n :param optimizer_class: The optimizer to use,\n ``th.optim.Adam`` by default\n :param optimizer_kwargs: Additional keyword arguments,\n excluding the learning rate, to pass to the optimizer\n ", "__init__": "<function ActorCriticPolicy.__init__ at 0x7f2803f51160>", "_get_constructor_parameters": "<function ActorCriticPolicy._get_constructor_parameters at 0x7f2803f511f0>", "reset_noise": "<function ActorCriticPolicy.reset_noise at 0x7f2803f51280>", "_build_mlp_extractor": "<function ActorCriticPolicy._build_mlp_extractor at 0x7f2803f51310>", "_build": "<function ActorCriticPolicy._build at 0x7f2803f513a0>", "forward": "<function ActorCriticPolicy.forward at 0x7f2803f51430>", "extract_features": "<function ActorCriticPolicy.extract_features at 0x7f2803f514c0>", "_get_action_dist_from_latent": "<function ActorCriticPolicy._get_action_dist_from_latent at 0x7f2803f51550>", "_predict": "<function ActorCriticPolicy._predict at 0x7f2803f515e0>", "evaluate_actions": "<function ActorCriticPolicy.evaluate_actions at 0x7f2803f51670>", "get_distribution": "<function ActorCriticPolicy.get_distribution at 0x7f2803f51700>", "predict_values": "<function ActorCriticPolicy.predict_values at 0x7f2803f51790>", "__abstractmethods__": "frozenset()", "_abc_impl": "<_abc_data object at 0x7f2803fb2d20>"}, "verbose": 0, "policy_kwargs": {}, "num_timesteps": 24002560, "_total_timesteps": 24000000, "_num_timesteps_at_start": 0, "seed": null, "action_noise": null, "start_time": 1744329544138845470, "learning_rate": 0.0003, "tensorboard_log": null, "_last_obs": {":type:": "<class 'numpy.ndarray'>", ":serialized:": "gAWVdQIAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJYAAgAAAAAAAGbmubsUjLq6IpNLsiQTaq7Ewps5Co+eMgAAgD8AAIA/ABoaPKSAY7sgHeu7PFFAPN2bl7ySMik9AACAPwAAgD9m1lG7jw0SvM6S8DyHyq48ZE93vQY8kD0AAIA/AACAP2aGADoUvIq6tVORtXfMXLDblhY6Q560NAAAgD8AAIA/ZqTTPFIXybsfK0u+ziGcPFiGI72HRoM9AACAPwAAgD/Nsl69Nji8P5F2Br/dGBQ+jJcFvOdUsr0AAAAAAAAAADPZ/TzsSo27Dg6Avv2lZj0DggI9MEobvgAAgD8AAIA/8qeAvgFlLT+Yn4k9pbwwv0KqDb8qD1w+AAAAAAAAAABAOd49TXECPnMJzr6h5Q+/wXF8vcKGrr4AAAAAAAAAAJpjIDx7LJW6PpoLNbyefS/2k/y6oB1itAAAgD8AAIA/fdyAPkupPz/EXoC9xCQ0v+xvFT/1dES+AAAAAAAAAACjQW6+zYBgP86ZOL7p7Pu+xmsSv8qXHL4AAAAAAAAAALPGHL1xKR27AlhquJvJijyYnhk8ixxwvQAAgD8AAIA/mvv0PLbQQrw743u+mtCAPeNBWD3KodM8AACAPwAAgD9W1GC+jU8jPwbkGj5fEki/2+PXvmiZhz4AAAAAAAAAAM04O70LMZI/umNKvo8EVr8bHx++yepJvgAAAAAAAAAAlIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUiYiHlFKUKEsDjAE8lE5OTkr/////Sv////9LAHSUYksQSwiGlIwBQ5R0lFKULg=="}, "_last_episode_starts": {":type:": "<class 'numpy.ndarray'>", ":serialized:": "gAWVgwAAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJYQAAAAAAAAAAAAAAAAAAAAAAAAAAAAAACUjAVudW1weZSMBWR0eXBllJOUjAJiMZSJiIeUUpQoSwOMAXyUTk5OSv////9K/////0sAdJRiSxCFlIwBQ5R0lFKULg=="}, "_last_original_obs": null, "_episode_num": 0, "use_sde": false, "sde_sample_freq": -1, "_current_progress_remaining": -0.00010666666666669933, "_stats_window_size": 100, "ep_info_buffer": {":type:": "<class 'collections.deque'>", ":serialized:": "gAWV4QsAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKUKH2UKIwBcpRHQHL0rmEGqxWMAWyUS6WMAXSUR0EIfePAVoHtdX2UKGgGR0Bz/de2NNrTaAdLqGgIR0EIff5VUVBVdX2UKGgGR0ByxdMDfWMCaAdLlWgIR0EIff5SmALBdX2UKGgGR0Bxrk35vcagaAdLpmgIR0EIff6iqp97dX2UKGgGR0By4U98qnWKaAdLomgIR0EIff7U8vEkdX2UKGgGR0BvYmY6XBxhaAdLmGgIR0EIff7cPWhAdX2UKGgGR0Bzh0Cp3os7aAdLu2gIR0EIff/t/nW8dX2UKGgGR0Bx8vXDm8ujaAdLk2gIR0EIfgABSLqEdX2UKGgGR0BoaDkdV/+baAdN6ANoCEdBCH4BBmseXHV9lChoBkdAckut+CsfaGgHS6doCEdBCH4BF6E8JXV9lChoBkdAcbRy3CsOomgHS51oCEdBCH4Bs1Mue3V9lChoBkdAcTuYFJQLu2gHS6RoCEdBCH4B8V32VXV9lChoBkdAcQbMUypJgGgHS55oCEdBCH4B/1DjR3V9lChoBkdAcTrN4qwyI2gHS61oCEdBCH4CurhisnV9lChoBkdAcWLsLORkmWgHS5hoCEdBCH4De5YozHV9lChoBkdAcrpf/3nIQ2gHS6loCEdBCH4D9fZ26nV9lChoBkdAcxat78ejmGgHS8doCEdBCH4EOlEZznV9lChoBkdAcqz86mwaBWgHS5hoCEdBCH4EqX0GvHV9lChoBkdAb41azNUwSWgHS5loCEdBCH4E6Qmu1XV9lChoBkdAcmR0xubZvmgHS5hoCEdBCH4E5eNT+HV9lChoBkdAdD2y1/lQuWgHS8hoCEdBCH4GQiLVF3V9lChoBkdAcmyKqGUOeGgHS6NoCEdBCH4GXw7T2HV9lChoBkdAc3q8XenAI2gHS85oCEdBCH4GgCZF5XV9lChoBkdAdCelJYkmhWgHS69oCEdBCH4GxylvZXV9lChoBkdAcliRDkU9IWgHS6JoCEdBCH4HM0WM0nV9lChoBkdAcVrJnQID5mgHS6BoCEdBCH4HMCHRC3V9lChoBkdAcguqBEroXGgHS45oCEdBCH4HKSTyKHV9lChoBkdAceohaTwDvGgHS4RoCEdBCH4Ht3wCsHV9lChoBkdAb3H2AXl8xGgHS5toCEdBCH4H1LYf4nV9lChoBkdAcg3uhsZYP2gHS7FoCEdBCH4IhZjhDXV9lChoBkdAcMIgte2NN2gHS5FoCEdBCH4Ix9RaYHV9lChoBkdAcbspvxYq5WgHS5VoCEdBCH4JzR0EHXV9lChoBkdAc3IgqVhTfmgHS6toCEdBCH4KfB+F13V9lChoBkdAcY/fGMn7YWgHS6loCEdBCH4LEhaC+XV9lChoBkdAcOUqubI91WgHS6poCEdBCH4LT0nPV3V9lChoBkdAc9lXmeUY9GgHS7JoCEdBCH4LkELYw3V9lChoBkdActCD5CWu5mgHS59oCEdBCH4ML/bTMXV9lChoBkdAcfIzByjpLWgHS5loCEdBCH4MeJiy6nV9lChoBkdAceI5v99+gGgHS41oCEdBCH4MfY+SsHV9lChoBkdAcmaqDsdDIGgHS7BoCEdBCH4M/H/953V9lChoBkdAcEOjTa0x/WgHS49oCEdBCH4NX6MzdnV9lChoBkdAcjY+r2g3+GgHS7doCEdBCH4NiC17Y3V9lChoBkdAc+TzshPj42gHS6doCEdBCH4Nr3RG+nV9lChoBkdAc8bdi2DxsmgHS7RoCEdBCH4OO+23KHV9lChoBkdAcPnnezlcQmgHS5hoCEdBCH4OxK15SnV9lChoBkdAc6UK0D2alWgHS69oCEdBCH4O3uXu3XV9lChoBkdAcpB23rleW2gHS6poCEdBCH4P8xKxs3V9lChoBkdAceyvJzT4L2gHS5doCEdBCH4QE9ZA6nV9lChoBkdAc0nKb8WKuWgHS5xoCEdBCH4REjNY83V9lChoBkdAcbRkOZssQWgHS5RoCEdBCH4SGd5IH3V9lChoBkdAcibgOBlMAWgHS7VoCEdBCH4S6K3uu3V9lChoBkdAcIN0J4SpSGgHS5ZoCEdBCH4TLxG2C3V9lChoBkdAcrLCo0hvBWgHS7poCEdBCH4Ta68QI3V9lChoBkdActeN4JNTLmgHS6toCEdBCH4UHXAdn3V9lChoBkdAcsig4wRGt2gHS6JoCEdBCH4USH/LknV9lChoBkdAcuzEidJ8OWgHS6doCEdBCH4Uw95hSnV9lChoBkdAc9R0elsP8WgHS6RoCEdBCH4U7R0EHXV9lChoBkdAcJZX6InBtWgHS5loCEdBCH4U88V58nV9lChoBkdAczG08vEjxGgHS41oCEdBCH4VExZdOnV9lChoBkdAcyJW3Sa3JGgHS6loCEdBCH4VBZr57HV9lChoBkdAdIxIvalDW2gHS9RoCEdBCH4VbYwqRXV9lChoBkdAcq9yKNyYHGgHS6VoCEdBCH4V9wYLs3V9lChoBkdAckU1r6+FlGgHS61oCEdBCH4XCqOtGXV9lChoBkdAc6njDbah6GgHS69oCEdBCH4XNU6xPnV9lChoBkdAb1g+UyHmBGgHS6FoCEdBCH4XdwR5DHV9lChoBkdAcHOpTuOS4mgHS59oCEdBCH4YOjASF3V9lChoBkdAcdgrPt2LYWgHS49oCEdBCH4Y3EwWWXV9lChoBkdAcFTvYODraGgHS65oCEdBCH4ZxlpXZHV9lChoBkdAcgKB2OhkAmgHS7RoCEdBCH4aYfr8i3V9lChoBkdAcVf6GQCCBmgHS5loCEdBCH4bG8274HV9lChoBkdAcdCM5wOvuGgHS61oCEdBCH4bNXaJynV9lChoBkdAcnTPjGT9sWgHS6JoCEdBCH4bj4agmXV9lChoBkdAcnyVLzwtrmgHS7loCEdBCH4bmNNrTHV9lChoBkdAcgPlHSWqtGgHS6loCEdBCH4blOmBOHV9lChoBkdAcqTdeIEbHmgHS55oCEdBCH4b5AGB4HV9lChoBkdAcC5XfZVXFWgHS5doCEdBCH4cNdC3PXV9lChoBkdAco1BGQSzxGgHS7loCEdBCH4cYyuZC3V9lChoBkdAdHjYvnKW9mgHS7xoCEdBCH4cg9JSSHV9lChoBkdAcRLKxLTQV2gHS4xoCEdBCH4c8OwxFnV9lChoBkdAdEQL8Jlar2gHS6BoCEdBCH4eC6+WW3V9lChoBkdAckzmKIi1RmgHS6loCEdBCH4e0gMc63V9lChoBkdAcw5uTA31jGgHS45oCEdBCH4gMpEx7HV9lChoBkdAcnVNe+mFamgHS6NoCEdBCH4gRlxwQ3V9lChoBkdAc5X9VFQVK2gHS7hoCEdBCH4gjMFEA3V9lChoBkdAcYrrYoRZlmgHS5toCEdBCH4hRwn6VXV9lChoBkdAccnQD3dsSGgHS4xoCEdBCH4hYg5imXV9lChoBkdAc2xIZIg/1WgHS6NoCEdBCH4iOGmDUXV9lChoBkdAcGpeeWfK6mgHS59oCEdBCH4ihUNrkHV9lChoBkdAcnR2PDHfdmgHS6FoCEdBCH4iqFK02XV9lChoBkdAccM11nuiOGgHS6ZoCEdBCH4i2KjzqnV9lChoBkdAcNrgsbvPT2gHS59oCEdBCH4i2Cwr2HV9lChoBkdAcZTFG5MDfWgHS5poCEdBCH4jOnGbTnV9lChoBkdAcsp9gWrOq2gHS6poCEdBCH4jpouf3HV9lChoBkdAcz1TewcHW2gHS69oCEdBCH4jqOJcgXV9lChoBkdAch2TBqKxcGgHS7loCEdBCH4kqH/LknV9lChoBkdAcxlEhaC+UWgHS5VoCEdBCH4k2B8QZnV9lChoBkdAc0rBreqJdmgHS6toCEdBCH4lDifg8HV9lChoBkdAcyEVhTfixWgHS5FoCEdBCH4ltsnAqXV9lChoBkdAcVvnCO3lS2gHS59oCEdBCH4mqYsunXV9lChoBkdAcMSB5HEuQWgHS7JoCEdBCH4nNld1MnVlLg=="}, "ep_success_buffer": {":type:": "<class 'collections.deque'>", ":serialized:": "gAWVIAAAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKULg=="}, "_n_updates": 73296, "observation_space": {":type:": "<class 'gymnasium.spaces.box.Box'>", ":serialized:": "gAWVcAIAAAAAAACMFGd5bW5hc2l1bS5zcGFjZXMuYm94lIwDQm94lJOUKYGUfZQojAVkdHlwZZSMBW51bXB5lGgFk5SMAmY0lImIh5RSlChLA4wBPJROTk5K/////0r/////SwB0lGKMDWJvdW5kZWRfYmVsb3eUjBJudW1weS5jb3JlLm51bWVyaWOUjAtfZnJvbWJ1ZmZlcpSTlCiWCAAAAAAAAAABAQEBAQEBAZRoB4wCYjGUiYiHlFKUKEsDjAF8lE5OTkr/////Sv////9LAHSUYksIhZSMAUOUdJRSlIwNYm91bmRlZF9hYm92ZZRoECiWCAAAAAAAAAABAQEBAQEBAZRoFEsIhZRoGHSUUpSMBl9zaGFwZZRLCIWUjANsb3eUaBAoliAAAAAAAAAAAAC0wgAAtMIAAKDAAACgwNsPScAAAKDAAAAAgAAAAICUaApLCIWUaBh0lFKUjARoaWdolGgQKJYgAAAAAAAAAAAAtEIAALRCAACgQAAAoEDbD0lAAACgQAAAgD8AAIA/lGgKSwiFlGgYdJRSlIwIbG93X3JlcHKUjFtbLTkwLiAgICAgICAgLTkwLiAgICAgICAgIC01LiAgICAgICAgIC01LiAgICAgICAgIC0zLjE0MTU5MjcgIC01LgogIC0wLiAgICAgICAgIC0wLiAgICAgICBdlIwJaGlnaF9yZXBylIxTWzkwLiAgICAgICAgOTAuICAgICAgICAgNS4gICAgICAgICA1LiAgICAgICAgIDMuMTQxNTkyNyAgNS4KICAxLiAgICAgICAgIDEuICAgICAgIF2UjApfbnBfcmFuZG9tlE51Yi4=", "dtype": "float32", "bounded_below": "[ True True True True True True True True]", "bounded_above": "[ True True True True True True True True]", "_shape": [8], "low": "[-90. -90. -5. -5. -3.1415927 -5.\n -0. -0. ]", "high": "[90. 90. 5. 5. 3.1415927 5.\n 1. 1. ]", "low_repr": "[-90. -90. -5. -5. -3.1415927 -5.\n -0. -0. ]", "high_repr": "[90. 90. 5. 5. 3.1415927 5.\n 1. 1. ]", "_np_random": null}, "action_space": {":type:": "<class 'gymnasium.spaces.discrete.Discrete'>", ":serialized:": "gAWV1QAAAAAAAACMGWd5bW5hc2l1bS5zcGFjZXMuZGlzY3JldGWUjAhEaXNjcmV0ZZSTlCmBlH2UKIwBbpSMFW51bXB5LmNvcmUubXVsdGlhcnJheZSMBnNjYWxhcpSTlIwFbnVtcHmUjAVkdHlwZZSTlIwCaTiUiYiHlFKUKEsDjAE8lE5OTkr/////Sv////9LAHSUYkMIBAAAAAAAAACUhpRSlIwFc3RhcnSUaAhoDkMIAAAAAAAAAACUhpRSlIwGX3NoYXBllCloCmgOjApfbnBfcmFuZG9tlE51Yi4=", "n": "4", "start": "0", "_shape": [], "dtype": "int64", "_np_random": null}, "n_envs": 16, "n_steps": 1024, "gamma": 0.999, "gae_lambda": 0.98, "ent_coef": 0.01, "vf_coef": 0.5, "max_grad_norm": 0.5, "batch_size": 64, "n_epochs": 4, "clip_range": {":type:": "<class 'function'>", ":serialized:": "gAWV5QIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwOX21ha2VfZnVuY3Rpb26Uk5QoaACMDV9idWlsdGluX3R5cGWUk5SMCENvZGVUeXBllIWUUpQoSwFLAEsASwFLAUsTQwSIAFMAlE6FlCmMAV+UhZSMZS9ob21lL3NlbmxpLy5jb25kYS9lbnZzL3NlbmxpX1B5dGhvbjMuOC9saWIvcHl0aG9uMy44L3NpdGUtcGFja2FnZXMvc3RhYmxlX2Jhc2VsaW5lczMvY29tbW9uL3V0aWxzLnB5lIwEZnVuY5RLhEMCAAGUjAN2YWyUhZQpdJRSlH2UKIwLX19wYWNrYWdlX1+UjBhzdGFibGVfYmFzZWxpbmVzMy5jb21tb26UjAhfX25hbWVfX5SMHnN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi51dGlsc5SMCF9fZmlsZV9flIxlL2hvbWUvc2VubGkvLmNvbmRhL2VudnMvc2VubGlfUHl0aG9uMy44L2xpYi9weXRob24zLjgvc2l0ZS1wYWNrYWdlcy9zdGFibGVfYmFzZWxpbmVzMy9jb21tb24vdXRpbHMucHmUdU5OaACMEF9tYWtlX2VtcHR5X2NlbGyUk5QpUpSFlHSUUpRoAIwSX2Z1bmN0aW9uX3NldHN0YXRllJOUaB99lH2UKGgWjARmdW5jlIwMX19xdWFsbmFtZV9flIwZY29uc3RhbnRfZm4uPGxvY2Fscz4uZnVuY5SMD19fYW5ub3RhdGlvbnNfX5R9lIwOX19rd2RlZmF1bHRzX1+UTowMX19kZWZhdWx0c19flE6MCl9fbW9kdWxlX1+UaBeMB19fZG9jX1+UTowLX19jbG9zdXJlX1+UaACMCl9tYWtlX2NlbGyUk5RHP8mZmZmZmZqFlFKUhZSMF19jbG91ZHBpY2tsZV9zdWJtb2R1bGVzlF2UjAtfX2dsb2JhbHNfX5R9lHWGlIZSMC4="}, "clip_range_vf": null, "normalize_advantage": true, "target_kl": null, "lr_schedule": {":type:": "<class 'function'>", ":serialized:": "gAWV5QIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwOX21ha2VfZnVuY3Rpb26Uk5QoaACMDV9idWlsdGluX3R5cGWUk5SMCENvZGVUeXBllIWUUpQoSwFLAEsASwFLAUsTQwSIAFMAlE6FlCmMAV+UhZSMZS9ob21lL3NlbmxpLy5jb25kYS9lbnZzL3NlbmxpX1B5dGhvbjMuOC9saWIvcHl0aG9uMy44L3NpdGUtcGFja2FnZXMvc3RhYmxlX2Jhc2VsaW5lczMvY29tbW9uL3V0aWxzLnB5lIwEZnVuY5RLhEMCAAGUjAN2YWyUhZQpdJRSlH2UKIwLX19wYWNrYWdlX1+UjBhzdGFibGVfYmFzZWxpbmVzMy5jb21tb26UjAhfX25hbWVfX5SMHnN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi51dGlsc5SMCF9fZmlsZV9flIxlL2hvbWUvc2VubGkvLmNvbmRhL2VudnMvc2VubGlfUHl0aG9uMy44L2xpYi9weXRob24zLjgvc2l0ZS1wYWNrYWdlcy9zdGFibGVfYmFzZWxpbmVzMy9jb21tb24vdXRpbHMucHmUdU5OaACMEF9tYWtlX2VtcHR5X2NlbGyUk5QpUpSFlHSUUpRoAIwSX2Z1bmN0aW9uX3NldHN0YXRllJOUaB99lH2UKGgWjARmdW5jlIwMX19xdWFsbmFtZV9flIwZY29uc3RhbnRfZm4uPGxvY2Fscz4uZnVuY5SMD19fYW5ub3RhdGlvbnNfX5R9lIwOX19rd2RlZmF1bHRzX1+UTowMX19kZWZhdWx0c19flE6MCl9fbW9kdWxlX1+UaBeMB19fZG9jX1+UTowLX19jbG9zdXJlX1+UaACMCl9tYWtlX2NlbGyUk5RHPzOpKjBVMmGFlFKUhZSMF19jbG91ZHBpY2tsZV9zdWJtb2R1bGVzlF2UjAtfX2dsb2JhbHNfX5R9lHWGlIZSMC4="}, "system_info": {"OS": "Linux-5.15.0-134-generic-x86_64-with-glibc2.17 # 145-Ubuntu SMP Wed Feb 12 20:08:39 UTC 2025", "Python": "3.8.20", "Stable-Baselines3": "2.0.0a5", "PyTorch": "2.4.1+cu121", "GPU Enabled": "True", "Numpy": "1.24.4", "Cloudpickle": "3.1.1", "Gymnasium": "0.28.1", "OpenAI Gym": "0.26.2"}}
|
replay.mp4
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c1cdf4de5caaf37e651c053003d3c930f3e5c6eba35926a095d40b11f7af4196
|
| 3 |
+
size 169879
|
results.json
CHANGED
|
@@ -1 +1 @@
|
|
| 1 |
-
{"mean_reward": 296.
|
|
|
|
| 1 |
+
{"mean_reward": 296.80941640000003, "std_reward": 20.4820047455361, "is_deterministic": true, "n_eval_episodes": 10, "eval_datetime": "2025-04-11T04:11:51.634753"}
|