Upload PPO LunarLander-v2 LunarLander-v2_PPO_ne16_ns1024_b64_e4_cpu_TotalStep3000K.zip
Browse files- LunarLander-v2_PPO_ne16_ns1024_b64_e4_cpu_TotalStep3000K.zip +2 -2
- LunarLander-v2_PPO_ne16_ns1024_b64_e4_cpu_TotalStep3000K/data +16 -16
- LunarLander-v2_PPO_ne16_ns1024_b64_e4_cpu_TotalStep3000K/policy.optimizer.pth +1 -1
- LunarLander-v2_PPO_ne16_ns1024_b64_e4_cpu_TotalStep3000K/policy.pth +1 -1
- README.md +1 -1
- config.json +1 -1
- replay.mp4 +2 -2
- results.json +1 -1
LunarLander-v2_PPO_ne16_ns1024_b64_e4_cpu_TotalStep3000K.zip
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0f0c844a23a16cfaf139b559618d612049752c06c77af502cc978accf7408522
|
| 3 |
+
size 147523
|
LunarLander-v2_PPO_ne16_ns1024_b64_e4_cpu_TotalStep3000K/data
CHANGED
|
@@ -4,20 +4,20 @@
|
|
| 4 |
":serialized:": "gAWVOwAAAAAAAACMIXN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi5wb2xpY2llc5SMEUFjdG9yQ3JpdGljUG9saWN5lJOULg==",
|
| 5 |
"__module__": "stable_baselines3.common.policies",
|
| 6 |
"__doc__": "\n Policy class for actor-critic algorithms (has both policy and value prediction).\n Used by A2C, PPO and the likes.\n\n :param observation_space: Observation space\n :param action_space: Action space\n :param lr_schedule: Learning rate schedule (could be constant)\n :param net_arch: The specification of the policy and value networks.\n :param activation_fn: Activation function\n :param ortho_init: Whether to use or not orthogonal initialization\n :param use_sde: Whether to use State Dependent Exploration or not\n :param log_std_init: Initial value for the log standard deviation\n :param full_std: Whether to use (n_features x n_actions) parameters\n for the std instead of only (n_features,) when using gSDE\n :param use_expln: Use ``expln()`` function instead of ``exp()`` to ensure\n a positive standard deviation (cf paper). It allows to keep variance\n above zero and prevent it from growing too fast. In practice, ``exp()`` is usually enough.\n :param squash_output: Whether to squash the output using a tanh function,\n this allows to ensure boundaries when using gSDE.\n :param features_extractor_class: Features extractor to use.\n :param features_extractor_kwargs: Keyword arguments\n to pass to the features extractor.\n :param share_features_extractor: If True, the features extractor is shared between the policy and value networks.\n :param normalize_images: Whether to normalize images or not,\n dividing by 255.0 (True by default)\n :param optimizer_class: The optimizer to use,\n ``th.optim.Adam`` by default\n :param optimizer_kwargs: Additional keyword arguments,\n excluding the learning rate, to pass to the optimizer\n ",
|
| 7 |
-
"__init__": "<function ActorCriticPolicy.__init__ at
|
| 8 |
-
"_get_constructor_parameters": "<function ActorCriticPolicy._get_constructor_parameters at
|
| 9 |
-
"reset_noise": "<function ActorCriticPolicy.reset_noise at
|
| 10 |
-
"_build_mlp_extractor": "<function ActorCriticPolicy._build_mlp_extractor at
|
| 11 |
-
"_build": "<function ActorCriticPolicy._build at
|
| 12 |
-
"forward": "<function ActorCriticPolicy.forward at
|
| 13 |
-
"extract_features": "<function ActorCriticPolicy.extract_features at
|
| 14 |
-
"_get_action_dist_from_latent": "<function ActorCriticPolicy._get_action_dist_from_latent at
|
| 15 |
-
"_predict": "<function ActorCriticPolicy._predict at
|
| 16 |
-
"evaluate_actions": "<function ActorCriticPolicy.evaluate_actions at
|
| 17 |
-
"get_distribution": "<function ActorCriticPolicy.get_distribution at
|
| 18 |
-
"predict_values": "<function ActorCriticPolicy.predict_values at
|
| 19 |
"__abstractmethods__": "frozenset()",
|
| 20 |
-
"_abc_impl": "<_abc_data object at
|
| 21 |
},
|
| 22 |
"verbose": 0,
|
| 23 |
"policy_kwargs": {},
|
|
@@ -26,12 +26,12 @@
|
|
| 26 |
"_num_timesteps_at_start": 0,
|
| 27 |
"seed": null,
|
| 28 |
"action_noise": null,
|
| 29 |
-
"start_time":
|
| 30 |
"learning_rate": 0.0003,
|
| 31 |
"tensorboard_log": null,
|
| 32 |
"_last_obs": {
|
| 33 |
":type:": "<class 'numpy.ndarray'>",
|
| 34 |
-
":serialized:": "
|
| 35 |
},
|
| 36 |
"_last_episode_starts": {
|
| 37 |
":type:": "<class 'numpy.ndarray'>",
|
|
@@ -45,7 +45,7 @@
|
|
| 45 |
"_stats_window_size": 100,
|
| 46 |
"ep_info_buffer": {
|
| 47 |
":type:": "<class 'collections.deque'>",
|
| 48 |
-
":serialized:": "
|
| 49 |
},
|
| 50 |
"ep_success_buffer": {
|
| 51 |
":type:": "<class 'collections.deque'>",
|
|
|
|
| 4 |
":serialized:": "gAWVOwAAAAAAAACMIXN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi5wb2xpY2llc5SMEUFjdG9yQ3JpdGljUG9saWN5lJOULg==",
|
| 5 |
"__module__": "stable_baselines3.common.policies",
|
| 6 |
"__doc__": "\n Policy class for actor-critic algorithms (has both policy and value prediction).\n Used by A2C, PPO and the likes.\n\n :param observation_space: Observation space\n :param action_space: Action space\n :param lr_schedule: Learning rate schedule (could be constant)\n :param net_arch: The specification of the policy and value networks.\n :param activation_fn: Activation function\n :param ortho_init: Whether to use or not orthogonal initialization\n :param use_sde: Whether to use State Dependent Exploration or not\n :param log_std_init: Initial value for the log standard deviation\n :param full_std: Whether to use (n_features x n_actions) parameters\n for the std instead of only (n_features,) when using gSDE\n :param use_expln: Use ``expln()`` function instead of ``exp()`` to ensure\n a positive standard deviation (cf paper). It allows to keep variance\n above zero and prevent it from growing too fast. In practice, ``exp()`` is usually enough.\n :param squash_output: Whether to squash the output using a tanh function,\n this allows to ensure boundaries when using gSDE.\n :param features_extractor_class: Features extractor to use.\n :param features_extractor_kwargs: Keyword arguments\n to pass to the features extractor.\n :param share_features_extractor: If True, the features extractor is shared between the policy and value networks.\n :param normalize_images: Whether to normalize images or not,\n dividing by 255.0 (True by default)\n :param optimizer_class: The optimizer to use,\n ``th.optim.Adam`` by default\n :param optimizer_kwargs: Additional keyword arguments,\n excluding the learning rate, to pass to the optimizer\n ",
|
| 7 |
+
"__init__": "<function ActorCriticPolicy.__init__ at 0x7ff7606941f0>",
|
| 8 |
+
"_get_constructor_parameters": "<function ActorCriticPolicy._get_constructor_parameters at 0x7ff760694280>",
|
| 9 |
+
"reset_noise": "<function ActorCriticPolicy.reset_noise at 0x7ff760694310>",
|
| 10 |
+
"_build_mlp_extractor": "<function ActorCriticPolicy._build_mlp_extractor at 0x7ff7606943a0>",
|
| 11 |
+
"_build": "<function ActorCriticPolicy._build at 0x7ff760694430>",
|
| 12 |
+
"forward": "<function ActorCriticPolicy.forward at 0x7ff7606944c0>",
|
| 13 |
+
"extract_features": "<function ActorCriticPolicy.extract_features at 0x7ff760694550>",
|
| 14 |
+
"_get_action_dist_from_latent": "<function ActorCriticPolicy._get_action_dist_from_latent at 0x7ff7606945e0>",
|
| 15 |
+
"_predict": "<function ActorCriticPolicy._predict at 0x7ff760694670>",
|
| 16 |
+
"evaluate_actions": "<function ActorCriticPolicy.evaluate_actions at 0x7ff760694700>",
|
| 17 |
+
"get_distribution": "<function ActorCriticPolicy.get_distribution at 0x7ff760694790>",
|
| 18 |
+
"predict_values": "<function ActorCriticPolicy.predict_values at 0x7ff760694820>",
|
| 19 |
"__abstractmethods__": "frozenset()",
|
| 20 |
+
"_abc_impl": "<_abc_data object at 0x7ff7606f3d20>"
|
| 21 |
},
|
| 22 |
"verbose": 0,
|
| 23 |
"policy_kwargs": {},
|
|
|
|
| 26 |
"_num_timesteps_at_start": 0,
|
| 27 |
"seed": null,
|
| 28 |
"action_noise": null,
|
| 29 |
+
"start_time": 1744147674661249894,
|
| 30 |
"learning_rate": 0.0003,
|
| 31 |
"tensorboard_log": null,
|
| 32 |
"_last_obs": {
|
| 33 |
":type:": "<class 'numpy.ndarray'>",
|
| 34 |
+
":serialized:": "gAWVdQIAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJYAAgAAAAAAAAAgBjspYH26l9citaab/68WLSa6FUdXNAAAgD8AAIA/WmCbvWshgj6NO3A9S8Ovvsftzj1y4cy8AAAAAAAAAADmtU0+dBrNPuVK/rsL1s2+NB4UPpaqB70AAAAAAAAAAOZlRb3jXAM9piyCPTfqq75JosU9ukRnvAAAAAAAAAAAbZ0PPkT+nj4xdDW9Nou4vnrcDj6OeIk9AAAAAAAAAACzAW2+W0hpP0jJaz3QXRm/Fiy8voJBCz4AAAAAAAAAAE0yfT04APm7aq4bvgzREj1fVFM9KozwvQAAgD8AAIA/mq9vPtRvlT8y7SQ/+clDvyt/oT6o/CI+AAAAAAAAAACmu+I9OxYRP4Mu3Txv7ee+jAbEPD92p70AAAAAAAAAADNrFD1x3z8/fiQIPh/+E79355u8qsIsPgAAAAAAAAAAmlMbvUhDlbrdC/I4K9zlM0BDijqaKAy4AACAPwAAgD8AI+q8e96OuihSfjmRMGw0Qt61uvWek7gAAIA/AACAP8AZgj7PxNE+i9VXvpC6o74ojuc9Hr2vvQAAAAAAAAAAk/sOvu00ZT+jkFM9TEwLv7Lsqb4yG/M9AAAAAAAAAABg3kM+AYUDPjAY5b4H9RS+rROnPIg8270AAAAAAAAAAPapgj6fdyo/YRDGPYBwEr+8BFI+LSpEPAAAAAAAAAAAlIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUiYiHlFKUKEsDjAE8lE5OTkr/////Sv////9LAHSUYksQSwiGlIwBQ5R0lFKULg=="
|
| 35 |
},
|
| 36 |
"_last_episode_starts": {
|
| 37 |
":type:": "<class 'numpy.ndarray'>",
|
|
|
|
| 45 |
"_stats_window_size": 100,
|
| 46 |
"ep_info_buffer": {
|
| 47 |
":type:": "<class 'collections.deque'>",
|
| 48 |
+
":serialized:": "gAWV6AsAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKUKH2UKIwBcpRHQHM/eOXE61eMAWyUS9mMAXSUR0C016OejEehdX2UKGgGR0Bym6L4vexfaAdL2mgIR0C017PX05EMdX2UKGgGR0BwjK5tm+TNaAdLvGgIR0C01/b8rI5pdX2UKGgGR0BwsUF4cFQmaAdLymgIR0C02EkPMB6sdX2UKGgGR0ByFkFwDNhWaAdLz2gIR0C02F08mrsCdX2UKGgGR0BrzZ5/smfHaAdLuGgIR0C02Hm5tm+TdX2UKGgGR0BxobaBZpztaAdL5mgIR0C02Le14Pf9dX2UKGgGR0BvYXD+BH09aAdLwWgIR0C02NuTFERbdX2UKGgGR0Bv+aPjn3cpaAdL0mgIR0C02OGo73fydX2UKGgGR0BxtnEUCaJAaAdLvmgIR0C02Of/WDpUdX2UKGgGR0BwQ/B2wFC+aAdL5WgIR0C02OvIS13MdX2UKGgGR0BxuSMGX5WSaAdL32gIR0C02OnndO6/dX2UKGgGR0Bx+C2Dxsl+aAdL+mgIR0C02OuIRAbAdX2UKGgGR0Byo/dpItlJaAdL+2gIR0C02P1nIyTIdX2UKGgGR0Bxy9RR/EwWaAdLy2gIR0C02QMe8wpOdX2UKGgGR0BymuepXIU8aAdLxGgIR0C02Q+jEehgdX2UKGgGR0BzOo8EFGG3aAdL5GgIR0C02RBvaURndX2UKGgGR0BzESJwbVBlaAdL62gIR0C02R2lQ/HHdX2UKGgGR0Byv7Xf642CaAdLv2gIR0C02UURaouPdX2UKGgGR0BwKNItlI3BaAdLumgIR0C02YZezD4ydX2UKGgGR0BzTMvoNd7faAdLz2gIR0C02cAjt5UtdX2UKGgGR0Bx84yN4qwyaAdLs2gIR0C02dumR/3GdX2UKGgGR0Bwe67CiypraAdLq2gIR0C02i5VXFLndX2UKGgGR0BvRmZqmCRPaAdLsmgIR0C02iy1AqusdX2UKGgGR0BxlHeuV5bAaAdLvWgIR0C03iXsC1Z1dX2UKGgGR0Bxq1IsiB5HaAdLx2gIR0C03iTTvy9VdX2UKGgGR0By2mi35N48aAdL+GgIR0C03iaBNEgGdX2UKGgGR0Bw/kQtjCpFaAdLzGgIR0C03ioTK1XvdX2UKGgGR0BykGzD4xk/aAdL0GgIR0C03ja6J66bdX2UKGgGR0BwghEb5uZUaAdLvWgIR0C03lK19fCzdX2UKGgGR0Byv5hH9WIXaAdL8WgIR0C03mbZFocrdX2UKGgGR0BxxfA31jAjaAdL7mgIR0C03m4bsF+vdX2UKGgGR0Bx6vj3mFJyaAdNAgFoCEdAtN6R0knkUHV9lChoBkdAcfdqC6H0smgHS8poCEdAtN6oaMrEtXV9lChoBkdAceKdDpkf92gHS7loCEdAtN7RjriVB3V9lChoBkdAc2SxH5Jsf2gHTRQBaAhHQLTe0olUp/h1fZQoaAZHQHFpz50r9VFoB0uyaAhHQLTe9N9H+ZR1fZQoaAZHQHD+yeiBXjloB0uxaAhHQLTfVBLPD511fZQoaAZHQHKWcG9pRGdoB0vvaAhHQLTffjriVB51fZQoaAZHQHNwF4oqkM1oB0u+aAhHQLTfguR9w3p1fZQoaAZHQHKhiAUcn3NoB0vQaAhHQLTflPX05EN1fZQoaAZHQHF+1NL127poB0vgaAhHQLTfrQj2SMd1fZQoaAZHQHMGX+6y0KJoB0vZaAhHQLTfsD6WPcV1fZQoaAZHQHLPGjfvWpZoB0voaAhHQLTfwRsuWbB1fZQoaAZHQHAUUcsDnvFoB0vMaAhHQLTfxVE/jbV1fZQoaAZHQHJmvH5rP+poB0v2aAhHQLTf6/82rGR1fZQoaAZHQEsZegte2NNoB0usaAhHQLTgKi0OVgR1fZQoaAZHQHELcLncL0BoB0vpaAhHQLTgLGtITXd1fZQoaAZHQHGtMl1KXfJoB0v0aAhHQLTgPIsRQJp1fZQoaAZHQHAxvJaJQ+FoB0vEaAhHQLTgX1A7gbZ1fZQoaAZHQHMDr/CIk7hoB0v1aAhHQLTgcDtw71Z1fZQoaAZHQHCY8HKOktVoB0u9aAhHQLTggS75Ec91fZQoaAZHQHCMy925hBtoB0vzaAhHQLTgiM9r4351fZQoaAZHQHLY5Etuk1xoB0vLaAhHQLThTcXm/351fZQoaAZHQHHnmZ/kNnZoB0uxaAhHQLThUwYLsrx1fZQoaAZHQHM6GAkLQX1oB0uyaAhHQLThWsO5J9R1fZQoaAZHQHFbo6CDmKZoB0vbaAhHQLTheciW3Sd1fZQoaAZHQHGnG+PBBRhoB0vXaAhHQLThhAnlXBB1fZQoaAZHQHExVAzHjp9oB0v+aAhHQLThl6IFeOZ1fZQoaAZHQG6tf0NBnjBoB0vIaAhHQLThue3hGYt1fZQoaAZHQHJ5nWSU1Q9oB0vyaAhHQLTh2pTMqz91fZQoaAZHQHBt9tdiUgVoB0veaAhHQLTiLKIi1Rd1fZQoaAZHQHGsxfKISDhoB00ZAWgIR0C04jef29L6dX2UKGgGR0BxBbDjzZpSaAdLw2gIR0C04jvdAPd3dX2UKGgGR0BwcYJlar3kaAdL8GgIR0C04lm3rleXdX2UKGgGR0ByAFrBTGYKaAdL2mgIR0C04l5EUj9odX2UKGgGR0BzdKbWmP5paAdL9GgIR0C04nED+zdDdX2UKGgGR0ByVdc9nscAaAdL3mgIR0C04oUbPyCndX2UKGgGR0BwqbOiWVu8aAdNEgFoCEdAtOL7lp48l3V9lChoBkdAc+s+lj3Eh2gHS95oCEdAtONPsu3+dnV9lChoBkdAc1D4QBgeBGgHS9NoCEdAtONgNI9TxXV9lChoBkdAcetLA57w8WgHS81oCEdAtONwCq6vq3V9lChoBkdAcQ+haC+UQmgHS8BoCEdAtONy+N96TnV9lChoBkdAcjOvFFUhm2gHTQEBaAhHQLTjnv/BFd91fZQoaAZHQHD63bAUL2JoB0v3aAhHQLTjtVh1DBx1fZQoaAZHQHDTX/T9bX9oB0vAaAhHQLTj6GbkOqh1fZQoaAZHQHJu0g8r7O5oB00jAWgIR0C04/R42S+ydX2UKGgGR0Bzhs5uIhyKaAdLyGgIR0C05DJbMX7+dX2UKGgGR0BxqgleF+NMaAdL02gIR0C05Dd/z8P4dX2UKGgGR0BweI3qAz55aAdL2WgIR0C05EAsGxD9dX2UKGgGR0BwwWTr3TNMaAdL82gIR0C05FA1vVEvdX2UKGgGR0Bx3g8jiXIEaAdL9GgIR0C05Fq7ZnL8dX2UKGgGR0BxVuDQJHAiaAdNIQFoCEdAtORprXUYsXV9lChoBkdAcznFXaJyhmgHS9FoCEdAtOTUtqYZ23V9lChoBkdAcTCZXuE252gHS7JoCEdAtOTZMlC1JHV9lChoBkdAcT+a6jFhomgHS8RoCEdAtOUed4FA3XV9lChoBkdAcUFnIQvpQmgHS9JoCEdAtOUp1W8yvnV9lChoBkdAbs/njABT42gHS7loCEdAtOVMDzRQanV9lChoBkdAb8z3lCCz1WgHS95oCEdAtOVU6ySmqHV9lChoBkdAccnLHdXT3WgHS+9oCEdAtOW0QZn+Q3V9lChoBkdAb8qKziS7oWgHS+JoCEdAtOXhk+X7cnV9lChoBkdAcW8DSPU8WGgHS71oCEdAtOXkhNdqtnV9lChoBkdAcmmOT7l7t2gHS8RoCEdAtOXnStvGZXV9lChoBkdAclUunuRcNmgHS+9oCEdAtOYK2uxKQXV9lChoBkdAcPlTMJQcgmgHS81oCEdAtOYZQVKwp3V9lChoBkdAcGsAq/dqL2gHS8VoCEdAtOYii+L3sXV9lChoBkdAcuLcdYGMXWgHTaEBaAhHQLTmLXAM2FZ1fZQoaAZHQHMFCS7oSthoB0vwaAhHQLTmSTfR/mV1fZQoaAZHQEmdn6l+EytoB0uraAhHQLTmVRiPQv91fZQoaAZHQHJ1SXlbNbFoB0vzaAhHQLTmaQ7tAs11ZS4="
|
| 49 |
},
|
| 50 |
"ep_success_buffer": {
|
| 51 |
":type:": "<class 'collections.deque'>",
|
LunarLander-v2_PPO_ne16_ns1024_b64_e4_cpu_TotalStep3000K/policy.optimizer.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 87978
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0fcfd4b38424c07dea9e8dc0dac7ea0ba539431a637bc79c2e48fd2f5f19fe1d
|
| 3 |
size 87978
|
LunarLander-v2_PPO_ne16_ns1024_b64_e4_cpu_TotalStep3000K/policy.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 43634
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c37ac510cf27ed5a45d3e2189fab8f191cc10389c1fd8250ebc08007b3471565
|
| 3 |
size 43634
|
README.md
CHANGED
|
@@ -16,7 +16,7 @@ model-index:
|
|
| 16 |
type: LunarLander-v2
|
| 17 |
metrics:
|
| 18 |
- type: mean_reward
|
| 19 |
-
value:
|
| 20 |
name: mean_reward
|
| 21 |
verified: false
|
| 22 |
---
|
|
|
|
| 16 |
type: LunarLander-v2
|
| 17 |
metrics:
|
| 18 |
- type: mean_reward
|
| 19 |
+
value: 295.40 +/- 14.38
|
| 20 |
name: mean_reward
|
| 21 |
verified: false
|
| 22 |
---
|
config.json
CHANGED
|
@@ -1 +1 @@
|
|
| 1 |
-
{"policy_class": {":type:": "<class 'abc.ABCMeta'>", ":serialized:": "gAWVOwAAAAAAAACMIXN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi5wb2xpY2llc5SMEUFjdG9yQ3JpdGljUG9saWN5lJOULg==", "__module__": "stable_baselines3.common.policies", "__doc__": "\n Policy class for actor-critic algorithms (has both policy and value prediction).\n Used by A2C, PPO and the likes.\n\n :param observation_space: Observation space\n :param action_space: Action space\n :param lr_schedule: Learning rate schedule (could be constant)\n :param net_arch: The specification of the policy and value networks.\n :param activation_fn: Activation function\n :param ortho_init: Whether to use or not orthogonal initialization\n :param use_sde: Whether to use State Dependent Exploration or not\n :param log_std_init: Initial value for the log standard deviation\n :param full_std: Whether to use (n_features x n_actions) parameters\n for the std instead of only (n_features,) when using gSDE\n :param use_expln: Use ``expln()`` function instead of ``exp()`` to ensure\n a positive standard deviation (cf paper). It allows to keep variance\n above zero and prevent it from growing too fast. In practice, ``exp()`` is usually enough.\n :param squash_output: Whether to squash the output using a tanh function,\n this allows to ensure boundaries when using gSDE.\n :param features_extractor_class: Features extractor to use.\n :param features_extractor_kwargs: Keyword arguments\n to pass to the features extractor.\n :param share_features_extractor: If True, the features extractor is shared between the policy and value networks.\n :param normalize_images: Whether to normalize images or not,\n dividing by 255.0 (True by default)\n :param optimizer_class: The optimizer to use,\n ``th.optim.Adam`` by default\n :param optimizer_kwargs: Additional keyword arguments,\n excluding the learning rate, to pass to the optimizer\n ", "__init__": "<function ActorCriticPolicy.__init__ at 0x7f2803f51160>", "_get_constructor_parameters": "<function ActorCriticPolicy._get_constructor_parameters at 0x7f2803f511f0>", "reset_noise": "<function ActorCriticPolicy.reset_noise at 0x7f2803f51280>", "_build_mlp_extractor": "<function ActorCriticPolicy._build_mlp_extractor at 0x7f2803f51310>", "_build": "<function ActorCriticPolicy._build at 0x7f2803f513a0>", "forward": "<function ActorCriticPolicy.forward at 0x7f2803f51430>", "extract_features": "<function ActorCriticPolicy.extract_features at 0x7f2803f514c0>", "_get_action_dist_from_latent": "<function ActorCriticPolicy._get_action_dist_from_latent at 0x7f2803f51550>", "_predict": "<function ActorCriticPolicy._predict at 0x7f2803f515e0>", "evaluate_actions": "<function ActorCriticPolicy.evaluate_actions at 0x7f2803f51670>", "get_distribution": "<function ActorCriticPolicy.get_distribution at 0x7f2803f51700>", "predict_values": "<function ActorCriticPolicy.predict_values at 0x7f2803f51790>", "__abstractmethods__": "frozenset()", "_abc_impl": "<_abc_data object at 0x7f2803fb2d20>"}, "verbose": 0, "policy_kwargs": {}, "num_timesteps": 3014656, "_total_timesteps": 3000000, "_num_timesteps_at_start": 0, "seed": null, "action_noise": null, "start_time": 1744147461375332633, "learning_rate": 0.0003, "tensorboard_log": null, "_last_obs": {":type:": "<class 'numpy.ndarray'>", ":serialized:": "gAWVdQIAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJYAAgAAAAAAAP2LpT5VPn4/sq4CP64QRb8hs9s+/do1PgAAAAAAAAAAs3JcPUNXJbxT+1e9qR9SvLaIqb2SgS29AACAPwAAgD9mPQS9roOuuvLUsTuBSnk8VV2ku5U9Wj0AAIA/AACAP4BXur0CKS4/1fIvPoswA7/cyfG9sYYtPgAAAAAAAAAATRQmvWAvCz/H7JA+vkT6vtO8zj27mgM+AAAAAAAAAAAz0oo8hXb/PDtVvjzqCK2+xs2VPY7scTsAAAAAAAAAAI182L0PYBM/MKtrPjqU8b7INza9fKs/PgAAAAAAAAAAZkU2vj41lz9SOA2/SVT4vkZPnb5t5sK+AAAAAAAAAAAALik8XMdcum5Phbm8m160Iu8Ku1KsnDgAAIA/AACAP2Zx/TxD6mS8eKyhvpfDoL2B4qs9vsSiOwAAgD8AAIA/AKA1uq6Zs7oA5S84bxAsMyowaTjft0i3AACAPwAAgD8Abbi9dBhVPjmpuD5xkdy+vsEfPvWNVz0AAAAAAAAAAJqZHDp7Qqm6q1v5tLxeAq8JZnC6jeJgNAAAgD8AAIA/M1U1PcPpV7qr43gz/pZprnbQj7sGGqazAACAPwAAgD/Aefc9gfyAPg3MaL6nPL++HHarPaqAzb0AAAAAAAAAAABa+jyu+YC68oPEMulx6zAR8ko6QDeQswAAgD8AAIA/lIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUiYiHlFKUKEsDjAE8lE5OTkr/////Sv////9LAHSUYksQSwiGlIwBQ5R0lFKULg=="}, "_last_episode_starts": {":type:": "<class 'numpy.ndarray'>", ":serialized:": "gAWVgwAAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJYQAAAAAAAAAAAAAAAAAAAAAAAAAAAAAACUjAVudW1weZSMBWR0eXBllJOUjAJiMZSJiIeUUpQoSwOMAXyUTk5OSv////9K/////0sAdJRiSxCFlIwBQ5R0lFKULg=="}, "_last_original_obs": null, "_episode_num": 0, "use_sde": false, "sde_sample_freq": -1, "_current_progress_remaining": -0.004885333333333408, "_stats_window_size": 100, "ep_info_buffer": {":type:": "<class 'collections.deque'>", ":serialized:": "gAWV4gsAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKUKH2UKIwBcpRHQG8HNcGC7K+MAWyUS8uMAXSUR0C3R/2gi/widX2UKGgGR0Bxdmmm+CbuaAdLzmgIR0C3SAR1PnB+dX2UKGgGR0ByYv3M6ij+aAdLx2gIR0C3SAm25QP7dX2UKGgGR0BweQTQE6kqaAdLwGgIR0C3SDouGsV+dX2UKGgGR0ByQ1IDoyKvaAdL3GgIR0C3SF1C1JDmdX2UKGgGR0BzO3QpnYg8aAdL1WgIR0C3SGYMSbpedX2UKGgGR0BwHrufEn9faAdLwmgIR0C3SH/ES/TLdX2UKGgGR0ByAaKiwjdIaAdLsWgIR0C3SIRPoFFEdX2UKGgGR0Bz+63I+4b0aAdL2GgIR0C3SLcjAzpHdX2UKGgGR0Bz0eRDCxeLaAdL12gIR0C3SLtpZfUndX2UKGgGR0BwzklSjxkNaAdL6mgIR0C3SMJTqB3BdX2UKGgGR0By2/Uz9CNTaAdL12gIR0C3SNhY3eendX2UKGgGR0ByvggU1yeaaAdL1WgIR0C3SNnEQ5FPdX2UKGgGR0BxCzZRKpT/aAdL1GgIR0C3SNr/jsD5dX2UKGgGR0Bx7av5gw49aAdLt2gIR0C3SQFF2FFldX2UKGgGR0ByGcLRa5f/aAdNJQFoCEdAt0kuifxtpHV9lChoBkdAcknRW912aGgHS8FoCEdAt0k5Ew35vnV9lChoBkdAboSQTVUdaWgHS8doCEdAt0k5PWQOnXV9lChoBkdAcN+hBqsU7GgHS9RoCEdAt0lZO/L1VnV9lChoBkdAdB1eY2Kl6GgHS8xoCEdAt0mFZIQOF3V9lChoBkdAcNY91EE1VGgHS79oCEdAt0mbHhjvu3V9lChoBkdAcS4f/3nIQ2gHS8RoCEdAt0m4m4RVZXV9lChoBkdAcO5kdmxt52gHS8toCEdAt0nI1Muez3V9lChoBkdAchz8HObAlGgHS+VoCEdAt0nNQO4G2XV9lChoBkdAcjFYLsrupmgHS8BoCEdAt01NtpEhJXV9lChoBkdAceLgZCOWB2gHS8NoCEdAt01pL8Jla3V9lChoBkdAcaZd2gWadGgHS8poCEdAt011xkupTHV9lChoBkdAcbOCnP3SKGgHS+BoCEdAt015sj3VTnV9lChoBkdAcaeg+QlrumgHS+hoCEdAt02B/tpmE3V9lChoBkdAccSKUmlZYGgHS9doCEdAt02G+rU9ZHV9lChoBkdAcn6fxtpEhWgHS9hoCEdAt02jUAksz3V9lChoBkdAcGAvalDWsmgHS8BoCEdAt02sny/bkHV9lChoBkdAcIENkvsZ52gHS7NoCEdAt024HJLdvnV9lChoBkdAceoDPGACn2gHS8xoCEdAt027X05EMXV9lChoBkdAccMcyWRigGgHS9toCEdAt03GdPLxJHV9lChoBkdAcUqI3zcynGgHS9VoCEdAt04D8Kohp3V9lChoBkdAcRJjH4oJA2gHS81oCEdAt04SL74zrXV9lChoBkdAcehG2CuloGgHS89oCEdAt040LCvX9XV9lChoBkdAcCm+9Jz1b2gHS8doCEdAt048evIOpnV9lChoBkdAcLx3SKFZgWgHS89oCEdAt05DRIBikXV9lChoBkdAcA8Hv+fh/GgHS8doCEdAt05ecjJMg3V9lChoBkdAco5u3trsSmgHS8NoCEdAt06EgPmPo3V9lChoBkdAcTn/c32mHmgHS8RoCEdAt06bCm/Fi3V9lChoBkdAcdW8UmD15GgHS9loCEdAt06fdpItlXV9lChoBkdAdCELUCq6v2gHS+BoCEdAt06+938n/nV9lChoBkdAc0RHT7VJ+WgHS75oCEdAt07SDUVi4XV9lChoBkdAcqtuKGcnV2gHS+JoCEdAt07UKb8WK3V9lChoBkdAcUwmBe5WimgHS9FoCEdAt07plbu+iHV9lChoBkdAcTh04R28qWgHS8loCEdAt076BJ7LMnV9lChoBkdAb/3m8M/hVGgHS8FoCEdAt08D8HfMwHV9lChoBkdAcLIGBWgezWgHS9toCEdAt08fdxhlUnV9lChoBkdAcJlU7jkuH2gHS8RoCEdAt09gXJo0ynV9lChoBkdAcFPD0163RWgHS7poCEdAt0+Cf+S8rnV9lChoBkdAccFgtOEdvWgHS9VoCEdAt0/DLQokRnV9lChoBkdAc99ms/6frmgHS9BoCEdAt0/DKFIuoXV9lChoBkdAcjcW3BpHqmgHS/xoCEdAt0/cDEFW4nV9lChoBkdAcjT/Vy3kP2gHS9doCEdAt0/3SSeRP3V9lChoBkdAcK5+nqFAV2gHS9FoCEdAt1AzgJkXlHV9lChoBkdAcWDymhufmWgHS+VoCEdAt1BDyRSxaHV9lChoBkdAcRX8UmD15GgHS99oCEdAt1BNvybx3HV9lChoBkdAcYEvLX+VDGgHS+BoCEdAt1Bzfl6qsHV9lChoBkdAcQSIoVmBfGgHS9hoCEdAt1B3SeAd4nV9lChoBkdAcBOXFLnLaGgHS89oCEdAt1CNN8E3bXV9lChoBkdAcnjTLW7OFGgHS+ZoCEdAt1CnvsqrinV9lChoBkdAc1KcNYr8SGgHS/ZoCEdAt1CwMBp5/3V9lChoBkdAbmRVJcxCY2gHS89oCEdAt1CzhvR7Z3V9lChoBkdAcLTdJJ5E+mgHS8JoCEdAt1EL6ZYxL3V9lChoBkdAcofYgq3EymgHS9toCEdAt1EVM495hXV9lChoBkdAcaR4ecQRPGgHS7ZoCEdAt1Fy+QEIPnV9lChoBkdAcQS7eEZiu2gHS9NoCEdAt1F8pPRAr3V9lChoBkdAcakN0NjLCGgHS9doCEdAt1GFPykKu3V9lChoBkdAcxvFR51Ng2gHS85oCEdAt1GNm4Ajp3V9lChoBkdAc9KZ6lchT2gHS8doCEdAt1HvA0sOG3V9lChoBkdAc6Soa1kUbmgHS89oCEdAt1HwAIY3vXV9lChoBkdAcAmWSEDhcmgHTXIBaAhHQLdR9di2Dxt1fZQoaAZHQG+vearmyPdoB0vGaAhHQLdR9oGIKtx1fZQoaAZHQG/vw71ZkkNoB0vHaAhHQLdSFXOW0JF1fZQoaAZHQHI61uivgWJoB0vLaAhHQLdSLNqQA+91fZQoaAZHQHRDsU/OdG1oB0vIaAhHQLdSQh8pkPN1fZQoaAZHQHLT2xD9fkZoB0vGaAhHQLdSSP3i7051fZQoaAZHQHMhtYwIt19oB0v0aAhHQLdSXYpDu0F1fZQoaAZHQHGFefEn9ehoB0vlaAhHQLdScy/sVtZ1fZQoaAZHQHHjn4bjtHBoB0u/aAhHQLdSg+cYqG11fZQoaAZHQHIw2XokiUxoB0v8aAhHQLdS051Ng0F1fZQoaAZHQHHRgrxy4nZoB0vQaAhHQLdS1j3mFJx1fZQoaAZHQHQyBw++ueVoB0vPaAhHQLdS2v+wTuh1fZQoaAZHQHM3N4eLehxoB0vQaAhHQLdS6Z+x4Y91fZQoaAZHQHGKxSP2f05oB0vaaAhHQLdS8lUZNwl1fZQoaAZHQG7NZb6guh9oB0u+aAhHQLdTEKqXF991fZQoaAZHQHEuCi/O+qRoB0vHaAhHQLdTH1zQu291fZQoaAZHQHF8W8h9srNoB0vTaAhHQLdTLMd92HN1fZQoaAZHQHML3hS9/SZoB0vYaAhHQLdTNfzz3AV1fZQoaAZHQG/BGDUVi4JoB0vQaAhHQLdTT36AOKB1fZQoaAZHQHEAydnTRY1oB0vGaAhHQLdTWzJ6po91fZQoaAZHQHGKemBOHnFoB0vPaAhHQLdTiMBZIQR1fZQoaAZHQHFrhtYSxqxoB0vUaAhHQLdTieyAxzt1fZQoaAZHQHLhYubqhURoB0vQaAhHQLdTp0O3DvV1fZQoaAZHQHMUumm+CbtoB0vRaAhHQLdTwx0uDjB1fZQoaAZHQHKAn9rGipNoB0vhaAhHQLdT9vUjLSx1fZQoaAZHQHJOGXHBDXxoB0u5aAhHQLdUGP1tfol1ZS4="}, "ep_success_buffer": {":type:": "<class 'collections.deque'>", ":serialized:": "gAWVIAAAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKULg=="}, "_n_updates": 1476, "observation_space": {":type:": "<class 'gymnasium.spaces.box.Box'>", ":serialized:": "gAWVcAIAAAAAAACMFGd5bW5hc2l1bS5zcGFjZXMuYm94lIwDQm94lJOUKYGUfZQojAVkdHlwZZSMBW51bXB5lGgFk5SMAmY0lImIh5RSlChLA4wBPJROTk5K/////0r/////SwB0lGKMDWJvdW5kZWRfYmVsb3eUjBJudW1weS5jb3JlLm51bWVyaWOUjAtfZnJvbWJ1ZmZlcpSTlCiWCAAAAAAAAAABAQEBAQEBAZRoB4wCYjGUiYiHlFKUKEsDjAF8lE5OTkr/////Sv////9LAHSUYksIhZSMAUOUdJRSlIwNYm91bmRlZF9hYm92ZZRoECiWCAAAAAAAAAABAQEBAQEBAZRoFEsIhZRoGHSUUpSMBl9zaGFwZZRLCIWUjANsb3eUaBAoliAAAAAAAAAAAAC0wgAAtMIAAKDAAACgwNsPScAAAKDAAAAAgAAAAICUaApLCIWUaBh0lFKUjARoaWdolGgQKJYgAAAAAAAAAAAAtEIAALRCAACgQAAAoEDbD0lAAACgQAAAgD8AAIA/lGgKSwiFlGgYdJRSlIwIbG93X3JlcHKUjFtbLTkwLiAgICAgICAgLTkwLiAgICAgICAgIC01LiAgICAgICAgIC01LiAgICAgICAgIC0zLjE0MTU5MjcgIC01LgogIC0wLiAgICAgICAgIC0wLiAgICAgICBdlIwJaGlnaF9yZXBylIxTWzkwLiAgICAgICAgOTAuICAgICAgICAgNS4gICAgICAgICA1LiAgICAgICAgIDMuMTQxNTkyNyAgNS4KICAxLiAgICAgICAgIDEuICAgICAgIF2UjApfbnBfcmFuZG9tlE51Yi4=", "dtype": "float32", "bounded_below": "[ True True True True True True True True]", "bounded_above": "[ True True True True True True True True]", "_shape": [8], "low": "[-90. -90. -5. -5. -3.1415927 -5.\n -0. -0. ]", "high": "[90. 90. 5. 5. 3.1415927 5.\n 1. 1. ]", "low_repr": "[-90. -90. -5. -5. -3.1415927 -5.\n -0. -0. ]", "high_repr": "[90. 90. 5. 5. 3.1415927 5.\n 1. 1. ]", "_np_random": null}, "action_space": {":type:": "<class 'gymnasium.spaces.discrete.Discrete'>", ":serialized:": "gAWV1QAAAAAAAACMGWd5bW5hc2l1bS5zcGFjZXMuZGlzY3JldGWUjAhEaXNjcmV0ZZSTlCmBlH2UKIwBbpSMFW51bXB5LmNvcmUubXVsdGlhcnJheZSMBnNjYWxhcpSTlIwFbnVtcHmUjAVkdHlwZZSTlIwCaTiUiYiHlFKUKEsDjAE8lE5OTkr/////Sv////9LAHSUYkMIBAAAAAAAAACUhpRSlIwFc3RhcnSUaAhoDkMIAAAAAAAAAACUhpRSlIwGX3NoYXBllCloCmgOjApfbnBfcmFuZG9tlE51Yi4=", "n": "4", "start": "0", "_shape": [], "dtype": "int64", "_np_random": null}, "n_envs": 16, "n_steps": 1024, "gamma": 0.999, "gae_lambda": 0.98, "ent_coef": 0.01, "vf_coef": 0.5, "max_grad_norm": 0.5, "batch_size": 64, "n_epochs": 4, "clip_range": {":type:": "<class 'function'>", ":serialized:": "gAWV5QIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwOX21ha2VfZnVuY3Rpb26Uk5QoaACMDV9idWlsdGluX3R5cGWUk5SMCENvZGVUeXBllIWUUpQoSwFLAEsASwFLAUsTQwSIAFMAlE6FlCmMAV+UhZSMZS9ob21lL3NlbmxpLy5jb25kYS9lbnZzL3NlbmxpX1B5dGhvbjMuOC9saWIvcHl0aG9uMy44L3NpdGUtcGFja2FnZXMvc3RhYmxlX2Jhc2VsaW5lczMvY29tbW9uL3V0aWxzLnB5lIwEZnVuY5RLhEMCAAGUjAN2YWyUhZQpdJRSlH2UKIwLX19wYWNrYWdlX1+UjBhzdGFibGVfYmFzZWxpbmVzMy5jb21tb26UjAhfX25hbWVfX5SMHnN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi51dGlsc5SMCF9fZmlsZV9flIxlL2hvbWUvc2VubGkvLmNvbmRhL2VudnMvc2VubGlfUHl0aG9uMy44L2xpYi9weXRob24zLjgvc2l0ZS1wYWNrYWdlcy9zdGFibGVfYmFzZWxpbmVzMy9jb21tb24vdXRpbHMucHmUdU5OaACMEF9tYWtlX2VtcHR5X2NlbGyUk5QpUpSFlHSUUpRoAIwSX2Z1bmN0aW9uX3NldHN0YXRllJOUaB99lH2UKGgWjARmdW5jlIwMX19xdWFsbmFtZV9flIwZY29uc3RhbnRfZm4uPGxvY2Fscz4uZnVuY5SMD19fYW5ub3RhdGlvbnNfX5R9lIwOX19rd2RlZmF1bHRzX1+UTowMX19kZWZhdWx0c19flE6MCl9fbW9kdWxlX1+UaBeMB19fZG9jX1+UTowLX19jbG9zdXJlX1+UaACMCl9tYWtlX2NlbGyUk5RHP8mZmZmZmZqFlFKUhZSMF19jbG91ZHBpY2tsZV9zdWJtb2R1bGVzlF2UjAtfX2dsb2JhbHNfX5R9lHWGlIZSMC4="}, "clip_range_vf": null, "normalize_advantage": true, "target_kl": null, "lr_schedule": {":type:": "<class 'function'>", ":serialized:": "gAWV5QIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwOX21ha2VfZnVuY3Rpb26Uk5QoaACMDV9idWlsdGluX3R5cGWUk5SMCENvZGVUeXBllIWUUpQoSwFLAEsASwFLAUsTQwSIAFMAlE6FlCmMAV+UhZSMZS9ob21lL3NlbmxpLy5jb25kYS9lbnZzL3NlbmxpX1B5dGhvbjMuOC9saWIvcHl0aG9uMy44L3NpdGUtcGFja2FnZXMvc3RhYmxlX2Jhc2VsaW5lczMvY29tbW9uL3V0aWxzLnB5lIwEZnVuY5RLhEMCAAGUjAN2YWyUhZQpdJRSlH2UKIwLX19wYWNrYWdlX1+UjBhzdGFibGVfYmFzZWxpbmVzMy5jb21tb26UjAhfX25hbWVfX5SMHnN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi51dGlsc5SMCF9fZmlsZV9flIxlL2hvbWUvc2VubGkvLmNvbmRhL2VudnMvc2VubGlfUHl0aG9uMy44L2xpYi9weXRob24zLjgvc2l0ZS1wYWNrYWdlcy9zdGFibGVfYmFzZWxpbmVzMy9jb21tb24vdXRpbHMucHmUdU5OaACMEF9tYWtlX2VtcHR5X2NlbGyUk5QpUpSFlHSUUpRoAIwSX2Z1bmN0aW9uX3NldHN0YXRllJOUaB99lH2UKGgWjARmdW5jlIwMX19xdWFsbmFtZV9flIwZY29uc3RhbnRfZm4uPGxvY2Fscz4uZnVuY5SMD19fYW5ub3RhdGlvbnNfX5R9lIwOX19rd2RlZmF1bHRzX1+UTowMX19kZWZhdWx0c19flE6MCl9fbW9kdWxlX1+UaBeMB19fZG9jX1+UTowLX19jbG9zdXJlX1+UaACMCl9tYWtlX2NlbGyUk5RHPzOpKjBVMmGFlFKUhZSMF19jbG91ZHBpY2tsZV9zdWJtb2R1bGVzlF2UjAtfX2dsb2JhbHNfX5R9lHWGlIZSMC4="}, "system_info": {"OS": "Linux-5.15.0-134-generic-x86_64-with-glibc2.17 # 145-Ubuntu SMP Wed Feb 12 20:08:39 UTC 2025", "Python": "3.8.20", "Stable-Baselines3": "2.0.0a5", "PyTorch": "2.4.1+cu121", "GPU Enabled": "True", "Numpy": "1.24.4", "Cloudpickle": "3.1.1", "Gymnasium": "0.28.1", "OpenAI Gym": "0.26.2"}}
|
|
|
|
| 1 |
+
{"policy_class": {":type:": "<class 'abc.ABCMeta'>", ":serialized:": "gAWVOwAAAAAAAACMIXN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi5wb2xpY2llc5SMEUFjdG9yQ3JpdGljUG9saWN5lJOULg==", "__module__": "stable_baselines3.common.policies", "__doc__": "\n Policy class for actor-critic algorithms (has both policy and value prediction).\n Used by A2C, PPO and the likes.\n\n :param observation_space: Observation space\n :param action_space: Action space\n :param lr_schedule: Learning rate schedule (could be constant)\n :param net_arch: The specification of the policy and value networks.\n :param activation_fn: Activation function\n :param ortho_init: Whether to use or not orthogonal initialization\n :param use_sde: Whether to use State Dependent Exploration or not\n :param log_std_init: Initial value for the log standard deviation\n :param full_std: Whether to use (n_features x n_actions) parameters\n for the std instead of only (n_features,) when using gSDE\n :param use_expln: Use ``expln()`` function instead of ``exp()`` to ensure\n a positive standard deviation (cf paper). It allows to keep variance\n above zero and prevent it from growing too fast. In practice, ``exp()`` is usually enough.\n :param squash_output: Whether to squash the output using a tanh function,\n this allows to ensure boundaries when using gSDE.\n :param features_extractor_class: Features extractor to use.\n :param features_extractor_kwargs: Keyword arguments\n to pass to the features extractor.\n :param share_features_extractor: If True, the features extractor is shared between the policy and value networks.\n :param normalize_images: Whether to normalize images or not,\n dividing by 255.0 (True by default)\n :param optimizer_class: The optimizer to use,\n ``th.optim.Adam`` by default\n :param optimizer_kwargs: Additional keyword arguments,\n excluding the learning rate, to pass to the optimizer\n ", "__init__": "<function ActorCriticPolicy.__init__ at 0x7ff7606941f0>", "_get_constructor_parameters": "<function ActorCriticPolicy._get_constructor_parameters at 0x7ff760694280>", "reset_noise": "<function ActorCriticPolicy.reset_noise at 0x7ff760694310>", "_build_mlp_extractor": "<function ActorCriticPolicy._build_mlp_extractor at 0x7ff7606943a0>", "_build": "<function ActorCriticPolicy._build at 0x7ff760694430>", "forward": "<function ActorCriticPolicy.forward at 0x7ff7606944c0>", "extract_features": "<function ActorCriticPolicy.extract_features at 0x7ff760694550>", "_get_action_dist_from_latent": "<function ActorCriticPolicy._get_action_dist_from_latent at 0x7ff7606945e0>", "_predict": "<function ActorCriticPolicy._predict at 0x7ff760694670>", "evaluate_actions": "<function ActorCriticPolicy.evaluate_actions at 0x7ff760694700>", "get_distribution": "<function ActorCriticPolicy.get_distribution at 0x7ff760694790>", "predict_values": "<function ActorCriticPolicy.predict_values at 0x7ff760694820>", "__abstractmethods__": "frozenset()", "_abc_impl": "<_abc_data object at 0x7ff7606f3d20>"}, "verbose": 0, "policy_kwargs": {}, "num_timesteps": 3014656, "_total_timesteps": 3000000, "_num_timesteps_at_start": 0, "seed": null, "action_noise": null, "start_time": 1744147674661249894, "learning_rate": 0.0003, "tensorboard_log": null, "_last_obs": {":type:": "<class 'numpy.ndarray'>", ":serialized:": "gAWVdQIAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJYAAgAAAAAAAAAgBjspYH26l9citaab/68WLSa6FUdXNAAAgD8AAIA/WmCbvWshgj6NO3A9S8Ovvsftzj1y4cy8AAAAAAAAAADmtU0+dBrNPuVK/rsL1s2+NB4UPpaqB70AAAAAAAAAAOZlRb3jXAM9piyCPTfqq75JosU9ukRnvAAAAAAAAAAAbZ0PPkT+nj4xdDW9Nou4vnrcDj6OeIk9AAAAAAAAAACzAW2+W0hpP0jJaz3QXRm/Fiy8voJBCz4AAAAAAAAAAE0yfT04APm7aq4bvgzREj1fVFM9KozwvQAAgD8AAIA/mq9vPtRvlT8y7SQ/+clDvyt/oT6o/CI+AAAAAAAAAACmu+I9OxYRP4Mu3Txv7ee+jAbEPD92p70AAAAAAAAAADNrFD1x3z8/fiQIPh/+E79355u8qsIsPgAAAAAAAAAAmlMbvUhDlbrdC/I4K9zlM0BDijqaKAy4AACAPwAAgD8AI+q8e96OuihSfjmRMGw0Qt61uvWek7gAAIA/AACAP8AZgj7PxNE+i9VXvpC6o74ojuc9Hr2vvQAAAAAAAAAAk/sOvu00ZT+jkFM9TEwLv7Lsqb4yG/M9AAAAAAAAAABg3kM+AYUDPjAY5b4H9RS+rROnPIg8270AAAAAAAAAAPapgj6fdyo/YRDGPYBwEr+8BFI+LSpEPAAAAAAAAAAAlIwFbnVtcHmUjAVkdHlwZZSTlIwCZjSUiYiHlFKUKEsDjAE8lE5OTkr/////Sv////9LAHSUYksQSwiGlIwBQ5R0lFKULg=="}, "_last_episode_starts": {":type:": "<class 'numpy.ndarray'>", ":serialized:": "gAWVgwAAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJYQAAAAAAAAAAAAAAAAAAAAAAAAAAAAAACUjAVudW1weZSMBWR0eXBllJOUjAJiMZSJiIeUUpQoSwOMAXyUTk5OSv////9K/////0sAdJRiSxCFlIwBQ5R0lFKULg=="}, "_last_original_obs": null, "_episode_num": 0, "use_sde": false, "sde_sample_freq": -1, "_current_progress_remaining": -0.004885333333333408, "_stats_window_size": 100, "ep_info_buffer": {":type:": "<class 'collections.deque'>", ":serialized:": "gAWV6AsAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKUKH2UKIwBcpRHQHM/eOXE61eMAWyUS9mMAXSUR0C016OejEehdX2UKGgGR0Bym6L4vexfaAdL2mgIR0C017PX05EMdX2UKGgGR0BwjK5tm+TNaAdLvGgIR0C01/b8rI5pdX2UKGgGR0BwsUF4cFQmaAdLymgIR0C02EkPMB6sdX2UKGgGR0ByFkFwDNhWaAdLz2gIR0C02F08mrsCdX2UKGgGR0BrzZ5/smfHaAdLuGgIR0C02Hm5tm+TdX2UKGgGR0BxobaBZpztaAdL5mgIR0C02Le14Pf9dX2UKGgGR0BvYXD+BH09aAdLwWgIR0C02NuTFERbdX2UKGgGR0Bv+aPjn3cpaAdL0mgIR0C02OGo73fydX2UKGgGR0BxtnEUCaJAaAdLvmgIR0C02Of/WDpUdX2UKGgGR0BwQ/B2wFC+aAdL5WgIR0C02OvIS13MdX2UKGgGR0BxuSMGX5WSaAdL32gIR0C02OnndO6/dX2UKGgGR0Bx+C2Dxsl+aAdL+mgIR0C02OuIRAbAdX2UKGgGR0Byo/dpItlJaAdL+2gIR0C02P1nIyTIdX2UKGgGR0Bxy9RR/EwWaAdLy2gIR0C02QMe8wpOdX2UKGgGR0BymuepXIU8aAdLxGgIR0C02Q+jEehgdX2UKGgGR0BzOo8EFGG3aAdL5GgIR0C02RBvaURndX2UKGgGR0BzESJwbVBlaAdL62gIR0C02R2lQ/HHdX2UKGgGR0Byv7Xf642CaAdLv2gIR0C02UURaouPdX2UKGgGR0BwKNItlI3BaAdLumgIR0C02YZezD4ydX2UKGgGR0BzTMvoNd7faAdLz2gIR0C02cAjt5UtdX2UKGgGR0Bx84yN4qwyaAdLs2gIR0C02dumR/3GdX2UKGgGR0Bwe67CiypraAdLq2gIR0C02i5VXFLndX2UKGgGR0BvRmZqmCRPaAdLsmgIR0C02iy1AqusdX2UKGgGR0BxlHeuV5bAaAdLvWgIR0C03iXsC1Z1dX2UKGgGR0Bxq1IsiB5HaAdLx2gIR0C03iTTvy9VdX2UKGgGR0By2mi35N48aAdL+GgIR0C03iaBNEgGdX2UKGgGR0Bw/kQtjCpFaAdLzGgIR0C03ioTK1XvdX2UKGgGR0BykGzD4xk/aAdL0GgIR0C03ja6J66bdX2UKGgGR0BwghEb5uZUaAdLvWgIR0C03lK19fCzdX2UKGgGR0Byv5hH9WIXaAdL8WgIR0C03mbZFocrdX2UKGgGR0BxxfA31jAjaAdL7mgIR0C03m4bsF+vdX2UKGgGR0Bx6vj3mFJyaAdNAgFoCEdAtN6R0knkUHV9lChoBkdAcfdqC6H0smgHS8poCEdAtN6oaMrEtXV9lChoBkdAceKdDpkf92gHS7loCEdAtN7RjriVB3V9lChoBkdAc2SxH5Jsf2gHTRQBaAhHQLTe0olUp/h1fZQoaAZHQHFpz50r9VFoB0uyaAhHQLTe9N9H+ZR1fZQoaAZHQHD+yeiBXjloB0uxaAhHQLTfVBLPD511fZQoaAZHQHKWcG9pRGdoB0vvaAhHQLTffjriVB51fZQoaAZHQHNwF4oqkM1oB0u+aAhHQLTfguR9w3p1fZQoaAZHQHKhiAUcn3NoB0vQaAhHQLTflPX05EN1fZQoaAZHQHF+1NL127poB0vgaAhHQLTfrQj2SMd1fZQoaAZHQHMGX+6y0KJoB0vZaAhHQLTfsD6WPcV1fZQoaAZHQHLPGjfvWpZoB0voaAhHQLTfwRsuWbB1fZQoaAZHQHAUUcsDnvFoB0vMaAhHQLTfxVE/jbV1fZQoaAZHQHJmvH5rP+poB0v2aAhHQLTf6/82rGR1fZQoaAZHQEsZegte2NNoB0usaAhHQLTgKi0OVgR1fZQoaAZHQHELcLncL0BoB0vpaAhHQLTgLGtITXd1fZQoaAZHQHGtMl1KXfJoB0v0aAhHQLTgPIsRQJp1fZQoaAZHQHAxvJaJQ+FoB0vEaAhHQLTgX1A7gbZ1fZQoaAZHQHMDr/CIk7hoB0v1aAhHQLTgcDtw71Z1fZQoaAZHQHCY8HKOktVoB0u9aAhHQLTggS75Ec91fZQoaAZHQHCMy925hBtoB0vzaAhHQLTgiM9r4351fZQoaAZHQHLY5Etuk1xoB0vLaAhHQLThTcXm/351fZQoaAZHQHHnmZ/kNnZoB0uxaAhHQLThUwYLsrx1fZQoaAZHQHM6GAkLQX1oB0uyaAhHQLThWsO5J9R1fZQoaAZHQHFbo6CDmKZoB0vbaAhHQLTheciW3Sd1fZQoaAZHQHGnG+PBBRhoB0vXaAhHQLThhAnlXBB1fZQoaAZHQHExVAzHjp9oB0v+aAhHQLThl6IFeOZ1fZQoaAZHQG6tf0NBnjBoB0vIaAhHQLThue3hGYt1fZQoaAZHQHJ5nWSU1Q9oB0vyaAhHQLTh2pTMqz91fZQoaAZHQHBt9tdiUgVoB0veaAhHQLTiLKIi1Rd1fZQoaAZHQHGsxfKISDhoB00ZAWgIR0C04jef29L6dX2UKGgGR0BxBbDjzZpSaAdLw2gIR0C04jvdAPd3dX2UKGgGR0BwcYJlar3kaAdL8GgIR0C04lm3rleXdX2UKGgGR0ByAFrBTGYKaAdL2mgIR0C04l5EUj9odX2UKGgGR0BzdKbWmP5paAdL9GgIR0C04nED+zdDdX2UKGgGR0ByVdc9nscAaAdL3mgIR0C04oUbPyCndX2UKGgGR0BwqbOiWVu8aAdNEgFoCEdAtOL7lp48l3V9lChoBkdAc+s+lj3Eh2gHS95oCEdAtONPsu3+dnV9lChoBkdAc1D4QBgeBGgHS9NoCEdAtONgNI9TxXV9lChoBkdAcetLA57w8WgHS81oCEdAtONwCq6vq3V9lChoBkdAcQ+haC+UQmgHS8BoCEdAtONy+N96TnV9lChoBkdAcjOvFFUhm2gHTQEBaAhHQLTjnv/BFd91fZQoaAZHQHD63bAUL2JoB0v3aAhHQLTjtVh1DBx1fZQoaAZHQHDTX/T9bX9oB0vAaAhHQLTj6GbkOqh1fZQoaAZHQHJu0g8r7O5oB00jAWgIR0C04/R42S+ydX2UKGgGR0Bzhs5uIhyKaAdLyGgIR0C05DJbMX7+dX2UKGgGR0BxqgleF+NMaAdL02gIR0C05Dd/z8P4dX2UKGgGR0BweI3qAz55aAdL2WgIR0C05EAsGxD9dX2UKGgGR0BwwWTr3TNMaAdL82gIR0C05FA1vVEvdX2UKGgGR0Bx3g8jiXIEaAdL9GgIR0C05Fq7ZnL8dX2UKGgGR0BxVuDQJHAiaAdNIQFoCEdAtORprXUYsXV9lChoBkdAcznFXaJyhmgHS9FoCEdAtOTUtqYZ23V9lChoBkdAcTCZXuE252gHS7JoCEdAtOTZMlC1JHV9lChoBkdAcT+a6jFhomgHS8RoCEdAtOUed4FA3XV9lChoBkdAcUFnIQvpQmgHS9JoCEdAtOUp1W8yvnV9lChoBkdAbs/njABT42gHS7loCEdAtOVMDzRQanV9lChoBkdAb8z3lCCz1WgHS95oCEdAtOVU6ySmqHV9lChoBkdAccnLHdXT3WgHS+9oCEdAtOW0QZn+Q3V9lChoBkdAb8qKziS7oWgHS+JoCEdAtOXhk+X7cnV9lChoBkdAcW8DSPU8WGgHS71oCEdAtOXkhNdqtnV9lChoBkdAcmmOT7l7t2gHS8RoCEdAtOXnStvGZXV9lChoBkdAclUunuRcNmgHS+9oCEdAtOYK2uxKQXV9lChoBkdAcPlTMJQcgmgHS81oCEdAtOYZQVKwp3V9lChoBkdAcGsAq/dqL2gHS8VoCEdAtOYii+L3sXV9lChoBkdAcuLcdYGMXWgHTaEBaAhHQLTmLXAM2FZ1fZQoaAZHQHMFCS7oSthoB0vwaAhHQLTmSTfR/mV1fZQoaAZHQEmdn6l+EytoB0uraAhHQLTmVRiPQv91fZQoaAZHQHJ1SXlbNbFoB0vzaAhHQLTmaQ7tAs11ZS4="}, "ep_success_buffer": {":type:": "<class 'collections.deque'>", ":serialized:": "gAWVIAAAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKULg=="}, "_n_updates": 1476, "observation_space": {":type:": "<class 'gymnasium.spaces.box.Box'>", ":serialized:": "gAWVcAIAAAAAAACMFGd5bW5hc2l1bS5zcGFjZXMuYm94lIwDQm94lJOUKYGUfZQojAVkdHlwZZSMBW51bXB5lGgFk5SMAmY0lImIh5RSlChLA4wBPJROTk5K/////0r/////SwB0lGKMDWJvdW5kZWRfYmVsb3eUjBJudW1weS5jb3JlLm51bWVyaWOUjAtfZnJvbWJ1ZmZlcpSTlCiWCAAAAAAAAAABAQEBAQEBAZRoB4wCYjGUiYiHlFKUKEsDjAF8lE5OTkr/////Sv////9LAHSUYksIhZSMAUOUdJRSlIwNYm91bmRlZF9hYm92ZZRoECiWCAAAAAAAAAABAQEBAQEBAZRoFEsIhZRoGHSUUpSMBl9zaGFwZZRLCIWUjANsb3eUaBAoliAAAAAAAAAAAAC0wgAAtMIAAKDAAACgwNsPScAAAKDAAAAAgAAAAICUaApLCIWUaBh0lFKUjARoaWdolGgQKJYgAAAAAAAAAAAAtEIAALRCAACgQAAAoEDbD0lAAACgQAAAgD8AAIA/lGgKSwiFlGgYdJRSlIwIbG93X3JlcHKUjFtbLTkwLiAgICAgICAgLTkwLiAgICAgICAgIC01LiAgICAgICAgIC01LiAgICAgICAgIC0zLjE0MTU5MjcgIC01LgogIC0wLiAgICAgICAgIC0wLiAgICAgICBdlIwJaGlnaF9yZXBylIxTWzkwLiAgICAgICAgOTAuICAgICAgICAgNS4gICAgICAgICA1LiAgICAgICAgIDMuMTQxNTkyNyAgNS4KICAxLiAgICAgICAgIDEuICAgICAgIF2UjApfbnBfcmFuZG9tlE51Yi4=", "dtype": "float32", "bounded_below": "[ True True True True True True True True]", "bounded_above": "[ True True True True True True True True]", "_shape": [8], "low": "[-90. -90. -5. -5. -3.1415927 -5.\n -0. -0. ]", "high": "[90. 90. 5. 5. 3.1415927 5.\n 1. 1. ]", "low_repr": "[-90. -90. -5. -5. -3.1415927 -5.\n -0. -0. ]", "high_repr": "[90. 90. 5. 5. 3.1415927 5.\n 1. 1. ]", "_np_random": null}, "action_space": {":type:": "<class 'gymnasium.spaces.discrete.Discrete'>", ":serialized:": "gAWV1QAAAAAAAACMGWd5bW5hc2l1bS5zcGFjZXMuZGlzY3JldGWUjAhEaXNjcmV0ZZSTlCmBlH2UKIwBbpSMFW51bXB5LmNvcmUubXVsdGlhcnJheZSMBnNjYWxhcpSTlIwFbnVtcHmUjAVkdHlwZZSTlIwCaTiUiYiHlFKUKEsDjAE8lE5OTkr/////Sv////9LAHSUYkMIBAAAAAAAAACUhpRSlIwFc3RhcnSUaAhoDkMIAAAAAAAAAACUhpRSlIwGX3NoYXBllCloCmgOjApfbnBfcmFuZG9tlE51Yi4=", "n": "4", "start": "0", "_shape": [], "dtype": "int64", "_np_random": null}, "n_envs": 16, "n_steps": 1024, "gamma": 0.999, "gae_lambda": 0.98, "ent_coef": 0.01, "vf_coef": 0.5, "max_grad_norm": 0.5, "batch_size": 64, "n_epochs": 4, "clip_range": {":type:": "<class 'function'>", ":serialized:": "gAWV5QIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwOX21ha2VfZnVuY3Rpb26Uk5QoaACMDV9idWlsdGluX3R5cGWUk5SMCENvZGVUeXBllIWUUpQoSwFLAEsASwFLAUsTQwSIAFMAlE6FlCmMAV+UhZSMZS9ob21lL3NlbmxpLy5jb25kYS9lbnZzL3NlbmxpX1B5dGhvbjMuOC9saWIvcHl0aG9uMy44L3NpdGUtcGFja2FnZXMvc3RhYmxlX2Jhc2VsaW5lczMvY29tbW9uL3V0aWxzLnB5lIwEZnVuY5RLhEMCAAGUjAN2YWyUhZQpdJRSlH2UKIwLX19wYWNrYWdlX1+UjBhzdGFibGVfYmFzZWxpbmVzMy5jb21tb26UjAhfX25hbWVfX5SMHnN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi51dGlsc5SMCF9fZmlsZV9flIxlL2hvbWUvc2VubGkvLmNvbmRhL2VudnMvc2VubGlfUHl0aG9uMy44L2xpYi9weXRob24zLjgvc2l0ZS1wYWNrYWdlcy9zdGFibGVfYmFzZWxpbmVzMy9jb21tb24vdXRpbHMucHmUdU5OaACMEF9tYWtlX2VtcHR5X2NlbGyUk5QpUpSFlHSUUpRoAIwSX2Z1bmN0aW9uX3NldHN0YXRllJOUaB99lH2UKGgWjARmdW5jlIwMX19xdWFsbmFtZV9flIwZY29uc3RhbnRfZm4uPGxvY2Fscz4uZnVuY5SMD19fYW5ub3RhdGlvbnNfX5R9lIwOX19rd2RlZmF1bHRzX1+UTowMX19kZWZhdWx0c19flE6MCl9fbW9kdWxlX1+UaBeMB19fZG9jX1+UTowLX19jbG9zdXJlX1+UaACMCl9tYWtlX2NlbGyUk5RHP8mZmZmZmZqFlFKUhZSMF19jbG91ZHBpY2tsZV9zdWJtb2R1bGVzlF2UjAtfX2dsb2JhbHNfX5R9lHWGlIZSMC4="}, "clip_range_vf": null, "normalize_advantage": true, "target_kl": null, "lr_schedule": {":type:": "<class 'function'>", ":serialized:": "gAWV5QIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwOX21ha2VfZnVuY3Rpb26Uk5QoaACMDV9idWlsdGluX3R5cGWUk5SMCENvZGVUeXBllIWUUpQoSwFLAEsASwFLAUsTQwSIAFMAlE6FlCmMAV+UhZSMZS9ob21lL3NlbmxpLy5jb25kYS9lbnZzL3NlbmxpX1B5dGhvbjMuOC9saWIvcHl0aG9uMy44L3NpdGUtcGFja2FnZXMvc3RhYmxlX2Jhc2VsaW5lczMvY29tbW9uL3V0aWxzLnB5lIwEZnVuY5RLhEMCAAGUjAN2YWyUhZQpdJRSlH2UKIwLX19wYWNrYWdlX1+UjBhzdGFibGVfYmFzZWxpbmVzMy5jb21tb26UjAhfX25hbWVfX5SMHnN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi51dGlsc5SMCF9fZmlsZV9flIxlL2hvbWUvc2VubGkvLmNvbmRhL2VudnMvc2VubGlfUHl0aG9uMy44L2xpYi9weXRob24zLjgvc2l0ZS1wYWNrYWdlcy9zdGFibGVfYmFzZWxpbmVzMy9jb21tb24vdXRpbHMucHmUdU5OaACMEF9tYWtlX2VtcHR5X2NlbGyUk5QpUpSFlHSUUpRoAIwSX2Z1bmN0aW9uX3NldHN0YXRllJOUaB99lH2UKGgWjARmdW5jlIwMX19xdWFsbmFtZV9flIwZY29uc3RhbnRfZm4uPGxvY2Fscz4uZnVuY5SMD19fYW5ub3RhdGlvbnNfX5R9lIwOX19rd2RlZmF1bHRzX1+UTowMX19kZWZhdWx0c19flE6MCl9fbW9kdWxlX1+UaBeMB19fZG9jX1+UTowLX19jbG9zdXJlX1+UaACMCl9tYWtlX2NlbGyUk5RHPzOpKjBVMmGFlFKUhZSMF19jbG91ZHBpY2tsZV9zdWJtb2R1bGVzlF2UjAtfX2dsb2JhbHNfX5R9lHWGlIZSMC4="}, "system_info": {"OS": "Linux-5.15.0-134-generic-x86_64-with-glibc2.17 # 145-Ubuntu SMP Wed Feb 12 20:08:39 UTC 2025", "Python": "3.8.20", "Stable-Baselines3": "2.0.0a5", "PyTorch": "2.4.1+cu121", "GPU Enabled": "True", "Numpy": "1.24.4", "Cloudpickle": "3.1.1", "Gymnasium": "0.28.1", "OpenAI Gym": "0.26.2"}}
|
replay.mp4
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f8d7ef94584edb1aaae13c897f15706c589d8870f2b0bbd8ee6e5e98a76b042a
|
| 3 |
+
size 146612
|
results.json
CHANGED
|
@@ -1 +1 @@
|
|
| 1 |
-
{"mean_reward":
|
|
|
|
| 1 |
+
{"mean_reward": 295.39934800000003, "std_reward": 14.381062367849559, "is_deterministic": true, "n_eval_episodes": 10, "eval_datetime": "2025-04-08T22:07:41.617129"}
|