First Commit
Browse files- README.md +1 -1
- config.json +1 -1
- ppo-LunarLander-v2.zip +2 -2
- ppo-LunarLander-v2/data +16 -16
- ppo-LunarLander-v2/policy.optimizer.pth +1 -1
- ppo-LunarLander-v2/policy.pth +1 -1
- replay.mp4 +2 -2
- results.json +1 -1
README.md
CHANGED
|
@@ -16,7 +16,7 @@ model-index:
|
|
| 16 |
type: LunarLander-v2
|
| 17 |
metrics:
|
| 18 |
- type: mean_reward
|
| 19 |
-
value:
|
| 20 |
name: mean_reward
|
| 21 |
verified: false
|
| 22 |
---
|
|
|
|
| 16 |
type: LunarLander-v2
|
| 17 |
metrics:
|
| 18 |
- type: mean_reward
|
| 19 |
+
value: 253.91 +/- 20.89
|
| 20 |
name: mean_reward
|
| 21 |
verified: false
|
| 22 |
---
|
config.json
CHANGED
|
@@ -1 +1 @@
|
|
| 1 |
-
{"policy_class": {":type:": "<class 'abc.ABCMeta'>", ":serialized:": "gAWVOwAAAAAAAACMIXN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi5wb2xpY2llc5SMEUFjdG9yQ3JpdGljUG9saWN5lJOULg==", "__module__": "stable_baselines3.common.policies", "__doc__": "\n Policy class for actor-critic algorithms (has both policy and value prediction).\n Used by A2C, PPO and the likes.\n\n :param observation_space: Observation space\n :param action_space: Action space\n :param lr_schedule: Learning rate schedule (could be constant)\n :param net_arch: The specification of the policy and value networks.\n :param activation_fn: Activation function\n :param ortho_init: Whether to use or not orthogonal initialization\n :param use_sde: Whether to use State Dependent Exploration or not\n :param log_std_init: Initial value for the log standard deviation\n :param full_std: Whether to use (n_features x n_actions) parameters\n for the std instead of only (n_features,) when using gSDE\n :param use_expln: Use ``expln()`` function instead of ``exp()`` to ensure\n a positive standard deviation (cf paper). It allows to keep variance\n above zero and prevent it from growing too fast. In practice, ``exp()`` is usually enough.\n :param squash_output: Whether to squash the output using a tanh function,\n this allows to ensure boundaries when using gSDE.\n :param features_extractor_class: Features extractor to use.\n :param features_extractor_kwargs: Keyword arguments\n to pass to the features extractor.\n :param share_features_extractor: If True, the features extractor is shared between the policy and value networks.\n :param normalize_images: Whether to normalize images or not,\n dividing by 255.0 (True by default)\n :param optimizer_class: The optimizer to use,\n ``th.optim.Adam`` by default\n :param optimizer_kwargs: Additional keyword arguments,\n excluding the learning rate, to pass to the optimizer\n ", "__init__": "<function ActorCriticPolicy.__init__ at 0x7f5cf95dfa60>", "_get_constructor_parameters": "<function ActorCriticPolicy._get_constructor_parameters at 0x7f5cf95dfb00>", "reset_noise": "<function ActorCriticPolicy.reset_noise at 0x7f5cf95dfba0>", "_build_mlp_extractor": "<function ActorCriticPolicy._build_mlp_extractor at 0x7f5cf95dfc40>", "_build": "<function ActorCriticPolicy._build at 0x7f5cf95dfce0>", "forward": "<function ActorCriticPolicy.forward at 0x7f5cf95dfd80>", "extract_features": "<function ActorCriticPolicy.extract_features at 0x7f5cf95dfe20>", "_get_action_dist_from_latent": "<function ActorCriticPolicy._get_action_dist_from_latent at 0x7f5cf95dfec0>", "_predict": "<function ActorCriticPolicy._predict at 0x7f5cf95dff60>", "evaluate_actions": "<function ActorCriticPolicy.evaluate_actions at 0x7f5cf95e8040>", "get_distribution": "<function ActorCriticPolicy.get_distribution at 0x7f5cf95e80e0>", "predict_values": "<function ActorCriticPolicy.predict_values at 0x7f5cf95e8180>", "__abstractmethods__": "frozenset()", "_abc_impl": "<_abc._abc_data object at 0x7f5cf9b58580>"}, "verbose": 1, "policy_kwargs": {}, "num_timesteps": 1015808, "_total_timesteps": 1000000.0, "_num_timesteps_at_start": 0, "seed": null, "action_noise": null, "start_time": 1749459478854926578, "learning_rate": 0.0003, "tensorboard_log": null, "_last_obs": {":type:": "<class 'numpy.ndarray'>", ":serialized:": "gAWVdgIAAAAAAACME251bXB5Ll9jb3JlLm51bWVyaWOUjAtfZnJvbWJ1ZmZlcpSTlCiWAAIAAAAAAACAA7690CmaP4JEzr4zd+C+ozWGvTXAzb0AAAAAAAAAAAAqFz2DSE68BhLzPJxZ67wGdGy9RrJovgAAgD8AAIA/s+djPszDhz+dkUQ+6TbwvvPWAj6YSIe8AAAAAAAAAABmJIa+zdiLPxwKk74AOe2+utAOvvY7yzwAAAAAAAAAAM3sabpcGhs7RJSWPWwgY763Lyc90yD3vAAAAAAAAAAATZxjvUh/47rFWjI8RfmRPCDQibur4nw9AACAPwAAgD9zZtO9KoynP9Vtcr5GWbu+ZnmhvUG6pL0AAAAAAAAAADMSQ72dLYo/yGGXvUwTv741fqc7+BRjPAAAAAAAAAAAmgC+vu0vhj9ClFS+A5ylvkylbb44J8A9AAAAAAAAAACNvry9WGDPPcAU/z2O01e+mMaNOz3Ab7wAAAAAAAAAAIAz0L25fEc+8n2GOyMvXL4aTAC9ZuxKuwAAAAAAAAAAM2cqvDasCj0g7w69nxVQvkvh07ua/4w8AAAAAAAAAAAAXdS81zZLu/eRrztnRI08nAqsvAVQcj0AAIA/AACAPyaIij0cXSu8U2wVvWMicD104Qq958OZvAAAgD8AAIA/WpyRvQNnMz2Ajrm8Y5CGvj8T4TpUy0O8AAAAAAAAAAAzV1g9rq2Zug0cmDNK97cvrGG3ukUHz7MAAIA/AACAP5SMBW51bXB5lIwFZHR5cGWUk5SMAmY0lImIh5RSlChLA4wBPJROTk5K/////0r/////SwB0lGJLEEsIhpSMAUOUdJRSlC4="}, "_last_episode_starts": {":type:": "<class 'numpy.ndarray'>", ":serialized:": "gAWVhAAAAAAAAACME251bXB5Ll9jb3JlLm51bWVyaWOUjAtfZnJvbWJ1ZmZlcpSTlCiWEAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAlIwFbnVtcHmUjAVkdHlwZZSTlIwCYjGUiYiHlFKUKEsDjAF8lE5OTkr/////Sv////9LAHSUYksQhZSMAUOUdJRSlC4="}, "_last_original_obs": null, "_episode_num": 0, "use_sde": false, "sde_sample_freq": -1, "_current_progress_remaining": -0.015808000000000044, "_stats_window_size": 100, "ep_info_buffer": {":type:": "<class 'collections.deque'>", ":serialized:": "gAWVLwwAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKUKH2UKIwBcpRHwBkNGI9C/oKMAWyUS9aMAXSUR0CZm1BZpztDdX2UKGgGR0BtxCQT238XaAdNDgFoCEdAmZuJNTLntHV9lChoBkdAcYTwH7gsLGgHTTcBaAhHQJmcwYdhiLF1fZQoaAZHQFAbsa86FM9oB0vgaAhHQJmfAN9YwIt1fZQoaAZHQDVyiEg4ffZoB0vNaAhHQJmfPifg75p1fZQoaAZHQE9bJeVs1sNoB0vZaAhHQJmfsraufVZ1fZQoaAZHQHAvVHnU2DRoB00WAWgIR0CZn7NMGorGdX2UKGgGR0BuZl+5OJtSaAdNCgFoCEdAmZ/RPfsNUnV9lChoBkdAcEkqyGBWgmgHTVoBaAhHQJmgaixmkFh1fZQoaAZHQG2qMzVMEidoB00qAWgIR0CZoHMTewcHdX2UKGgGR0BxaISZjQRgaAdNKwFoCEdAmaG3bRF7U3V9lChoBkdAcQYpcophF2gHTRkBaAhHQJmh3SQYDT11fZQoaAZHwBPjO1OTJQtoB0vDaAhHQJmiSm+Cbtt1fZQoaAZHQG+t/zasZHdoB02RAWgIR0CZom3wTdtVdX2UKGgGR0BwE/kXDWK/aAdNdgFoCEdAmaKQpnYg73V9lChoBkdAch/CT2WY4WgHTXMBaAhHQJmjWtMfzSV1fZQoaAZHQHBgqo/A0sRoB017AWgIR0CZpebBoEjgdX2UKGgGR0BxQe1Z1V5saAdL92gIR0CZpg/+KjzqdX2UKGgGR0Bxr1pWV/tqaAdNCgFoCEdAmadNHUc4pHV9lChoBkdAcsLLDAJswmgHTVUBaAhHQJmnan5zo2Z1fZQoaAZHQEvxOFg2IftoB0vnaAhHQJmoihh6Skl1fZQoaAZHQHGNBzRx95RoB01CAWgIR0CZqK3yI55rdX2UKGgGR0BPit5UtI07aAdL5GgIR0CZqPFCLMs6dX2UKGgGR0BuK43o9s7/aAdNPwFoCEdAmakhvWH1vnV9lChoBkdAa0wMOPNmlWgHTTMBaAhHQJmpWJ0nw5N1fZQoaAZHQHD4CpWFN+NoB0v1aAhHQJmpwChew9t1fZQoaAZHQHGqcyzolldoB00qAWgIR0CZql3fhuO0dX2UKGgGR0BxF7H/95yEaAdNQAFoCEdAmau7s8gZCXV9lChoBkdAcZ9i1iONpGgHTbcBaAhHQJmtPLq2SdR1fZQoaAZHQE8+vTPSlWRoB0voaAhHQJmugkhRqGl1fZQoaAZHQDIY/cFhXsBoB0vNaAhHQJmuxtk4FRp1fZQoaAZHQHAuKzE74i5oB00kAWgIR0CZrxM3qAz6dX2UKGgGR0BwfIDfWMCLaAdNkQFoCEdAma+AmZ3LWHV9lChoBkdAa5WEqUeMh2gHTTEBaAhHQJmw0YyfthN1fZQoaAZHQHI+1H4GlhxoB01jAWgIR0CZsPqo60Y1dX2UKGgGR0BxawnssxwiaAdNCQFoCEdAmbFYuPFNtnV9lChoBkdAcPMYG+sYEWgHTRcBaAhHQJmxWSTyJ9B1fZQoaAZHQET/kQPI4l1oB0u6aAhHQJmxteUpuuR1fZQoaAZHQHLAtkjHGS9oB00zAWgIR0CZsmV2Rq46dX2UKGgGR0Bwfb6BRQ7+aAdNNwFoCEdAmbM/OQhfSnV9lChoBkdAcGztCRfWtmgHTUkBaAhHQJmzZpaiblR1fZQoaAZHQG9Jc7IT4+NoB008AWgIR0CZs/+YtxuLdX2UKGgGR0AymRQJokAxaAdLv2gIR0CZtKecQRPHdX2UKGgGR0ByVRvwVj7RaAdNEQFoCEdAmbavzSThYXV9lChoBkdAThpV2icoY2gHS+5oCEdAmcruKoAGS3V9lChoBkdAcWOUdaMaTGgHTX8BaAhHQJnL1aNdZ7p1fZQoaAZHQHEbsF+uvEFoB01cAWgIR0CZy//8VHnVdX2UKGgGR0Bu8kRBeHBUaAdNIQFoCEdAmcxv3FkxynV9lChoBkdAbyh3os7MgWgHTSMBaAhHQJnMt70Fr2x1fZQoaAZHQHMFqA8SwnpoB01VAWgIR0CZzO9E1EVndX2UKGgGR0Bye1TAFgUlaAdL+mgIR0CZzQJU5uIidX2UKGgGR0BlLuMl1KXfaAdN6ANoCEdAmc3vDgqEvnV9lChoBkdAcG6752yLRGgHTUUBaAhHQJnO6sjmjj91fZQoaAZHQHFJaouPFNtoB01RAWgIR0CZzuvBrN4adX2UKGgGR0Bv+jpNbkfcaAdNGwFoCEdAmc9oBvJiiXV9lChoBkdAceu925hBq2gHTS0BaAhHQJnQlRl6JIl1fZQoaAZHQGRb1B+nZTRoB03oA2gIR0CZ0PyjYZl4dX2UKGgGR0BzCiu4gA6uaAdNVQFoCEdAmdD7/Khcq3V9lChoBkdAP/wiFCb+cmgHS7loCEdAmdE4OQQtjHV9lChoBkdAcbxci4axYGgHTVMBaAhHQJnSJ3NcGC91fZQoaAZHQG9MT4+KTB9oB00LAWgIR0CZ0i6g/TsqdX2UKGgGR0Bx0IVHnU2DaAdNGAFoCEdAmdRLMX7+DXV9lChoBkdAbkySVW0Z32gHTRwBaAhHQJnU9efI0ZZ1fZQoaAZHQGzQqS5iExtoB00VAWgIR0CZ1PYJ3PiUdX2UKGgGR0BwqQlE7W/baAdNBAFoCEdAmdV55/smfHV9lChoBkdAcFZ7NSqEOGgHTTABaAhHQJnWAwWWQfZ1fZQoaAZHQG/dDa4+bExoB01zAWgIR0CZ15M1jy4GdX2UKGgGR0BySmPn0TURaAdNMQFoCEdAmdfmGucME3V9lChoBkdAccVkq+ajOGgHTSUBaAhHQJnX8MQVbiZ1fZQoaAZHQHGLvGhmGudoB0v8aAhHQJnYFvHcUM51fZQoaAZHQG+ExUWEbo9oB01aAWgIR0CZ2TIY3vQXdX2UKGgGR0Bu2tU0elsQaAdNDwFoCEdAmdmMj3VTaXV9lChoBkdAbstt6X0GvGgHTRsBaAhHQJnZl8PWhAZ1fZQoaAZHQHDXCckMTexoB00lAWgIR0CZ2dS619fDdX2UKGgGR0BwQzQ0GeMAaAdNCgFoCEdAmdpk8aGYbHV9lChoBkdAcLWENe+mFmgHTdwBaAhHQJnbPa7EpAl1fZQoaAZHQHCkjLr5ZbJoB006AWgIR0CZ241+y7f6dX2UKGgGR0BNN3Ehq0tzaAdLsWgIR0CZ3HhjOLR8dX2UKGgGR0BwK/ck+otMaAdNFwFoCEdAmd1EXpGFz3V9lChoBkdAcMFPikwevWgHTRIBaAhHQJndmef7Jnx1fZQoaAZHQG8kX1zySV5oB01AAWgIR0CZ3fPwNLDidX2UKGgGR0BwZeNrCWNWaAdNMgFoCEdAmd4Q/xDst3V9lChoBkdAbvfQ3xWkrWgHTTQBaAhHQJne88NhE0B1fZQoaAZHQG2m/1YhdMVoB0v8aAhHQJngDy9VWCF1fZQoaAZHQHBzAbIcR15oB00wAWgIR0CZ4JfCyhSMdX2UKGgGR0BwM3EP1+RYaAdNPAFoCEdAmeDQbEP1+XV9lChoBkdAbnjmT1TR6WgHTVcBaAhHQJnhuaVlf7d1fZQoaAZHQHECywOe8PFoB00kAWgIR0CZ4cx3V09ydX2UKGgGR0BwzRrSE12raAdNKQFoCEdAmeIyBXjlxXV9lChoBkdAb8bPiT+vQmgHTTcBaAhHQJniQ+fRNRF1fZQoaAZHQG9WBgmZ3LVoB00hAWgIR0CZ4oqSHM2WdX2UKGgGR0A05WhAWzniaAdNAQFoCEdAmeKYgvDgqHV9lChoBkdAb/0eg+Qlr2gHTQUBaAhHQJni9fb9If91fZQoaAZHQEdKgFotcwBoB0vjaAhHQJni/jlxOtZ1fZQoaAZHQGzf/g75mAdoB00sAWgIR0CZ5d17Y02tdX2UKGgGR0BwC47EHdGiaAdNLgFoCEdAmeZSHRCx/3V9lChoBkdAb3ElrM1TBWgHTUkBaAhHQJnmcm+j/Mp1fZQoaAZHQG/A1rylN11oB00SAWgIR0CZ5pwbEP1+dX2UKGgGR0BC2iiyprDZaAdL5GgIR0CZ5uF6iTMadX2UKGgGR0BvaHLeQ+2WaAdNWAFoCEdAmeebi2lVLnVlLg=="}, "ep_success_buffer": {":type:": "<class 'collections.deque'>", ":serialized:": "gAWVIAAAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKULg=="}, "_n_updates": 248, "observation_space": {":type:": "<class 'gymnasium.spaces.box.Box'>", ":serialized:": "gAWVdwIAAAAAAACMFGd5bW5hc2l1bS5zcGFjZXMuYm94lIwDQm94lJOUKYGUfZQojAVkdHlwZZSMBW51bXB5lIwFZHR5cGWUk5SMAmY0lImIh5RSlChLA4wBPJROTk5K/////0r/////SwB0lGKMDWJvdW5kZWRfYmVsb3eUjBNudW1weS5fY29yZS5udW1lcmljlIwLX2Zyb21idWZmZXKUk5QolggAAAAAAAAAAQEBAQEBAQGUaAiMAmIxlImIh5RSlChLA4wBfJROTk5K/////0r/////SwB0lGJLCIWUjAFDlHSUUpSMDWJvdW5kZWRfYWJvdmWUaBEolggAAAAAAAAAAQEBAQEBAQGUaBVLCIWUaBl0lFKUjAZfc2hhcGWUSwiFlIwDbG93lGgRKJYgAAAAAAAAAAAAtMIAALTCAACgwAAAoMDbD0nAAACgwAAAAIAAAACAlGgLSwiFlGgZdJRSlIwEaGlnaJRoESiWIAAAAAAAAAAAALRCAAC0QgAAoEAAAKBA2w9JQAAAoEAAAIA/AACAP5RoC0sIhZRoGXSUUpSMCGxvd19yZXBylIxbWy05MC4gICAgICAgIC05MC4gICAgICAgICAtNS4gICAgICAgICAtNS4gICAgICAgICAtMy4xNDE1OTI3ICAtNS4KICAtMC4gICAgICAgICAtMC4gICAgICAgXZSMCWhpZ2hfcmVwcpSMU1s5MC4gICAgICAgIDkwLiAgICAgICAgIDUuICAgICAgICAgNS4gICAgICAgICAzLjE0MTU5MjcgIDUuCiAgMS4gICAgICAgICAxLiAgICAgICBdlIwKX25wX3JhbmRvbZROdWIu", "dtype": "float32", "bounded_below": "[ True True True True True True True True]", "bounded_above": "[ True True True True True True True True]", "_shape": [8], "low": "[-90. -90. -5. -5. -3.1415927 -5.\n -0. -0. ]", "high": "[90. 90. 5. 5. 3.1415927 5.\n 1. 1. ]", "low_repr": "[-90. -90. -5. -5. -3.1415927 -5.\n -0. -0. ]", "high_repr": "[90. 90. 5. 5. 3.1415927 5.\n 1. 1. ]", "_np_random": null}, "action_space": {":type:": "<class 'gymnasium.spaces.discrete.Discrete'>", ":serialized:": "gAWV3AAAAAAAAACMGWd5bW5hc2l1bS5zcGFjZXMuZGlzY3JldGWUjAhEaXNjcmV0ZZSTlCmBlH2UKIwBbpSMFm51bXB5Ll9jb3JlLm11bHRpYXJyYXmUjAZzY2FsYXKUk5SMBW51bXB5lIwFZHR5cGWUk5SMAmk4lImIh5RSlChLA4wBPJROTk5K/////0r/////SwB0lGJDCAQAAAAAAAAAlIaUUpSMBXN0YXJ0lGgIaA5DCAAAAAAAAAAAlIaUUpSMBl9zaGFwZZQpjAVkdHlwZZRoDowKX25wX3JhbmRvbZROdWIu", "n": "4", "start": "0", "_shape": [], "dtype": "int64", "_np_random": null}, "n_envs": 16, "n_steps": 1024, "gamma": 0.999, "gae_lambda": 0.98, "ent_coef": 0.01, "vf_coef": 0.5, "max_grad_norm": 0.5, "batch_size": 64, "n_epochs": 4, "clip_range": {":type:": "<class 'function'>", ":serialized:": "gAWV1gIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwOX21ha2VfZnVuY3Rpb26Uk5QoaACMDV9idWlsdGluX3R5cGWUk5SMCENvZGVUeXBllIWUUpQoSwFLAEsASwFLAUsTQwiVAZcAiQFTAJROhZQpjAFflIWUjEkvdXNyL2xvY2FsL2xpYi9weXRob24zLjExL2Rpc3QtcGFja2FnZXMvc3RhYmxlX2Jhc2VsaW5lczMvY29tbW9uL3V0aWxzLnB5lIwEZnVuY5SMGWNvbnN0YW50X2ZuLjxsb2NhbHM+LmZ1bmOUS4RDCPiAANgPEogKlEMAlIwDdmFslIWUKXSUUpR9lCiMC19fcGFja2FnZV9flIwYc3RhYmxlX2Jhc2VsaW5lczMuY29tbW9ulIwIX19uYW1lX1+UjB5zdGFibGVfYmFzZWxpbmVzMy5jb21tb24udXRpbHOUjAhfX2ZpbGVfX5SMSS91c3IvbG9jYWwvbGliL3B5dGhvbjMuMTEvZGlzdC1wYWNrYWdlcy9zdGFibGVfYmFzZWxpbmVzMy9jb21tb24vdXRpbHMucHmUdU5OaACMEF9tYWtlX2VtcHR5X2NlbGyUk5QpUpSFlHSUUpRoAIwSX2Z1bmN0aW9uX3NldHN0YXRllJOUaCF9lH2UKGgYjARmdW5jlIwMX19xdWFsbmFtZV9flIwZY29uc3RhbnRfZm4uPGxvY2Fscz4uZnVuY5SMD19fYW5ub3RhdGlvbnNfX5R9lIwOX19rd2RlZmF1bHRzX1+UTowMX19kZWZhdWx0c19flE6MCl9fbW9kdWxlX1+UaBmMB19fZG9jX1+UTowLX19jbG9zdXJlX1+UaACMCl9tYWtlX2NlbGyUk5RHP8mZmZmZmZqFlFKUhZSMF19jbG91ZHBpY2tsZV9zdWJtb2R1bGVzlF2UjAtfX2dsb2JhbHNfX5R9lHWGlIZSMC4="}, "clip_range_vf": null, "normalize_advantage": true, "target_kl": null, "lr_schedule": {":type:": "<class 'function'>", ":serialized:": "gAWV1gIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwOX21ha2VfZnVuY3Rpb26Uk5QoaACMDV9idWlsdGluX3R5cGWUk5SMCENvZGVUeXBllIWUUpQoSwFLAEsASwFLAUsTQwiVAZcAiQFTAJROhZQpjAFflIWUjEkvdXNyL2xvY2FsL2xpYi9weXRob24zLjExL2Rpc3QtcGFja2FnZXMvc3RhYmxlX2Jhc2VsaW5lczMvY29tbW9uL3V0aWxzLnB5lIwEZnVuY5SMGWNvbnN0YW50X2ZuLjxsb2NhbHM+LmZ1bmOUS4RDCPiAANgPEogKlEMAlIwDdmFslIWUKXSUUpR9lCiMC19fcGFja2FnZV9flIwYc3RhYmxlX2Jhc2VsaW5lczMuY29tbW9ulIwIX19uYW1lX1+UjB5zdGFibGVfYmFzZWxpbmVzMy5jb21tb24udXRpbHOUjAhfX2ZpbGVfX5SMSS91c3IvbG9jYWwvbGliL3B5dGhvbjMuMTEvZGlzdC1wYWNrYWdlcy9zdGFibGVfYmFzZWxpbmVzMy9jb21tb24vdXRpbHMucHmUdU5OaACMEF9tYWtlX2VtcHR5X2NlbGyUk5QpUpSFlHSUUpRoAIwSX2Z1bmN0aW9uX3NldHN0YXRllJOUaCF9lH2UKGgYjARmdW5jlIwMX19xdWFsbmFtZV9flIwZY29uc3RhbnRfZm4uPGxvY2Fscz4uZnVuY5SMD19fYW5ub3RhdGlvbnNfX5R9lIwOX19rd2RlZmF1bHRzX1+UTowMX19kZWZhdWx0c19flE6MCl9fbW9kdWxlX1+UaBmMB19fZG9jX1+UTowLX19jbG9zdXJlX1+UaACMCl9tYWtlX2NlbGyUk5RHPzOpKjBVMmGFlFKUhZSMF19jbG91ZHBpY2tsZV9zdWJtb2R1bGVzlF2UjAtfX2dsb2JhbHNfX5R9lHWGlIZSMC4="}, "system_info": {"OS": "Linux-6.1.123+-x86_64-with-glibc2.35 # 1 SMP PREEMPT_DYNAMIC Sun Mar 30 16:01:29 UTC 2025", "Python": "3.11.13", "Stable-Baselines3": "2.0.0a5", "PyTorch": "2.6.0+cu124", "GPU Enabled": "True", "Numpy": "2.0.2", "Cloudpickle": "3.1.1", "Gymnasium": "0.28.1", "OpenAI Gym": "0.25.2"}}
|
|
|
|
| 1 |
+
{"policy_class": {":type:": "<class 'abc.ABCMeta'>", ":serialized:": "gAWVOwAAAAAAAACMIXN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi5wb2xpY2llc5SMEUFjdG9yQ3JpdGljUG9saWN5lJOULg==", "__module__": "stable_baselines3.common.policies", "__doc__": "\n Policy class for actor-critic algorithms (has both policy and value prediction).\n Used by A2C, PPO and the likes.\n\n :param observation_space: Observation space\n :param action_space: Action space\n :param lr_schedule: Learning rate schedule (could be constant)\n :param net_arch: The specification of the policy and value networks.\n :param activation_fn: Activation function\n :param ortho_init: Whether to use or not orthogonal initialization\n :param use_sde: Whether to use State Dependent Exploration or not\n :param log_std_init: Initial value for the log standard deviation\n :param full_std: Whether to use (n_features x n_actions) parameters\n for the std instead of only (n_features,) when using gSDE\n :param use_expln: Use ``expln()`` function instead of ``exp()`` to ensure\n a positive standard deviation (cf paper). It allows to keep variance\n above zero and prevent it from growing too fast. In practice, ``exp()`` is usually enough.\n :param squash_output: Whether to squash the output using a tanh function,\n this allows to ensure boundaries when using gSDE.\n :param features_extractor_class: Features extractor to use.\n :param features_extractor_kwargs: Keyword arguments\n to pass to the features extractor.\n :param share_features_extractor: If True, the features extractor is shared between the policy and value networks.\n :param normalize_images: Whether to normalize images or not,\n dividing by 255.0 (True by default)\n :param optimizer_class: The optimizer to use,\n ``th.optim.Adam`` by default\n :param optimizer_kwargs: Additional keyword arguments,\n excluding the learning rate, to pass to the optimizer\n ", "__init__": "<function ActorCriticPolicy.__init__ at 0x793f0e8f11c0>", "_get_constructor_parameters": "<function ActorCriticPolicy._get_constructor_parameters at 0x793f0e8f1260>", "reset_noise": "<function ActorCriticPolicy.reset_noise at 0x793f0e8f1300>", "_build_mlp_extractor": "<function ActorCriticPolicy._build_mlp_extractor at 0x793f0e8f13a0>", "_build": "<function ActorCriticPolicy._build at 0x793f0e8f1440>", "forward": "<function ActorCriticPolicy.forward at 0x793f0e8f14e0>", "extract_features": "<function ActorCriticPolicy.extract_features at 0x793f0e8f1580>", "_get_action_dist_from_latent": "<function ActorCriticPolicy._get_action_dist_from_latent at 0x793f0e8f1620>", "_predict": "<function ActorCriticPolicy._predict at 0x793f0e8f16c0>", "evaluate_actions": "<function ActorCriticPolicy.evaluate_actions at 0x793f0e8f1760>", "get_distribution": "<function ActorCriticPolicy.get_distribution at 0x793f0e8f1800>", "predict_values": "<function ActorCriticPolicy.predict_values at 0x793f0e8f18a0>", "__abstractmethods__": "frozenset()", "_abc_impl": "<_abc._abc_data object at 0x793f0e9f6340>"}, "verbose": 1, "policy_kwargs": {}, "num_timesteps": 1015808, "_total_timesteps": 1000000.0, "_num_timesteps_at_start": 0, "seed": null, "action_noise": null, "start_time": 1749518909944734722, "learning_rate": 0.0003, "tensorboard_log": null, "_last_obs": {":type:": "<class 'numpy.ndarray'>", ":serialized:": "gAWVdgIAAAAAAACME251bXB5Ll9jb3JlLm51bWVyaWOUjAtfZnJvbWJ1ZmZlcpSTlCiWAAIAAAAAAACaGxa8bCSvu/p+Ir7/qeO9mRLvPMJtpz4AAIA/AACAP4DEVT0uS8g7KBC4PdCUQb7Yjfs8HAsIvQAAAAAAAAAAzbSqvFIQuLnLbw28/g1YOxQyp7hP3Qe8AACAPwAAgD+gJiY+xRoEP8lojr13Doa+10qCPSV/wD0AAAAAAAAAAC2fS77aYA8/fKPJPv0fg74q+Fu8B4G2OwAAAAAAAAAAM9sxO19hKD62W3e+72uEvk0sgb214IO9AAAAAAAAAAC6BGk+uId+P+JMuL2k6Li+Ll95PT6PZrsAAAAAAAAAALNpO73sIaG5zQAaum/DxbS11Z+7cHg2OQAAgD8AAIA/baduPuRzZj82GFM9cWSdvpZt7j076mW9AAAAAAAAAADg/DW+RyYcP2VuTT4LDGC+L1PAvA6o+z0AAAAAAAAAAM0HmTwi01E+HuqGPUEFmr4Q67o8qE+rPQAAAAAAAAAAU2cwPkcEID/wskS+5Y+IviIzh7zKS569AAAAAAAAAAD6iBQ+56pZPouqVb7JYoO+qGU9vYKpBD0AAAAAAAAAABrNLj1DFxk9lbNxPb7HSb4KHbM8VmebPQAAAAAAAAAAZuC1vLjHsj+iPoO+sO0wvisAerwSbOy9AAAAAAAAAAAADq+9qmNUP8KkZz0D1Ya++pt6vXObUT0AAAAAAAAAAJSMBW51bXB5lIwFZHR5cGWUk5SMAmY0lImIh5RSlChLA4wBPJROTk5K/////0r/////SwB0lGJLEEsIhpSMAUOUdJRSlC4="}, "_last_episode_starts": {":type:": "<class 'numpy.ndarray'>", ":serialized:": "gAWVhAAAAAAAAACME251bXB5Ll9jb3JlLm51bWVyaWOUjAtfZnJvbWJ1ZmZlcpSTlCiWEAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAAlIwFbnVtcHmUjAVkdHlwZZSTlIwCYjGUiYiHlFKUKEsDjAF8lE5OTkr/////Sv////9LAHSUYksQhZSMAUOUdJRSlC4="}, "_last_original_obs": null, "_episode_num": 0, "use_sde": false, "sde_sample_freq": -1, "_current_progress_remaining": -0.015808000000000044, "_stats_window_size": 100, "ep_info_buffer": {":type:": "<class 'collections.deque'>", ":serialized:": "gAWVQwwAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKUKH2UKIwBcpRHQHGYYvN/vv2MAWyUTaoBjAF0lEdAkEaw4n4O+nV9lChoBkdAcBbsUIsyz2gHTaQCaAhHQJBH470WdmR1fZQoaAZHQGt/9i+cpb5oB02jAWgIR0CQSPgjyFwldX2UKGgGR0Bw8ObmU4aQaAdNfAFoCEdAkEl1Id2gWnV9lChoBkdAb0GYEW69TWgHTfcCaAhHQJBJzltCRfZ1fZQoaAZHQG8PjQ7cO9ZoB02HAmgIR0CQSnWszVMFdX2UKGgGR0BxiUldC3PSaAdNrgJoCEdAkEsIVuaWonV9lChoBkdAcG9+sHSncmgHTa4CaAhHQJBLoXizcAR1fZQoaAZHQGyU+yquKXRoB01pAmgIR0CQS7JkoWpIdX2UKGgGR0BcSclolD4QaAdN6ANoCEdAkEwx37k4m3V9lChoBkdAcSYFYuCf6GgHTZsBaAhHQJBNPwDvE0l1fZQoaAZHQGZC6MaS9uhoB03oA2gIR0CQTzFAVwgldX2UKGgGR0ByGv8qFyq/aAdNzwFoCEdAkE+KIznA7HV9lChoBkdAcAOH6dlNDmgHTSgBaAhHQJBQyzTnaFp1fZQoaAZHQHI7tNnGsFNoB02uAWgIR0CQUWdBjWkKdX2UKGgGR0BwsXcwg1WKaAdNAgJoCEdAkFVtKujh1nV9lChoBkdAcoVXZoPCmGgHTRIBaAhHQJBWLEm6XjV1fZQoaAZHQG2zVXvH93toB02wAWgIR0CQXDJ2t+1CdX2UKGgGR0ByIFmFrVOLaAdNAAJoCEdAkF3AUDdP+HV9lChoBkdAcH7ILw4KhWgHTV0BaAhHQJBjjKMefZp1fZQoaAZHQHF5F+mWMS9oB005AWgIR0CQY/UKRdQgdX2UKGgGR0BwT2l3yI56aAdNLgJoCEdAkHavLTx5LXV9lChoBkdAcMN7gKnei2gHTawBaAhHQJB5FnGsFMZ1fZQoaAZHQHAOBSxZ+x5oB001AmgIR0CQeXNcGC7LdX2UKGgGR0Buvm+h4+r3aAdNQwFoCEdAkHyRa1TisHV9lChoBkdAcexzo2XLNmgHTZ8CaAhHQJB9FOVPepJ1fZQoaAZHQHBcbAHmig1oB02sAmgIR0CQfU7L+xW1dX2UKGgGR0BwFQFUyYXwaAdNbAFoCEdAkH2s8cMmW3V9lChoBkdAceBXFtKqXGgHTfACaAhHQJB9y6Ymb9Z1fZQoaAZHQHI0tLlFMIxoB02VAmgIR0CQfowblzU7dX2UKGgGR0BwphDgIhQnaAdN7gFoCEdAkH7aQ3gk1XV9lChoBkdAcYLWkJrtV2gHTaYCaAhHQJB/CZE2Hcl1fZQoaAZHQHLE/pljEvVoB02aAmgIR0CQgD0DEFW5dX2UKGgGR0BxE88xKxs3aAdNqwFoCEdAkISSIDYAbXV9lChoBkdAb0ycz67/XGgHTecBaAhHQJCLVj8UEgZ1fZQoaAZHQHDcSaEzwc5oB01JAWgIR0CQi1oc7yQQdX2UKGgGR0Bv4wPXkHUuaAdNJAJoCEdAkIxNEb5uZXV9lChoBkdAcX03vQWvbGgHTU4BaAhHQJCMh44ZMtd1fZQoaAZHQHLesdo371toB00DAmgIR0CQjujyFwkxdX2UKGgGR0BxmJfeDWbxaAdNcgFoCEdAkI76rBCUo3V9lChoBkdAcktqjafzz2gHTXcBaAhHQJCQysDGLk11fZQoaAZHQG20Np22XsxoB01CAWgIR0CQkOzl90A+dX2UKGgGR0Buj0b1h9b5aAdNcQFoCEdAkJD8VtXPq3V9lChoBkdAcGOc2R7qp2gHTXEBaAhHQJCRR02cawV1fZQoaAZHQHIxwKKHfuVoB03hAWgIR0CQlGQTEit8dX2UKGgGR0BuSiAnUlRhaAdNAAJoCEdAkJWFivxH5XV9lChoBkdAcbjIN3GGVWgHTY8CaAhHQJCYhooNNJx1fZQoaAZHQGv5h+nZTQ5oB02fAmgIR0CQmYajvd/KdX2UKGgGR0Btzp8neBQOaAdNCANoCEdAkJq4BmwqzHV9lChoBkdAcCtjk+5e7mgHTV0BaAhHQJCflf4REnd1fZQoaAZHQHC/WjGkvbpoB01EAWgIR0CQoH8kUsWgdX2UKGgGR0BxiFZq20AtaAdN/wFoCEdAkKFNihFmWnV9lChoBkdAcjq/yGzrvGgHTVIBaAhHQJCkQdOqNqB1fZQoaAZHQHEENjCpFThoB02rAWgIR0CQpJG2CulodX2UKGgGR0BwsRs3yZrpaAdNfAFoCEdAkKaen62v0XV9lChoBkdAcBsj1PFefWgHTZgBaAhHQJCouiGnGbV1fZQoaAZHQHAPIsAeaKFoB01lAWgIR0CQqqRZlnRLdX2UKGgGR0BtL9xlxwQ2aAdNggFoCEdAkKsN1IRRM3V9lChoBkdAbZ0S+xnnMmgHTQMCaAhHQJC+3jbSJCV1fZQoaAZHQHDJejqOcUdoB01tAWgIR0CQwSX18LKFdX2UKGgGR0BtNksBhhH9aAdNJAFoCEdAkMWvMjeKsXV9lChoBkdAcDT11nuiOGgHTbUCaAhHQJDGvsa86FN1fZQoaAZHQGzjWDHwPRRoB01mAmgIR0CQxtmHxjJ/dX2UKGgGR0ByppxzaK1paAdNWwFoCEdAkMe2AG0NSnV9lChoBkdAbvC0ngHeJ2gHTdQCaAhHQJDI8pEx7At1fZQoaAZHQHACNQ0oBq9oB01CAWgIR0CQySMhouf3dX2UKGgGR0Bu/TqQiiZfaAdN1wFoCEdAkMn/8dgfEHV9lChoBkdAckpl/YraumgHTYkBaAhHQJDKNScbzbx1fZQoaAZHQHHfwzHjp9toB00eAWgIR0CQysprk8zRdX2UKGgGR0BvPKyOaOPvaAdNRAFoCEdAkMrUvCdjG3V9lChoBkdAcLTAZKnNxGgHTQwCaAhHQJDK8CDEm6Z1fZQoaAZHQGxiPZh8YyhoB03QAWgIR0CQzkc9nscAdX2UKGgGR0ByBDevZAY6aAdNKQFoCEdAkM6WY4Qz13V9lChoBkdARQ+0E5hjOWgHS+NoCEdAkM6YjW07bXV9lChoBkdAcmTwKBun/GgHTZoBaAhHQJDPtTJhfBx1fZQoaAZHQHIf0gGKQ7toB02bAWgIR0CQ0Sq7ROUMdX2UKGgGR0AvbW+XZ5AyaAdNGgFoCEdAkNHkHQhOg3V9lChoBkdAcqHj4pMHr2gHTdMBaAhHQJDSHNHH3lF1fZQoaAZHQG/lupbUwztoB01EAWgIR0CQ0sa6z3RHdX2UKGgGR0Bt4oFiay8jaAdNeAFoCEdAkNS1JpWV/3V9lChoBkdAcQMAU+LWJGgHTWYBaAhHQJDV49Net0V1fZQoaAZHQHI3DzVc2R9oB010AWgIR0CQ1kR0EHMVdX2UKGgGR0Bw+DTx5LRKaAdNdwFoCEdAkNg12vB7/nV9lChoBkdAcAIocaOxS2gHTasBaAhHQJDZ3nwG4Zx1fZQoaAZHQHAa2uxKQJZoB03aAWgIR0CQ27kBjnV5dX2UKGgGR0Bx5IJzDGcXaAdNygFoCEdAkNvtbs4T9XV9lChoBkdAbjG+cpb2UWgHTeQBaAhHQJDdMQe3hGZ1fZQoaAZHQHDqGpMpPRBoB02AAWgIR0CQ32WN3np0dX2UKGgGR0BuKHVPN3W4aAdNUAFoCEdAkN/0BjnV5XV9lChoBkdAcIXYYBNmDmgHTbUBaAhHQJDgPrD63y91fZQoaAZHQHDwGVAzHjpoB00YAWgIR0CQ4ivegte2dX2UKGgGR0BueUdLg4wRaAdNgAFoCEdAkOJY6S1VpHV9lChoBkdAbJbuGbkOqmgHTQYCaAhHQJDja3fAKv51fZQoaAZHQHAOPm1YyO9oB020AWgIR0CQ44aef7JodX2UKGgGR0Bx/yQJXyRTaAdNDwJoCEdAkOQhaxHG0nV9lChoBkdAcR+WdEsrd2gHTYsBaAhHQJDle+ajN6h1fZQoaAZHQHCodKRMewNoB03DAWgIR0CQ5bDAJswddX2UKGgGR0BxZCLfk3juaAdNKwFoCEdAkOZ0GeMAFXV9lChoBkdAbREam4y44WgHTVUBaAhHQJDmn2oNutR1fZQoaAZHQHCWWXTmW+poB01NAWgIR0CQ6RBe5WildWUu"}, "ep_success_buffer": {":type:": "<class 'collections.deque'>", ":serialized:": "gAWVIAAAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKULg=="}, "_n_updates": 248, "observation_space": {":type:": "<class 'gymnasium.spaces.box.Box'>", ":serialized:": "gAWVdwIAAAAAAACMFGd5bW5hc2l1bS5zcGFjZXMuYm94lIwDQm94lJOUKYGUfZQojAVkdHlwZZSMBW51bXB5lIwFZHR5cGWUk5SMAmY0lImIh5RSlChLA4wBPJROTk5K/////0r/////SwB0lGKMDWJvdW5kZWRfYmVsb3eUjBNudW1weS5fY29yZS5udW1lcmljlIwLX2Zyb21idWZmZXKUk5QolggAAAAAAAAAAQEBAQEBAQGUaAiMAmIxlImIh5RSlChLA4wBfJROTk5K/////0r/////SwB0lGJLCIWUjAFDlHSUUpSMDWJvdW5kZWRfYWJvdmWUaBEolggAAAAAAAAAAQEBAQEBAQGUaBVLCIWUaBl0lFKUjAZfc2hhcGWUSwiFlIwDbG93lGgRKJYgAAAAAAAAAAAAtMIAALTCAACgwAAAoMDbD0nAAACgwAAAAIAAAACAlGgLSwiFlGgZdJRSlIwEaGlnaJRoESiWIAAAAAAAAAAAALRCAAC0QgAAoEAAAKBA2w9JQAAAoEAAAIA/AACAP5RoC0sIhZRoGXSUUpSMCGxvd19yZXBylIxbWy05MC4gICAgICAgIC05MC4gICAgICAgICAtNS4gICAgICAgICAtNS4gICAgICAgICAtMy4xNDE1OTI3ICAtNS4KICAtMC4gICAgICAgICAtMC4gICAgICAgXZSMCWhpZ2hfcmVwcpSMU1s5MC4gICAgICAgIDkwLiAgICAgICAgIDUuICAgICAgICAgNS4gICAgICAgICAzLjE0MTU5MjcgIDUuCiAgMS4gICAgICAgICAxLiAgICAgICBdlIwKX25wX3JhbmRvbZROdWIu", "dtype": "float32", "bounded_below": "[ True True True True True True True True]", "bounded_above": "[ True True True True True True True True]", "_shape": [8], "low": "[-90. -90. -5. -5. -3.1415927 -5.\n -0. -0. ]", "high": "[90. 90. 5. 5. 3.1415927 5.\n 1. 1. ]", "low_repr": "[-90. -90. -5. -5. -3.1415927 -5.\n -0. -0. ]", "high_repr": "[90. 90. 5. 5. 3.1415927 5.\n 1. 1. ]", "_np_random": null}, "action_space": {":type:": "<class 'gymnasium.spaces.discrete.Discrete'>", ":serialized:": "gAWV3AAAAAAAAACMGWd5bW5hc2l1bS5zcGFjZXMuZGlzY3JldGWUjAhEaXNjcmV0ZZSTlCmBlH2UKIwBbpSMFm51bXB5Ll9jb3JlLm11bHRpYXJyYXmUjAZzY2FsYXKUk5SMBW51bXB5lIwFZHR5cGWUk5SMAmk4lImIh5RSlChLA4wBPJROTk5K/////0r/////SwB0lGJDCAQAAAAAAAAAlIaUUpSMBXN0YXJ0lGgIaA5DCAAAAAAAAAAAlIaUUpSMBl9zaGFwZZQpjAVkdHlwZZRoDowKX25wX3JhbmRvbZROdWIu", "n": "4", "start": "0", "_shape": [], "dtype": "int64", "_np_random": null}, "n_envs": 16, "n_steps": 1024, "gamma": 0.999, "gae_lambda": 0.98, "ent_coef": 0.01, "vf_coef": 0.5, "max_grad_norm": 0.5, "batch_size": 64, "n_epochs": 4, "clip_range": {":type:": "<class 'function'>", ":serialized:": "gAWV1gIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwOX21ha2VfZnVuY3Rpb26Uk5QoaACMDV9idWlsdGluX3R5cGWUk5SMCENvZGVUeXBllIWUUpQoSwFLAEsASwFLAUsTQwiVAZcAiQFTAJROhZQpjAFflIWUjEkvdXNyL2xvY2FsL2xpYi9weXRob24zLjExL2Rpc3QtcGFja2FnZXMvc3RhYmxlX2Jhc2VsaW5lczMvY29tbW9uL3V0aWxzLnB5lIwEZnVuY5SMGWNvbnN0YW50X2ZuLjxsb2NhbHM+LmZ1bmOUS4RDCPiAANgPEogKlEMAlIwDdmFslIWUKXSUUpR9lCiMC19fcGFja2FnZV9flIwYc3RhYmxlX2Jhc2VsaW5lczMuY29tbW9ulIwIX19uYW1lX1+UjB5zdGFibGVfYmFzZWxpbmVzMy5jb21tb24udXRpbHOUjAhfX2ZpbGVfX5SMSS91c3IvbG9jYWwvbGliL3B5dGhvbjMuMTEvZGlzdC1wYWNrYWdlcy9zdGFibGVfYmFzZWxpbmVzMy9jb21tb24vdXRpbHMucHmUdU5OaACMEF9tYWtlX2VtcHR5X2NlbGyUk5QpUpSFlHSUUpRoAIwSX2Z1bmN0aW9uX3NldHN0YXRllJOUaCF9lH2UKGgYjARmdW5jlIwMX19xdWFsbmFtZV9flIwZY29uc3RhbnRfZm4uPGxvY2Fscz4uZnVuY5SMD19fYW5ub3RhdGlvbnNfX5R9lIwOX19rd2RlZmF1bHRzX1+UTowMX19kZWZhdWx0c19flE6MCl9fbW9kdWxlX1+UaBmMB19fZG9jX1+UTowLX19jbG9zdXJlX1+UaACMCl9tYWtlX2NlbGyUk5RHP8mZmZmZmZqFlFKUhZSMF19jbG91ZHBpY2tsZV9zdWJtb2R1bGVzlF2UjAtfX2dsb2JhbHNfX5R9lHWGlIZSMC4="}, "clip_range_vf": null, "normalize_advantage": true, "target_kl": null, "lr_schedule": {":type:": "<class 'function'>", ":serialized:": "gAWV1gIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwOX21ha2VfZnVuY3Rpb26Uk5QoaACMDV9idWlsdGluX3R5cGWUk5SMCENvZGVUeXBllIWUUpQoSwFLAEsASwFLAUsTQwiVAZcAiQFTAJROhZQpjAFflIWUjEkvdXNyL2xvY2FsL2xpYi9weXRob24zLjExL2Rpc3QtcGFja2FnZXMvc3RhYmxlX2Jhc2VsaW5lczMvY29tbW9uL3V0aWxzLnB5lIwEZnVuY5SMGWNvbnN0YW50X2ZuLjxsb2NhbHM+LmZ1bmOUS4RDCPiAANgPEogKlEMAlIwDdmFslIWUKXSUUpR9lCiMC19fcGFja2FnZV9flIwYc3RhYmxlX2Jhc2VsaW5lczMuY29tbW9ulIwIX19uYW1lX1+UjB5zdGFibGVfYmFzZWxpbmVzMy5jb21tb24udXRpbHOUjAhfX2ZpbGVfX5SMSS91c3IvbG9jYWwvbGliL3B5dGhvbjMuMTEvZGlzdC1wYWNrYWdlcy9zdGFibGVfYmFzZWxpbmVzMy9jb21tb24vdXRpbHMucHmUdU5OaACMEF9tYWtlX2VtcHR5X2NlbGyUk5QpUpSFlHSUUpRoAIwSX2Z1bmN0aW9uX3NldHN0YXRllJOUaCF9lH2UKGgYjARmdW5jlIwMX19xdWFsbmFtZV9flIwZY29uc3RhbnRfZm4uPGxvY2Fscz4uZnVuY5SMD19fYW5ub3RhdGlvbnNfX5R9lIwOX19rd2RlZmF1bHRzX1+UTowMX19kZWZhdWx0c19flE6MCl9fbW9kdWxlX1+UaBmMB19fZG9jX1+UTowLX19jbG9zdXJlX1+UaACMCl9tYWtlX2NlbGyUk5RHPzOpKjBVMmGFlFKUhZSMF19jbG91ZHBpY2tsZV9zdWJtb2R1bGVzlF2UjAtfX2dsb2JhbHNfX5R9lHWGlIZSMC4="}, "system_info": {"OS": "Linux-6.1.123+-x86_64-with-glibc2.35 # 1 SMP PREEMPT_DYNAMIC Sun Mar 30 16:01:29 UTC 2025", "Python": "3.11.13", "Stable-Baselines3": "2.0.0a5", "PyTorch": "2.6.0+cu124", "GPU Enabled": "True", "Numpy": "2.0.2", "Cloudpickle": "3.1.1", "Gymnasium": "0.28.1", "OpenAI Gym": "0.25.2"}}
|
ppo-LunarLander-v2.zip
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:016bb4d3f3c3b4bfc3a223241d2e8c3ae72c41f99ba8f12a925bb5a3c97d28b8
|
| 3 |
+
size 148134
|
ppo-LunarLander-v2/data
CHANGED
|
@@ -4,20 +4,20 @@
|
|
| 4 |
":serialized:": "gAWVOwAAAAAAAACMIXN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi5wb2xpY2llc5SMEUFjdG9yQ3JpdGljUG9saWN5lJOULg==",
|
| 5 |
"__module__": "stable_baselines3.common.policies",
|
| 6 |
"__doc__": "\n Policy class for actor-critic algorithms (has both policy and value prediction).\n Used by A2C, PPO and the likes.\n\n :param observation_space: Observation space\n :param action_space: Action space\n :param lr_schedule: Learning rate schedule (could be constant)\n :param net_arch: The specification of the policy and value networks.\n :param activation_fn: Activation function\n :param ortho_init: Whether to use or not orthogonal initialization\n :param use_sde: Whether to use State Dependent Exploration or not\n :param log_std_init: Initial value for the log standard deviation\n :param full_std: Whether to use (n_features x n_actions) parameters\n for the std instead of only (n_features,) when using gSDE\n :param use_expln: Use ``expln()`` function instead of ``exp()`` to ensure\n a positive standard deviation (cf paper). It allows to keep variance\n above zero and prevent it from growing too fast. In practice, ``exp()`` is usually enough.\n :param squash_output: Whether to squash the output using a tanh function,\n this allows to ensure boundaries when using gSDE.\n :param features_extractor_class: Features extractor to use.\n :param features_extractor_kwargs: Keyword arguments\n to pass to the features extractor.\n :param share_features_extractor: If True, the features extractor is shared between the policy and value networks.\n :param normalize_images: Whether to normalize images or not,\n dividing by 255.0 (True by default)\n :param optimizer_class: The optimizer to use,\n ``th.optim.Adam`` by default\n :param optimizer_kwargs: Additional keyword arguments,\n excluding the learning rate, to pass to the optimizer\n ",
|
| 7 |
-
"__init__": "<function ActorCriticPolicy.__init__ at
|
| 8 |
-
"_get_constructor_parameters": "<function ActorCriticPolicy._get_constructor_parameters at
|
| 9 |
-
"reset_noise": "<function ActorCriticPolicy.reset_noise at
|
| 10 |
-
"_build_mlp_extractor": "<function ActorCriticPolicy._build_mlp_extractor at
|
| 11 |
-
"_build": "<function ActorCriticPolicy._build at
|
| 12 |
-
"forward": "<function ActorCriticPolicy.forward at
|
| 13 |
-
"extract_features": "<function ActorCriticPolicy.extract_features at
|
| 14 |
-
"_get_action_dist_from_latent": "<function ActorCriticPolicy._get_action_dist_from_latent at
|
| 15 |
-
"_predict": "<function ActorCriticPolicy._predict at
|
| 16 |
-
"evaluate_actions": "<function ActorCriticPolicy.evaluate_actions at
|
| 17 |
-
"get_distribution": "<function ActorCriticPolicy.get_distribution at
|
| 18 |
-
"predict_values": "<function ActorCriticPolicy.predict_values at
|
| 19 |
"__abstractmethods__": "frozenset()",
|
| 20 |
-
"_abc_impl": "<_abc._abc_data object at
|
| 21 |
},
|
| 22 |
"verbose": 1,
|
| 23 |
"policy_kwargs": {},
|
|
@@ -26,12 +26,12 @@
|
|
| 26 |
"_num_timesteps_at_start": 0,
|
| 27 |
"seed": null,
|
| 28 |
"action_noise": null,
|
| 29 |
-
"start_time":
|
| 30 |
"learning_rate": 0.0003,
|
| 31 |
"tensorboard_log": null,
|
| 32 |
"_last_obs": {
|
| 33 |
":type:": "<class 'numpy.ndarray'>",
|
| 34 |
-
":serialized:": "
|
| 35 |
},
|
| 36 |
"_last_episode_starts": {
|
| 37 |
":type:": "<class 'numpy.ndarray'>",
|
|
@@ -45,7 +45,7 @@
|
|
| 45 |
"_stats_window_size": 100,
|
| 46 |
"ep_info_buffer": {
|
| 47 |
":type:": "<class 'collections.deque'>",
|
| 48 |
-
":serialized:": "
|
| 49 |
},
|
| 50 |
"ep_success_buffer": {
|
| 51 |
":type:": "<class 'collections.deque'>",
|
|
|
|
| 4 |
":serialized:": "gAWVOwAAAAAAAACMIXN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi5wb2xpY2llc5SMEUFjdG9yQ3JpdGljUG9saWN5lJOULg==",
|
| 5 |
"__module__": "stable_baselines3.common.policies",
|
| 6 |
"__doc__": "\n Policy class for actor-critic algorithms (has both policy and value prediction).\n Used by A2C, PPO and the likes.\n\n :param observation_space: Observation space\n :param action_space: Action space\n :param lr_schedule: Learning rate schedule (could be constant)\n :param net_arch: The specification of the policy and value networks.\n :param activation_fn: Activation function\n :param ortho_init: Whether to use or not orthogonal initialization\n :param use_sde: Whether to use State Dependent Exploration or not\n :param log_std_init: Initial value for the log standard deviation\n :param full_std: Whether to use (n_features x n_actions) parameters\n for the std instead of only (n_features,) when using gSDE\n :param use_expln: Use ``expln()`` function instead of ``exp()`` to ensure\n a positive standard deviation (cf paper). It allows to keep variance\n above zero and prevent it from growing too fast. In practice, ``exp()`` is usually enough.\n :param squash_output: Whether to squash the output using a tanh function,\n this allows to ensure boundaries when using gSDE.\n :param features_extractor_class: Features extractor to use.\n :param features_extractor_kwargs: Keyword arguments\n to pass to the features extractor.\n :param share_features_extractor: If True, the features extractor is shared between the policy and value networks.\n :param normalize_images: Whether to normalize images or not,\n dividing by 255.0 (True by default)\n :param optimizer_class: The optimizer to use,\n ``th.optim.Adam`` by default\n :param optimizer_kwargs: Additional keyword arguments,\n excluding the learning rate, to pass to the optimizer\n ",
|
| 7 |
+
"__init__": "<function ActorCriticPolicy.__init__ at 0x793f0e8f11c0>",
|
| 8 |
+
"_get_constructor_parameters": "<function ActorCriticPolicy._get_constructor_parameters at 0x793f0e8f1260>",
|
| 9 |
+
"reset_noise": "<function ActorCriticPolicy.reset_noise at 0x793f0e8f1300>",
|
| 10 |
+
"_build_mlp_extractor": "<function ActorCriticPolicy._build_mlp_extractor at 0x793f0e8f13a0>",
|
| 11 |
+
"_build": "<function ActorCriticPolicy._build at 0x793f0e8f1440>",
|
| 12 |
+
"forward": "<function ActorCriticPolicy.forward at 0x793f0e8f14e0>",
|
| 13 |
+
"extract_features": "<function ActorCriticPolicy.extract_features at 0x793f0e8f1580>",
|
| 14 |
+
"_get_action_dist_from_latent": "<function ActorCriticPolicy._get_action_dist_from_latent at 0x793f0e8f1620>",
|
| 15 |
+
"_predict": "<function ActorCriticPolicy._predict at 0x793f0e8f16c0>",
|
| 16 |
+
"evaluate_actions": "<function ActorCriticPolicy.evaluate_actions at 0x793f0e8f1760>",
|
| 17 |
+
"get_distribution": "<function ActorCriticPolicy.get_distribution at 0x793f0e8f1800>",
|
| 18 |
+
"predict_values": "<function ActorCriticPolicy.predict_values at 0x793f0e8f18a0>",
|
| 19 |
"__abstractmethods__": "frozenset()",
|
| 20 |
+
"_abc_impl": "<_abc._abc_data object at 0x793f0e9f6340>"
|
| 21 |
},
|
| 22 |
"verbose": 1,
|
| 23 |
"policy_kwargs": {},
|
|
|
|
| 26 |
"_num_timesteps_at_start": 0,
|
| 27 |
"seed": null,
|
| 28 |
"action_noise": null,
|
| 29 |
+
"start_time": 1749518909944734722,
|
| 30 |
"learning_rate": 0.0003,
|
| 31 |
"tensorboard_log": null,
|
| 32 |
"_last_obs": {
|
| 33 |
":type:": "<class 'numpy.ndarray'>",
|
| 34 |
+
":serialized:": "gAWVdgIAAAAAAACME251bXB5Ll9jb3JlLm51bWVyaWOUjAtfZnJvbWJ1ZmZlcpSTlCiWAAIAAAAAAACaGxa8bCSvu/p+Ir7/qeO9mRLvPMJtpz4AAIA/AACAP4DEVT0uS8g7KBC4PdCUQb7Yjfs8HAsIvQAAAAAAAAAAzbSqvFIQuLnLbw28/g1YOxQyp7hP3Qe8AACAPwAAgD+gJiY+xRoEP8lojr13Doa+10qCPSV/wD0AAAAAAAAAAC2fS77aYA8/fKPJPv0fg74q+Fu8B4G2OwAAAAAAAAAAM9sxO19hKD62W3e+72uEvk0sgb214IO9AAAAAAAAAAC6BGk+uId+P+JMuL2k6Li+Ll95PT6PZrsAAAAAAAAAALNpO73sIaG5zQAaum/DxbS11Z+7cHg2OQAAgD8AAIA/baduPuRzZj82GFM9cWSdvpZt7j076mW9AAAAAAAAAADg/DW+RyYcP2VuTT4LDGC+L1PAvA6o+z0AAAAAAAAAAM0HmTwi01E+HuqGPUEFmr4Q67o8qE+rPQAAAAAAAAAAU2cwPkcEID/wskS+5Y+IviIzh7zKS569AAAAAAAAAAD6iBQ+56pZPouqVb7JYoO+qGU9vYKpBD0AAAAAAAAAABrNLj1DFxk9lbNxPb7HSb4KHbM8VmebPQAAAAAAAAAAZuC1vLjHsj+iPoO+sO0wvisAerwSbOy9AAAAAAAAAAAADq+9qmNUP8KkZz0D1Ya++pt6vXObUT0AAAAAAAAAAJSMBW51bXB5lIwFZHR5cGWUk5SMAmY0lImIh5RSlChLA4wBPJROTk5K/////0r/////SwB0lGJLEEsIhpSMAUOUdJRSlC4="
|
| 35 |
},
|
| 36 |
"_last_episode_starts": {
|
| 37 |
":type:": "<class 'numpy.ndarray'>",
|
|
|
|
| 45 |
"_stats_window_size": 100,
|
| 46 |
"ep_info_buffer": {
|
| 47 |
":type:": "<class 'collections.deque'>",
|
| 48 |
+
":serialized:": "gAWVQwwAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKUKH2UKIwBcpRHQHGYYvN/vv2MAWyUTaoBjAF0lEdAkEaw4n4O+nV9lChoBkdAcBbsUIsyz2gHTaQCaAhHQJBH470WdmR1fZQoaAZHQGt/9i+cpb5oB02jAWgIR0CQSPgjyFwldX2UKGgGR0Bw8ObmU4aQaAdNfAFoCEdAkEl1Id2gWnV9lChoBkdAb0GYEW69TWgHTfcCaAhHQJBJzltCRfZ1fZQoaAZHQG8PjQ7cO9ZoB02HAmgIR0CQSnWszVMFdX2UKGgGR0BxiUldC3PSaAdNrgJoCEdAkEsIVuaWonV9lChoBkdAcG9+sHSncmgHTa4CaAhHQJBLoXizcAR1fZQoaAZHQGyU+yquKXRoB01pAmgIR0CQS7JkoWpIdX2UKGgGR0BcSclolD4QaAdN6ANoCEdAkEwx37k4m3V9lChoBkdAcSYFYuCf6GgHTZsBaAhHQJBNPwDvE0l1fZQoaAZHQGZC6MaS9uhoB03oA2gIR0CQTzFAVwgldX2UKGgGR0ByGv8qFyq/aAdNzwFoCEdAkE+KIznA7HV9lChoBkdAcAOH6dlNDmgHTSgBaAhHQJBQyzTnaFp1fZQoaAZHQHI7tNnGsFNoB02uAWgIR0CQUWdBjWkKdX2UKGgGR0BwsXcwg1WKaAdNAgJoCEdAkFVtKujh1nV9lChoBkdAcoVXZoPCmGgHTRIBaAhHQJBWLEm6XjV1fZQoaAZHQG2zVXvH93toB02wAWgIR0CQXDJ2t+1CdX2UKGgGR0ByIFmFrVOLaAdNAAJoCEdAkF3AUDdP+HV9lChoBkdAcH7ILw4KhWgHTV0BaAhHQJBjjKMefZp1fZQoaAZHQHF5F+mWMS9oB005AWgIR0CQY/UKRdQgdX2UKGgGR0BwT2l3yI56aAdNLgJoCEdAkHavLTx5LXV9lChoBkdAcMN7gKnei2gHTawBaAhHQJB5FnGsFMZ1fZQoaAZHQHAOBSxZ+x5oB001AmgIR0CQeXNcGC7LdX2UKGgGR0Buvm+h4+r3aAdNQwFoCEdAkHyRa1TisHV9lChoBkdAcexzo2XLNmgHTZ8CaAhHQJB9FOVPepJ1fZQoaAZHQHBcbAHmig1oB02sAmgIR0CQfU7L+xW1dX2UKGgGR0BwFQFUyYXwaAdNbAFoCEdAkH2s8cMmW3V9lChoBkdAceBXFtKqXGgHTfACaAhHQJB9y6Ymb9Z1fZQoaAZHQHI0tLlFMIxoB02VAmgIR0CQfowblzU7dX2UKGgGR0BwphDgIhQnaAdN7gFoCEdAkH7aQ3gk1XV9lChoBkdAcYLWkJrtV2gHTaYCaAhHQJB/CZE2Hcl1fZQoaAZHQHLE/pljEvVoB02aAmgIR0CQgD0DEFW5dX2UKGgGR0BxE88xKxs3aAdNqwFoCEdAkISSIDYAbXV9lChoBkdAb0ycz67/XGgHTecBaAhHQJCLVj8UEgZ1fZQoaAZHQHDcSaEzwc5oB01JAWgIR0CQi1oc7yQQdX2UKGgGR0Bv4wPXkHUuaAdNJAJoCEdAkIxNEb5uZXV9lChoBkdAcX03vQWvbGgHTU4BaAhHQJCMh44ZMtd1fZQoaAZHQHLesdo371toB00DAmgIR0CQjujyFwkxdX2UKGgGR0BxmJfeDWbxaAdNcgFoCEdAkI76rBCUo3V9lChoBkdAcktqjafzz2gHTXcBaAhHQJCQysDGLk11fZQoaAZHQG20Np22XsxoB01CAWgIR0CQkOzl90A+dX2UKGgGR0Buj0b1h9b5aAdNcQFoCEdAkJD8VtXPq3V9lChoBkdAcGOc2R7qp2gHTXEBaAhHQJCRR02cawV1fZQoaAZHQHIxwKKHfuVoB03hAWgIR0CQlGQTEit8dX2UKGgGR0BuSiAnUlRhaAdNAAJoCEdAkJWFivxH5XV9lChoBkdAcbjIN3GGVWgHTY8CaAhHQJCYhooNNJx1fZQoaAZHQGv5h+nZTQ5oB02fAmgIR0CQmYajvd/KdX2UKGgGR0Btzp8neBQOaAdNCANoCEdAkJq4BmwqzHV9lChoBkdAcCtjk+5e7mgHTV0BaAhHQJCflf4REnd1fZQoaAZHQHC/WjGkvbpoB01EAWgIR0CQoH8kUsWgdX2UKGgGR0BxiFZq20AtaAdN/wFoCEdAkKFNihFmWnV9lChoBkdAcjq/yGzrvGgHTVIBaAhHQJCkQdOqNqB1fZQoaAZHQHEENjCpFThoB02rAWgIR0CQpJG2CulodX2UKGgGR0BwsRs3yZrpaAdNfAFoCEdAkKaen62v0XV9lChoBkdAcBsj1PFefWgHTZgBaAhHQJCouiGnGbV1fZQoaAZHQHAPIsAeaKFoB01lAWgIR0CQqqRZlnRLdX2UKGgGR0BtL9xlxwQ2aAdNggFoCEdAkKsN1IRRM3V9lChoBkdAbZ0S+xnnMmgHTQMCaAhHQJC+3jbSJCV1fZQoaAZHQHDJejqOcUdoB01tAWgIR0CQwSX18LKFdX2UKGgGR0BtNksBhhH9aAdNJAFoCEdAkMWvMjeKsXV9lChoBkdAcDT11nuiOGgHTbUCaAhHQJDGvsa86FN1fZQoaAZHQGzjWDHwPRRoB01mAmgIR0CQxtmHxjJ/dX2UKGgGR0ByppxzaK1paAdNWwFoCEdAkMe2AG0NSnV9lChoBkdAbvC0ngHeJ2gHTdQCaAhHQJDI8pEx7At1fZQoaAZHQHACNQ0oBq9oB01CAWgIR0CQySMhouf3dX2UKGgGR0Bu/TqQiiZfaAdN1wFoCEdAkMn/8dgfEHV9lChoBkdAckpl/YraumgHTYkBaAhHQJDKNScbzbx1fZQoaAZHQHHfwzHjp9toB00eAWgIR0CQysprk8zRdX2UKGgGR0BvPKyOaOPvaAdNRAFoCEdAkMrUvCdjG3V9lChoBkdAcLTAZKnNxGgHTQwCaAhHQJDK8CDEm6Z1fZQoaAZHQGxiPZh8YyhoB03QAWgIR0CQzkc9nscAdX2UKGgGR0ByBDevZAY6aAdNKQFoCEdAkM6WY4Qz13V9lChoBkdARQ+0E5hjOWgHS+NoCEdAkM6YjW07bXV9lChoBkdAcmTwKBun/GgHTZoBaAhHQJDPtTJhfBx1fZQoaAZHQHIf0gGKQ7toB02bAWgIR0CQ0Sq7ROUMdX2UKGgGR0AvbW+XZ5AyaAdNGgFoCEdAkNHkHQhOg3V9lChoBkdAcqHj4pMHr2gHTdMBaAhHQJDSHNHH3lF1fZQoaAZHQG/lupbUwztoB01EAWgIR0CQ0sa6z3RHdX2UKGgGR0Bt4oFiay8jaAdNeAFoCEdAkNS1JpWV/3V9lChoBkdAcQMAU+LWJGgHTWYBaAhHQJDV49Net0V1fZQoaAZHQHI3DzVc2R9oB010AWgIR0CQ1kR0EHMVdX2UKGgGR0Bw+DTx5LRKaAdNdwFoCEdAkNg12vB7/nV9lChoBkdAcAIocaOxS2gHTasBaAhHQJDZ3nwG4Zx1fZQoaAZHQHAa2uxKQJZoB03aAWgIR0CQ27kBjnV5dX2UKGgGR0Bx5IJzDGcXaAdNygFoCEdAkNvtbs4T9XV9lChoBkdAbjG+cpb2UWgHTeQBaAhHQJDdMQe3hGZ1fZQoaAZHQHDqGpMpPRBoB02AAWgIR0CQ32WN3np0dX2UKGgGR0BuKHVPN3W4aAdNUAFoCEdAkN/0BjnV5XV9lChoBkdAcIXYYBNmDmgHTbUBaAhHQJDgPrD63y91fZQoaAZHQHDwGVAzHjpoB00YAWgIR0CQ4ivegte2dX2UKGgGR0BueUdLg4wRaAdNgAFoCEdAkOJY6S1VpHV9lChoBkdAbJbuGbkOqmgHTQYCaAhHQJDja3fAKv51fZQoaAZHQHAOPm1YyO9oB020AWgIR0CQ44aef7JodX2UKGgGR0Bx/yQJXyRTaAdNDwJoCEdAkOQhaxHG0nV9lChoBkdAcR+WdEsrd2gHTYsBaAhHQJDle+ajN6h1fZQoaAZHQHCodKRMewNoB03DAWgIR0CQ5bDAJswddX2UKGgGR0BxZCLfk3juaAdNKwFoCEdAkOZ0GeMAFXV9lChoBkdAbREam4y44WgHTVUBaAhHQJDmn2oNutR1fZQoaAZHQHCWWXTmW+poB01NAWgIR0CQ6RBe5WildWUu"
|
| 49 |
},
|
| 50 |
"ep_success_buffer": {
|
| 51 |
":type:": "<class 'collections.deque'>",
|
ppo-LunarLander-v2/policy.optimizer.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 88362
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:267458ee500fa7d74019f859cdaf870407f50618fb10d3be44d02b9fc8bc3e80
|
| 3 |
size 88362
|
ppo-LunarLander-v2/policy.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 43762
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:da4bd9c0f4f2b2e137dd07e3fadf2a10f6f4298dd8e4a66e585a742b5cc39c38
|
| 3 |
size 43762
|
replay.mp4
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f9fdaee268f4e508a10325ab6a378f82d7019512be301c2e3ee385d6bdd96d84
|
| 3 |
+
size 170478
|
results.json
CHANGED
|
@@ -1 +1 @@
|
|
| 1 |
-
{"mean_reward":
|
|
|
|
| 1 |
+
{"mean_reward": 253.9088456, "std_reward": 20.890147745055817, "is_deterministic": true, "n_eval_episodes": 10, "eval_datetime": "2025-06-10T01:51:11.324683"}
|