Initial commit
Browse files- args.yml +3 -3
- ppo-seals-Walker2d-v0.zip +1 -1
- ppo-seals-Walker2d-v0/data +15 -15
- results.json +1 -1
- train_eval_metrics.zip +2 -2
args.yml
CHANGED
|
@@ -16,7 +16,7 @@
|
|
| 16 |
- - hyperparams
|
| 17 |
- null
|
| 18 |
- - log_folder
|
| 19 |
-
-
|
| 20 |
- - log_interval
|
| 21 |
- -1
|
| 22 |
- - max_total_trials
|
|
@@ -56,7 +56,7 @@
|
|
| 56 |
- - study_name
|
| 57 |
- null
|
| 58 |
- - tensorboard_log
|
| 59 |
-
- runs/seals/Walker2d-
|
| 60 |
- - track
|
| 61 |
- true
|
| 62 |
- - trained_agent
|
|
@@ -72,4 +72,4 @@
|
|
| 72 |
- - wandb_entity
|
| 73 |
- null
|
| 74 |
- - wandb_project_name
|
| 75 |
-
- seals-experts
|
|
|
|
| 16 |
- - hyperparams
|
| 17 |
- null
|
| 18 |
- - log_folder
|
| 19 |
+
- seals_experts_wandb_oldpickle/seed_5/
|
| 20 |
- - log_interval
|
| 21 |
- -1
|
| 22 |
- - max_total_trials
|
|
|
|
| 56 |
- - study_name
|
| 57 |
- null
|
| 58 |
- - tensorboard_log
|
| 59 |
+
- runs/seals/Walker2d-v0__ppo__5__1658860231
|
| 60 |
- - track
|
| 61 |
- true
|
| 62 |
- - trained_agent
|
|
|
|
| 72 |
- - wandb_entity
|
| 73 |
- null
|
| 74 |
- - wandb_project_name
|
| 75 |
+
- seals-experts-oldpickle
|
ppo-seals-Walker2d-v0.zip
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1753237
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e322057fa9c4c6540b8c1f145c9e8b5d13adb0c7801836fcc5465f86b53aed09
|
| 3 |
size 1753237
|
ppo-seals-Walker2d-v0/data
CHANGED
|
@@ -4,19 +4,19 @@
|
|
| 4 |
":serialized:": "gAWVOwAAAAAAAACMIXN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi5wb2xpY2llc5SMEUFjdG9yQ3JpdGljUG9saWN5lJOULg==",
|
| 5 |
"__module__": "stable_baselines3.common.policies",
|
| 6 |
"__doc__": "\n Policy class for actor-critic algorithms (has both policy and value prediction).\n Used by A2C, PPO and the likes.\n\n :param observation_space: Observation space\n :param action_space: Action space\n :param lr_schedule: Learning rate schedule (could be constant)\n :param net_arch: The specification of the policy and value networks.\n :param activation_fn: Activation function\n :param ortho_init: Whether to use or not orthogonal initialization\n :param use_sde: Whether to use State Dependent Exploration or not\n :param log_std_init: Initial value for the log standard deviation\n :param full_std: Whether to use (n_features x n_actions) parameters\n for the std instead of only (n_features,) when using gSDE\n :param sde_net_arch: Network architecture for extracting features\n when using gSDE. If None, the latent features from the policy will be used.\n Pass an empty list to use the states as features.\n :param use_expln: Use ``expln()`` function instead of ``exp()`` to ensure\n a positive standard deviation (cf paper). It allows to keep variance\n above zero and prevent it from growing too fast. In practice, ``exp()`` is usually enough.\n :param squash_output: Whether to squash the output using a tanh function,\n this allows to ensure boundaries when using gSDE.\n :param features_extractor_class: Features extractor to use.\n :param features_extractor_kwargs: Keyword arguments\n to pass to the features extractor.\n :param normalize_images: Whether to normalize images or not,\n dividing by 255.0 (True by default)\n :param optimizer_class: The optimizer to use,\n ``th.optim.Adam`` by default\n :param optimizer_kwargs: Additional keyword arguments,\n excluding the learning rate, to pass to the optimizer\n ",
|
| 7 |
-
"__init__": "<function ActorCriticPolicy.__init__ at
|
| 8 |
-
"_get_constructor_parameters": "<function ActorCriticPolicy._get_constructor_parameters at
|
| 9 |
-
"reset_noise": "<function ActorCriticPolicy.reset_noise at
|
| 10 |
-
"_build_mlp_extractor": "<function ActorCriticPolicy._build_mlp_extractor at
|
| 11 |
-
"_build": "<function ActorCriticPolicy._build at
|
| 12 |
-
"forward": "<function ActorCriticPolicy.forward at
|
| 13 |
-
"_get_action_dist_from_latent": "<function ActorCriticPolicy._get_action_dist_from_latent at
|
| 14 |
-
"_predict": "<function ActorCriticPolicy._predict at
|
| 15 |
-
"evaluate_actions": "<function ActorCriticPolicy.evaluate_actions at
|
| 16 |
-
"get_distribution": "<function ActorCriticPolicy.get_distribution at
|
| 17 |
-
"predict_values": "<function ActorCriticPolicy.predict_values at
|
| 18 |
"__abstractmethods__": "frozenset()",
|
| 19 |
-
"_abc_impl": "<_abc_data object at
|
| 20 |
},
|
| 21 |
"verbose": 1,
|
| 22 |
"policy_kwargs": {
|
|
@@ -68,12 +68,12 @@
|
|
| 68 |
"_num_timesteps_at_start": 0,
|
| 69 |
"seed": 0,
|
| 70 |
"action_noise": null,
|
| 71 |
-
"start_time":
|
| 72 |
"learning_rate": {
|
| 73 |
":type:": "<class 'function'>",
|
| 74 |
":serialized:": "gAWVhwIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwOX21ha2VfZnVuY3Rpb26Uk5QoaACMDV9idWlsdGluX3R5cGWUk5SMCENvZGVUeXBllIWUUpQoSwFLAEsASwFLAUsTQwSIAFMAlE6FlCmMAV+UhZSMVS9ob21lL21heGltaWxpYW4vLmxvY2FsL2xpYi9weXRob24zLjgvc2l0ZS1wYWNrYWdlcy9zdGFibGVfYmFzZWxpbmVzMy9jb21tb24vdXRpbHMucHmUjARmdW5jlEuAQwIAAZSMA3ZhbJSFlCl0lFKUfZQojAtfX3BhY2thZ2VfX5SMGHN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbpSMCF9fbmFtZV9flIwec3RhYmxlX2Jhc2VsaW5lczMuY29tbW9uLnV0aWxzlIwIX19maWxlX1+UaAx1Tk5oAIwQX21ha2VfZW1wdHlfY2VsbJSTlClSlIWUdJRSlIwcY2xvdWRwaWNrbGUuY2xvdWRwaWNrbGVfZmFzdJSMEl9mdW5jdGlvbl9zZXRzdGF0ZZSTlGgefZR9lChoFmgNjAxfX3F1YWxuYW1lX1+UjBljb25zdGFudF9mbi48bG9jYWxzPi5mdW5jlIwPX19hbm5vdGF0aW9uc19flH2UjA5fX2t3ZGVmYXVsdHNfX5ROjAxfX2RlZmF1bHRzX1+UTowKX19tb2R1bGVfX5RoF4wHX19kb2NfX5ROjAtfX2Nsb3N1cmVfX5RoAIwKX21ha2VfY2VsbJSTlEc/A+Ek31AkT4WUUpSFlIwXX2Nsb3VkcGlja2xlX3N1Ym1vZHVsZXOUXZSMC19fZ2xvYmFsc19flH2UdYaUhlIwLg=="
|
| 75 |
},
|
| 76 |
-
"tensorboard_log": "runs/seals/Walker2d-
|
| 77 |
"lr_schedule": {
|
| 78 |
":type:": "<class 'function'>",
|
| 79 |
":serialized:": "gAWVhwIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwOX21ha2VfZnVuY3Rpb26Uk5QoaACMDV9idWlsdGluX3R5cGWUk5SMCENvZGVUeXBllIWUUpQoSwFLAEsASwFLAUsTQwSIAFMAlE6FlCmMAV+UhZSMVS9ob21lL21heGltaWxpYW4vLmxvY2FsL2xpYi9weXRob24zLjgvc2l0ZS1wYWNrYWdlcy9zdGFibGVfYmFzZWxpbmVzMy9jb21tb24vdXRpbHMucHmUjARmdW5jlEuAQwIAAZSMA3ZhbJSFlCl0lFKUfZQojAtfX3BhY2thZ2VfX5SMGHN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbpSMCF9fbmFtZV9flIwec3RhYmxlX2Jhc2VsaW5lczMuY29tbW9uLnV0aWxzlIwIX19maWxlX1+UaAx1Tk5oAIwQX21ha2VfZW1wdHlfY2VsbJSTlClSlIWUdJRSlIwcY2xvdWRwaWNrbGUuY2xvdWRwaWNrbGVfZmFzdJSMEl9mdW5jdGlvbl9zZXRzdGF0ZZSTlGgefZR9lChoFmgNjAxfX3F1YWxuYW1lX1+UjBljb25zdGFudF9mbi48bG9jYWxzPi5mdW5jlIwPX19hbm5vdGF0aW9uc19flH2UjA5fX2t3ZGVmYXVsdHNfX5ROjAxfX2RlZmF1bHRzX1+UTowKX19tb2R1bGVfX5RoF4wHX19kb2NfX5ROjAtfX2Nsb3N1cmVfX5RoAIwKX21ha2VfY2VsbJSTlEc/A+Ek31AkT4WUUpSFlIwXX2Nsb3VkcGlja2xlX3N1Ym1vZHVsZXOUXZSMC19fZ2xvYmFsc19flH2UdYaUhlIwLg=="
|
|
@@ -93,7 +93,7 @@
|
|
| 93 |
"_current_progress_remaining": -0.0014719999999999178,
|
| 94 |
"ep_info_buffer": {
|
| 95 |
":type:": "<class 'collections.deque'>",
|
| 96 |
-
":serialized:": "gAWVgRAAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKUKH2UKIwBcpSMFW51bXB5LmNvcmUubXVsdGlhcnJheZSMBnNjYWxhcpSTlIwFbnVtcHmUjAVkdHlwZZSTlIwCZjiUiYiHlFKUKEsDjAE8lE5OTkr/////Sv////
|
| 97 |
},
|
| 98 |
"ep_success_buffer": {
|
| 99 |
":type:": "<class 'collections.deque'>",
|
|
|
|
| 4 |
":serialized:": "gAWVOwAAAAAAAACMIXN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi5wb2xpY2llc5SMEUFjdG9yQ3JpdGljUG9saWN5lJOULg==",
|
| 5 |
"__module__": "stable_baselines3.common.policies",
|
| 6 |
"__doc__": "\n Policy class for actor-critic algorithms (has both policy and value prediction).\n Used by A2C, PPO and the likes.\n\n :param observation_space: Observation space\n :param action_space: Action space\n :param lr_schedule: Learning rate schedule (could be constant)\n :param net_arch: The specification of the policy and value networks.\n :param activation_fn: Activation function\n :param ortho_init: Whether to use or not orthogonal initialization\n :param use_sde: Whether to use State Dependent Exploration or not\n :param log_std_init: Initial value for the log standard deviation\n :param full_std: Whether to use (n_features x n_actions) parameters\n for the std instead of only (n_features,) when using gSDE\n :param sde_net_arch: Network architecture for extracting features\n when using gSDE. If None, the latent features from the policy will be used.\n Pass an empty list to use the states as features.\n :param use_expln: Use ``expln()`` function instead of ``exp()`` to ensure\n a positive standard deviation (cf paper). It allows to keep variance\n above zero and prevent it from growing too fast. In practice, ``exp()`` is usually enough.\n :param squash_output: Whether to squash the output using a tanh function,\n this allows to ensure boundaries when using gSDE.\n :param features_extractor_class: Features extractor to use.\n :param features_extractor_kwargs: Keyword arguments\n to pass to the features extractor.\n :param normalize_images: Whether to normalize images or not,\n dividing by 255.0 (True by default)\n :param optimizer_class: The optimizer to use,\n ``th.optim.Adam`` by default\n :param optimizer_kwargs: Additional keyword arguments,\n excluding the learning rate, to pass to the optimizer\n ",
|
| 7 |
+
"__init__": "<function ActorCriticPolicy.__init__ at 0x7fae1fbbfca0>",
|
| 8 |
+
"_get_constructor_parameters": "<function ActorCriticPolicy._get_constructor_parameters at 0x7fae1fbbfd30>",
|
| 9 |
+
"reset_noise": "<function ActorCriticPolicy.reset_noise at 0x7fae1fbbfdc0>",
|
| 10 |
+
"_build_mlp_extractor": "<function ActorCriticPolicy._build_mlp_extractor at 0x7fae1fbbfe50>",
|
| 11 |
+
"_build": "<function ActorCriticPolicy._build at 0x7fae1fbbfee0>",
|
| 12 |
+
"forward": "<function ActorCriticPolicy.forward at 0x7fae1fbbff70>",
|
| 13 |
+
"_get_action_dist_from_latent": "<function ActorCriticPolicy._get_action_dist_from_latent at 0x7fae1fbbc040>",
|
| 14 |
+
"_predict": "<function ActorCriticPolicy._predict at 0x7fae1fbbc0d0>",
|
| 15 |
+
"evaluate_actions": "<function ActorCriticPolicy.evaluate_actions at 0x7fae1fbbc160>",
|
| 16 |
+
"get_distribution": "<function ActorCriticPolicy.get_distribution at 0x7fae1fbbc1f0>",
|
| 17 |
+
"predict_values": "<function ActorCriticPolicy.predict_values at 0x7fae1fbbc280>",
|
| 18 |
"__abstractmethods__": "frozenset()",
|
| 19 |
+
"_abc_impl": "<_abc_data object at 0x7fae1fbc1090>"
|
| 20 |
},
|
| 21 |
"verbose": 1,
|
| 22 |
"policy_kwargs": {
|
|
|
|
| 68 |
"_num_timesteps_at_start": 0,
|
| 69 |
"seed": 0,
|
| 70 |
"action_noise": null,
|
| 71 |
+
"start_time": 1658860236.103207,
|
| 72 |
"learning_rate": {
|
| 73 |
":type:": "<class 'function'>",
|
| 74 |
":serialized:": "gAWVhwIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwOX21ha2VfZnVuY3Rpb26Uk5QoaACMDV9idWlsdGluX3R5cGWUk5SMCENvZGVUeXBllIWUUpQoSwFLAEsASwFLAUsTQwSIAFMAlE6FlCmMAV+UhZSMVS9ob21lL21heGltaWxpYW4vLmxvY2FsL2xpYi9weXRob24zLjgvc2l0ZS1wYWNrYWdlcy9zdGFibGVfYmFzZWxpbmVzMy9jb21tb24vdXRpbHMucHmUjARmdW5jlEuAQwIAAZSMA3ZhbJSFlCl0lFKUfZQojAtfX3BhY2thZ2VfX5SMGHN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbpSMCF9fbmFtZV9flIwec3RhYmxlX2Jhc2VsaW5lczMuY29tbW9uLnV0aWxzlIwIX19maWxlX1+UaAx1Tk5oAIwQX21ha2VfZW1wdHlfY2VsbJSTlClSlIWUdJRSlIwcY2xvdWRwaWNrbGUuY2xvdWRwaWNrbGVfZmFzdJSMEl9mdW5jdGlvbl9zZXRzdGF0ZZSTlGgefZR9lChoFmgNjAxfX3F1YWxuYW1lX1+UjBljb25zdGFudF9mbi48bG9jYWxzPi5mdW5jlIwPX19hbm5vdGF0aW9uc19flH2UjA5fX2t3ZGVmYXVsdHNfX5ROjAxfX2RlZmF1bHRzX1+UTowKX19tb2R1bGVfX5RoF4wHX19kb2NfX5ROjAtfX2Nsb3N1cmVfX5RoAIwKX21ha2VfY2VsbJSTlEc/A+Ek31AkT4WUUpSFlIwXX2Nsb3VkcGlja2xlX3N1Ym1vZHVsZXOUXZSMC19fZ2xvYmFsc19flH2UdYaUhlIwLg=="
|
| 75 |
},
|
| 76 |
+
"tensorboard_log": "runs/seals/Walker2d-v0__ppo__5__1658860231/seals-Walker2d-v0",
|
| 77 |
"lr_schedule": {
|
| 78 |
":type:": "<class 'function'>",
|
| 79 |
":serialized:": "gAWVhwIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwOX21ha2VfZnVuY3Rpb26Uk5QoaACMDV9idWlsdGluX3R5cGWUk5SMCENvZGVUeXBllIWUUpQoSwFLAEsASwFLAUsTQwSIAFMAlE6FlCmMAV+UhZSMVS9ob21lL21heGltaWxpYW4vLmxvY2FsL2xpYi9weXRob24zLjgvc2l0ZS1wYWNrYWdlcy9zdGFibGVfYmFzZWxpbmVzMy9jb21tb24vdXRpbHMucHmUjARmdW5jlEuAQwIAAZSMA3ZhbJSFlCl0lFKUfZQojAtfX3BhY2thZ2VfX5SMGHN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbpSMCF9fbmFtZV9flIwec3RhYmxlX2Jhc2VsaW5lczMuY29tbW9uLnV0aWxzlIwIX19maWxlX1+UaAx1Tk5oAIwQX21ha2VfZW1wdHlfY2VsbJSTlClSlIWUdJRSlIwcY2xvdWRwaWNrbGUuY2xvdWRwaWNrbGVfZmFzdJSMEl9mdW5jdGlvbl9zZXRzdGF0ZZSTlGgefZR9lChoFmgNjAxfX3F1YWxuYW1lX1+UjBljb25zdGFudF9mbi48bG9jYWxzPi5mdW5jlIwPX19hbm5vdGF0aW9uc19flH2UjA5fX2t3ZGVmYXVsdHNfX5ROjAxfX2RlZmF1bHRzX1+UTowKX19tb2R1bGVfX5RoF4wHX19kb2NfX5ROjAtfX2Nsb3N1cmVfX5RoAIwKX21ha2VfY2VsbJSTlEc/A+Ek31AkT4WUUpSFlIwXX2Nsb3VkcGlja2xlX3N1Ym1vZHVsZXOUXZSMC19fZ2xvYmFsc19flH2UdYaUhlIwLg=="
|
|
|
|
| 93 |
"_current_progress_remaining": -0.0014719999999999178,
|
| 94 |
"ep_info_buffer": {
|
| 95 |
":type:": "<class 'collections.deque'>",
|
| 96 |
+
":serialized:": "gAWVgRAAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKUKH2UKIwBcpSMFW51bXB5LmNvcmUubXVsdGlhcnJheZSMBnNjYWxhcpSTlIwFbnVtcHmUjAVkdHlwZZSTlIwCZjiUiYiHlFKUKEsDjAE8lE5OTkr/////Sv////9LAHSUYkMI7idjfCDwqUCUhpRSlIwBbJRN6AOMAXSUR0C7S5/NJOFhdX2UKGgGaAloD0MIvko+dhdjrkCUhpRSlGgVTegDaBZHQLtN+ZvkzXV1fZQoaAZoCWgPQwiH3XcM39mqQJSGlFKUaBVN6ANoFkdAu1rN9hJAdHV9lChoBmgJaA9DCFqCjIBqCKlAlIaUUpRoFU3oA2gWR0C7XQNtygf2dX2UKGgGaAloD0MIqmIq/cxgp0CUhpRSlGgVTegDaBZHQLtqFfjjrAx1fZQoaAZoCWgPQwhMbhRZ2yKuQJSGlFKUaBVN6ANoFkdAu2yFWV/tpnV9lChoBmgJaA9DCACrI0cSVKtAlIaUUpRoFU3oA2gWR0C7eUOZG8VYdX2UKGgGaAloD0MIpYKKqg+DrECUhpRSlGgVTegDaBZHQLt7wgJTl1d1fZQoaAZoCWgPQwgu46YG2vWmQJSGlFKUaBVN6ANoFkdAu4jssxwhn3V9lChoBmgJaA9DCHJSmPcgUqxAlIaUUpRoFU3oA2gWR0C7ix+R9w3pdX2UKGgGaAloD0MIYTdsW9QZgMCUhpRSlGgVTegDaBZHQLuX1Ur08Nh1fZQoaAZoCWgPQwhcqtIWz5ynQJSGlFKUaBVN6ANoFkdAu5ouKiwjdHV9lChoBmgJaA9DCBi0kICZV6VAlIaUUpRoFU3oA2gWR0C7ptD9n9NvdX2UKGgGaAloD0MIQUXVr8xzpkCUhpRSlGgVTegDaBZHQLupFErXlKd1fZQoaAZoCWgPQwgo0v2cMuarQJSGlFKUaBVN6ANoFkdAu7WjaakRBnV9lChoBmgJaA9DCBjrG5jsEq1AlIaUUpRoFU3oA2gWR0C7uCDyz5XVdX2UKGgGaAloD0MIb7n6sWHprECUhpRSlGgVTegDaBZHQLvE+fsNUfh1fZQoaAZoCWgPQwgW3XpNj26JwJSGlFKUaBVN6ANoFkdAu8dOEBbOeXV9lChoBmgJaA9DCNdnzvosbqxAlIaUUpRoFU3oA2gWR0C71LWtp22YdX2UKGgGaAloD0MI7C5QUnirqUCUhpRSlGgVTegDaBZHQLvW+d56dDp1fZQoaAZoCWgPQwjXw5eJUkarQJSGlFKUaBVN6ANoFkdAu+RLS4OMEXV9lChoBmgJaA9DCG2rWWcsYqlAlIaUUpRoFU3oA2gWR0C75tvwmVqvdX2UKGgGaAloD0MIzuLFwqiWqUCUhpRSlGgVTegDaBZHQLvz+gbZOBV1fZQoaAZoCWgPQwhvhEVFTAytQJSGlFKUaBVN6ANoFkdAu/ZIidJ8OXV9lChoBmgJaA9DCJBnl289xK9AlIaUUpRoFU3oA2gWR0C8DRhfa6BidX2UKGgGaAloD0MItMpMaS1Wq0CUhpRSlGgVTegDaBZHQLwPSC/XXiB1fZQoaAZoCWgPQwgRx7q4XTKmQJSGlFKUaBVN6ANoFkdAvBzegezUqnV9lChoBmgJaA9DCMLdWbuV+KpAlIaUUpRoFU3oA2gWR0C8H6olhPTHdX2UKGgGaAloD0MImBO0yUEqgsCUhpRSlGgVTegDaBZHQLwsu6I3zc11fZQoaAZoCWgPQwg/OnXlu/itQJSGlFKUaBVN6ANoFkdAvC7Qb4rSVnV9lChoBmgJaA9DCKeWrfWdvqhAlIaUUpRoFU3oA2gWR0C8O/AWBSUDdX2UKGgGaAloD0MItvXTf9YeUsCUhpRSlGgVTegDaBZHQLw+nZamoBJ1fZQoaAZoCWgPQwgxCRfybHuwQJSGlFKUaBVN6ANoFkdAvEusenyd4HV9lChoBmgJaA9DCGCUoL+w2atAlIaUUpRoFU3oA2gWR0C8Tehm9QGfdX2UKGgGaAloD0MI3xYs1cVHpkCUhpRSlGgVTegDaBZHQLxaWyjpLVZ1fZQoaAZoCWgPQwh5lEp4atmqQJSGlFKUaBVN6ANoFkdAvFzxg3Lmp3V9lChoBmgJaA9DCLyuX7CbQ6pAlIaUUpRoFU3oA2gWR0C8aZVLOAy3dX2UKGgGaAloD0MIC9C2mv3KqkCUhpRSlGgVTegDaBZHQLxrs26kIop1fZQoaAZoCWgPQwj11sBWoVqvQJSGlFKUaBVN6ANoFkdAvG3gfHPu5XV9lChoBmgJaA9DCAVQjCxZznzAlIaUUpRoFU3oA2gWR0C8eqjmSyMUdX2UKGgGaAloD0MI6tDpeUfLqECUhpRSlGgVTegDaBZHQLx9TUY8+zN1fZQoaAZoCWgPQwiiKTv9WNOrQJSGlFKUaBVN6ANoFkdAvIpJPpIMB3V9lChoBmgJaA9DCGPwMO3Lzq1AlIaUUpRoFU3oA2gWR0C8jLcs+V1PdX2UKGgGaAloD0MIL204LBXupUCUhpRSlGgVTegDaBZHQLyZU7dznzR1fZQoaAZoCWgPQwg7GRwll6uoQJSGlFKUaBVN6ANoFkdAvJt2ubI91XV9lChoBmgJaA9DCGJNZVGQUq1AlIaUUpRoFU3oA2gWR0C8qLeSOinHdX2UKGgGaAloD0MIgGJkyXyPrUCUhpRSlGgVTegDaBZHQLyrMrVOKwZ1fZQoaAZoCWgPQwgZ5ZmXMwqlQJSGlFKUaBVN6ANoFkdAvLip43WFvnV9lChoBmgJaA9DCF392CS3Jq1AlIaUUpRoFU3oA2gWR0C8uspIlMRIdX2UKGgGaAloD0MIwJKrWNSxqkCUhpRSlGgVTegDaBZHQLzRn17Y02t1fZQoaAZoCWgPQwhAUG7bd7erQJSGlFKUaBVN6ANoFkdAvNQiCz1K5HV9lChoBmgJaA9DCJRpNLkwUK5AlIaUUpRoFU3oA2gWR0C84Qtcv/R3dX2UKGgGaAloD0MIwZDVrXYuqECUhpRSlGgVTegDaBZHQLzjjAkLQX11fZQoaAZoCWgPQwitS43QL/KvQJSGlFKUaBVN6ANoFkdAvPB5oZhrnHV9lChoBmgJaA9DCPXyO032sq1AlIaUUpRoFU3oA2gWR0C88svGEPDpdX2UKGgGaAloD0MIvviiPe7npkCUhpRSlGgVTegDaBZHQL0AGn2Iwdt1fZQoaAZoCWgPQwiOW8zP1ZGsQJSGlFKUaBVN6ANoFkdAvQK8pYs/ZHV9lChoBmgJaA9DCGAgCJDhs6pAlIaUUpRoFU3oA2gWR0C9D5FFUhmodX2UKGgGaAloD0MIg02dR821rECUhpRSlGgVTegDaBZHQL0SDC2c8T11fZQoaAZoCWgPQwh6GjBIOq+tQJSGlFKUaBVN6ANoFkdAvR41NWU8m3V9lChoBmgJaA9DCHdOs0AzDK9AlIaUUpRoFU3oA2gWR0C9IFcJY1YRdX2UKGgGaAloD0MI0ZFc/sO+q0CUhpRSlGgVTegDaBZHQL0tfwHZ9NN1fZQoaAZoCWgPQwhzvW2meiawQJSGlFKUaBVN6ANoFkdAvS/Mw1zhgnV9lChoBmgJaA9DCFVNEHWH8KpAlIaUUpRoFU3oA2gWR0C9PPpGvwEydX2UKGgGaAloD0MIDyvc8uGdrECUhpRSlGgVTegDaBZHQL0/idgOSW91fZQoaAZoCWgPQwiJQsu670atQJSGlFKUaBVN6ANoFkdAvUyvlV94NnV9lChoBmgJaA9DCFvR5jgfDqpAlIaUUpRoFU3oA2gWR0C9Tyuv6j33dX2UKGgGaAloD0MI4UbKFmFzqECUhpRSlGgVTegDaBZHQL1b+Lfk3jx1fZQoaAZoCWgPQwj2lnK+yCyrQJSGlFKUaBVN6ANoFkdAvV5NXxOLznV9lChoBmgJaA9DCPhVuVDRWqtAlIaUUpRoFU3oA2gWR0C9azUqlP8AdX2UKGgGaAloD0MI7bd2oszHp0CUhpRSlGgVTegDaBZHQL1tgJgb6xh1fZQoaAZoCWgPQwguG53zWxWpQJSGlFKUaBVN6ANoFkdAvXpTIcR15nV9lChoBmgJaA9DCLq9pDGS0KlAlIaUUpRoFU3oA2gWR0C9fN00BOpLdX2UKGgGaAloD0MIGqN1VE1xqkCUhpRSlGgVTegDaBZHQL2J2UUfxMF1fZQoaAZoCWgPQwiXH7jK24yuQJSGlFKUaBVN6ANoFkdAvZYvm3fAK3V9lChoBmgJaA9DCAM/qmE/uKtAlIaUUpRoFU3oA2gWR0C9o7ckleF+dX2UKGgGaAloD0MIs/D1td7wqkCUhpRSlGgVTegDaBZHQL2mry0rsjV1fZQoaAZoCWgPQwjAJmvU0xmnQJSGlFKUaBVN6ANoFkdAvbObfvWpZXV9lChoBmgJaA9DCKBrX0AHl6hAlIaUUpRoFU3oA2gWR0C9tfLTx5LRdX2UKGgGaAloD0MIU67wLv8SrECUhpRSlGgVTegDaBZHQL3C2ZlnRLN1fZQoaAZoCWgPQwjQfM7dRuirQJSGlFKUaBVN6ANoFkdAvcVmshgVoHV9lChoBmgJaA9DCGIwf4W8eK5AlIaUUpRoFU3oA2gWR0C9yDpMg2ZRdX2UKGgGaAloD0MIfzDw3LPIr0CUhpRSlGgVTegDaBZHQL3VzI9TxXp1fZQoaAZoCWgPQwj7lc6HhwmtQJSGlFKUaBVN6ANoFkdAvdg8c0cfeXV9lChoBmgJaA9DCN8WLNXFfqdAlIaUUpRoFU3oA2gWR0C95SFNQCSzdX2UKGgGaAloD0MIBmNEotAeqkCUhpRSlGgVTegDaBZHQL3nj5eZ5Rl1fZQoaAZoCWgPQwhMp3UbRN+pQJSGlFKUaBVN6ANoFkdAvfTriMo+fXV9lChoBmgJaA9DCOIhjJ8ub6pAlIaUUpRoFU3oA2gWR0C994tY8uBddX2UKGgGaAloD0MIVP1K5wtLp0CUhpRSlGgVTegDaBZHQL4EnDWK/Eh1fZQoaAZoCWgPQwgQO1PoZOOjQJSGlFKUaBVN6ANoFkdAvgcBy+6AfHV9lChoBmgJaA9DCHpvDAHQEatAlIaUUpRoFU3oA2gWR0C+E/9SuQp4dX2UKGgGaAloD0MI5dTOMMWcrkCUhpRSlGgVTegDaBZHQL4WUPwNLDh1fZQoaAZoCWgPQwhDxw4qoZmrQJSGlFKUaBVN6ANoFkdAviM4QXhwVHV9lChoBmgJaA9DCB7iH7asZ6ZAlIaUUpRoFU3oA2gWR0C+JXSJoCdSdX2UKGgGaAloD0MIFqQZi34ZrkCUhpRSlGgVTegDaBZHQL4yYeV9nbt1fZQoaAZoCWgPQwi6aMh41P2qQJSGlFKUaBVN6ANoFkdAvjTY1LrX2HV9lChoBmgJaA9DCFdgyOo2569AlIaUUpRoFU3oA2gWR0C+QkzEvTPTdX2UKGgGaAloD0MIU0Da/9BvrECUhpRSlGgVTegDaBZHQL5EhqD9Oyp1fZQoaAZoCWgPQwhmguFc+/qnQJSGlFKUaBVN6ANoFkdAvlE46xPfsXV9lChoBmgJaA9DCFhYcD+Qc6lAlIaUUpRoFU3oA2gWR0C+XaAjps42dWUu"
|
| 97 |
},
|
| 98 |
"ep_success_buffer": {
|
| 99 |
":type:": "<class 'collections.deque'>",
|
results.json
CHANGED
|
@@ -1 +1 @@
|
|
| 1 |
-
{"mean_reward": 3714.7483351, "std_reward": 374.4360849549338, "is_deterministic": true, "n_eval_episodes": 10, "eval_datetime": "2022-07-
|
|
|
|
| 1 |
+
{"mean_reward": 3714.7483351, "std_reward": 374.4360849549338, "is_deterministic": true, "n_eval_episodes": 10, "eval_datetime": "2022-07-27T16:51:30.692783"}
|
train_eval_metrics.zip
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e7016779610d98e5a5a4b51f7d522f5945a72ffbeab7fc0b4e16778bfe1c11de
|
| 3 |
+
size 34055
|