Initial commit
Browse files- args.yml +3 -3
- ppo-seals-Hopper-v0.zip +2 -2
- ppo-seals-Hopper-v0/data +15 -15
- results.json +1 -1
- train_eval_metrics.zip +2 -2
args.yml
CHANGED
|
@@ -16,7 +16,7 @@
|
|
| 16 |
- - hyperparams
|
| 17 |
- null
|
| 18 |
- - log_folder
|
| 19 |
-
-
|
| 20 |
- - log_interval
|
| 21 |
- -1
|
| 22 |
- - max_total_trials
|
|
@@ -56,7 +56,7 @@
|
|
| 56 |
- - study_name
|
| 57 |
- null
|
| 58 |
- - tensorboard_log
|
| 59 |
-
- runs/seals/Hopper-
|
| 60 |
- - track
|
| 61 |
- true
|
| 62 |
- - trained_agent
|
|
@@ -72,4 +72,4 @@
|
|
| 72 |
- - wandb_entity
|
| 73 |
- null
|
| 74 |
- - wandb_project_name
|
| 75 |
-
- seals-experts
|
|
|
|
| 16 |
- - hyperparams
|
| 17 |
- null
|
| 18 |
- - log_folder
|
| 19 |
+
- seals_experts_wandb_oldpickle/seed_3/
|
| 20 |
- - log_interval
|
| 21 |
- -1
|
| 22 |
- - max_total_trials
|
|
|
|
| 56 |
- - study_name
|
| 57 |
- null
|
| 58 |
- - tensorboard_log
|
| 59 |
+
- runs/seals/Hopper-v0__ppo__3__1658845752
|
| 60 |
- - track
|
| 61 |
- true
|
| 62 |
- - trained_agent
|
|
|
|
| 72 |
- - wandb_entity
|
| 73 |
- null
|
| 74 |
- - wandb_project_name
|
| 75 |
+
- seals-experts-oldpickle
|
ppo-seals-Hopper-v0.zip
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a82d243c4d55c39a2297ebfdb2b25ef779622b682dcfe08e8da3b03ac1e6d133
|
| 3 |
+
size 158296
|
ppo-seals-Hopper-v0/data
CHANGED
|
@@ -4,19 +4,19 @@
|
|
| 4 |
":serialized:": "gAWVOwAAAAAAAACMIXN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi5wb2xpY2llc5SMEUFjdG9yQ3JpdGljUG9saWN5lJOULg==",
|
| 5 |
"__module__": "stable_baselines3.common.policies",
|
| 6 |
"__doc__": "\n Policy class for actor-critic algorithms (has both policy and value prediction).\n Used by A2C, PPO and the likes.\n\n :param observation_space: Observation space\n :param action_space: Action space\n :param lr_schedule: Learning rate schedule (could be constant)\n :param net_arch: The specification of the policy and value networks.\n :param activation_fn: Activation function\n :param ortho_init: Whether to use or not orthogonal initialization\n :param use_sde: Whether to use State Dependent Exploration or not\n :param log_std_init: Initial value for the log standard deviation\n :param full_std: Whether to use (n_features x n_actions) parameters\n for the std instead of only (n_features,) when using gSDE\n :param sde_net_arch: Network architecture for extracting features\n when using gSDE. If None, the latent features from the policy will be used.\n Pass an empty list to use the states as features.\n :param use_expln: Use ``expln()`` function instead of ``exp()`` to ensure\n a positive standard deviation (cf paper). It allows to keep variance\n above zero and prevent it from growing too fast. In practice, ``exp()`` is usually enough.\n :param squash_output: Whether to squash the output using a tanh function,\n this allows to ensure boundaries when using gSDE.\n :param features_extractor_class: Features extractor to use.\n :param features_extractor_kwargs: Keyword arguments\n to pass to the features extractor.\n :param normalize_images: Whether to normalize images or not,\n dividing by 255.0 (True by default)\n :param optimizer_class: The optimizer to use,\n ``th.optim.Adam`` by default\n :param optimizer_kwargs: Additional keyword arguments,\n excluding the learning rate, to pass to the optimizer\n ",
|
| 7 |
-
"__init__": "<function ActorCriticPolicy.__init__ at
|
| 8 |
-
"_get_constructor_parameters": "<function ActorCriticPolicy._get_constructor_parameters at
|
| 9 |
-
"reset_noise": "<function ActorCriticPolicy.reset_noise at
|
| 10 |
-
"_build_mlp_extractor": "<function ActorCriticPolicy._build_mlp_extractor at
|
| 11 |
-
"_build": "<function ActorCriticPolicy._build at
|
| 12 |
-
"forward": "<function ActorCriticPolicy.forward at
|
| 13 |
-
"_get_action_dist_from_latent": "<function ActorCriticPolicy._get_action_dist_from_latent at
|
| 14 |
-
"_predict": "<function ActorCriticPolicy._predict at
|
| 15 |
-
"evaluate_actions": "<function ActorCriticPolicy.evaluate_actions at
|
| 16 |
-
"get_distribution": "<function ActorCriticPolicy.get_distribution at
|
| 17 |
-
"predict_values": "<function ActorCriticPolicy.predict_values at
|
| 18 |
"__abstractmethods__": "frozenset()",
|
| 19 |
-
"_abc_impl": "<_abc_data object at
|
| 20 |
},
|
| 21 |
"verbose": 1,
|
| 22 |
"policy_kwargs": {
|
|
@@ -68,12 +68,12 @@
|
|
| 68 |
"_num_timesteps_at_start": 0,
|
| 69 |
"seed": 0,
|
| 70 |
"action_noise": null,
|
| 71 |
-
"start_time":
|
| 72 |
"learning_rate": {
|
| 73 |
":type:": "<class 'function'>",
|
| 74 |
":serialized:": "gAWVhwIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwOX21ha2VfZnVuY3Rpb26Uk5QoaACMDV9idWlsdGluX3R5cGWUk5SMCENvZGVUeXBllIWUUpQoSwFLAEsASwFLAUsTQwSIAFMAlE6FlCmMAV+UhZSMVS9ob21lL21heGltaWxpYW4vLmxvY2FsL2xpYi9weXRob24zLjgvc2l0ZS1wYWNrYWdlcy9zdGFibGVfYmFzZWxpbmVzMy9jb21tb24vdXRpbHMucHmUjARmdW5jlEuAQwIAAZSMA3ZhbJSFlCl0lFKUfZQojAtfX3BhY2thZ2VfX5SMGHN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbpSMCF9fbmFtZV9flIwec3RhYmxlX2Jhc2VsaW5lczMuY29tbW9uLnV0aWxzlIwIX19maWxlX1+UaAx1Tk5oAIwQX21ha2VfZW1wdHlfY2VsbJSTlClSlIWUdJRSlIwcY2xvdWRwaWNrbGUuY2xvdWRwaWNrbGVfZmFzdJSMEl9mdW5jdGlvbl9zZXRzdGF0ZZSTlGgefZR9lChoFmgNjAxfX3F1YWxuYW1lX1+UjBljb25zdGFudF9mbi48bG9jYWxzPi5mdW5jlIwPX19hbm5vdGF0aW9uc19flH2UjA5fX2t3ZGVmYXVsdHNfX5ROjAxfX2RlZmF1bHRzX1+UTowKX19tb2R1bGVfX5RoF4wHX19kb2NfX5ROjAtfX2Nsb3N1cmVfX5RoAIwKX21ha2VfY2VsbJSTlEc/OZceI3Gfy4WUUpSFlIwXX2Nsb3VkcGlja2xlX3N1Ym1vZHVsZXOUXZSMC19fZ2xvYmFsc19flH2UdYaUhlIwLg=="
|
| 75 |
},
|
| 76 |
-
"tensorboard_log": "runs/seals/Hopper-
|
| 77 |
"lr_schedule": {
|
| 78 |
":type:": "<class 'function'>",
|
| 79 |
":serialized:": "gAWVhwIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwOX21ha2VfZnVuY3Rpb26Uk5QoaACMDV9idWlsdGluX3R5cGWUk5SMCENvZGVUeXBllIWUUpQoSwFLAEsASwFLAUsTQwSIAFMAlE6FlCmMAV+UhZSMVS9ob21lL21heGltaWxpYW4vLmxvY2FsL2xpYi9weXRob24zLjgvc2l0ZS1wYWNrYWdlcy9zdGFibGVfYmFzZWxpbmVzMy9jb21tb24vdXRpbHMucHmUjARmdW5jlEuAQwIAAZSMA3ZhbJSFlCl0lFKUfZQojAtfX3BhY2thZ2VfX5SMGHN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbpSMCF9fbmFtZV9flIwec3RhYmxlX2Jhc2VsaW5lczMuY29tbW9uLnV0aWxzlIwIX19maWxlX1+UaAx1Tk5oAIwQX21ha2VfZW1wdHlfY2VsbJSTlClSlIWUdJRSlIwcY2xvdWRwaWNrbGUuY2xvdWRwaWNrbGVfZmFzdJSMEl9mdW5jdGlvbl9zZXRzdGF0ZZSTlGgefZR9lChoFmgNjAxfX3F1YWxuYW1lX1+UjBljb25zdGFudF9mbi48bG9jYWxzPi5mdW5jlIwPX19hbm5vdGF0aW9uc19flH2UjA5fX2t3ZGVmYXVsdHNfX5ROjAxfX2RlZmF1bHRzX1+UTowKX19tb2R1bGVfX5RoF4wHX19kb2NfX5ROjAtfX2Nsb3N1cmVfX5RoAIwKX21ha2VfY2VsbJSTlEc/OZceI3Gfy4WUUpSFlIwXX2Nsb3VkcGlja2xlX3N1Ym1vZHVsZXOUXZSMC19fZ2xvYmFsc19flH2UdYaUhlIwLg=="
|
|
@@ -93,7 +93,7 @@
|
|
| 93 |
"_current_progress_remaining": -0.0014719999999999178,
|
| 94 |
"ep_info_buffer": {
|
| 95 |
":type:": "<class 'collections.deque'>",
|
| 96 |
-
":serialized:": "gAWVgRAAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKUKH2UKIwBcpSMFW51bXB5LmNvcmUubXVsdGlhcnJheZSMBnNjYWxhcpSTlIwFbnVtcHmUjAVkdHlwZZSTlIwCZjiUiYiHlFKUKEsDjAE8lE5OTkr/////Sv////
|
| 97 |
},
|
| 98 |
"ep_success_buffer": {
|
| 99 |
":type:": "<class 'collections.deque'>",
|
|
|
|
| 4 |
":serialized:": "gAWVOwAAAAAAAACMIXN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi5wb2xpY2llc5SMEUFjdG9yQ3JpdGljUG9saWN5lJOULg==",
|
| 5 |
"__module__": "stable_baselines3.common.policies",
|
| 6 |
"__doc__": "\n Policy class for actor-critic algorithms (has both policy and value prediction).\n Used by A2C, PPO and the likes.\n\n :param observation_space: Observation space\n :param action_space: Action space\n :param lr_schedule: Learning rate schedule (could be constant)\n :param net_arch: The specification of the policy and value networks.\n :param activation_fn: Activation function\n :param ortho_init: Whether to use or not orthogonal initialization\n :param use_sde: Whether to use State Dependent Exploration or not\n :param log_std_init: Initial value for the log standard deviation\n :param full_std: Whether to use (n_features x n_actions) parameters\n for the std instead of only (n_features,) when using gSDE\n :param sde_net_arch: Network architecture for extracting features\n when using gSDE. If None, the latent features from the policy will be used.\n Pass an empty list to use the states as features.\n :param use_expln: Use ``expln()`` function instead of ``exp()`` to ensure\n a positive standard deviation (cf paper). It allows to keep variance\n above zero and prevent it from growing too fast. In practice, ``exp()`` is usually enough.\n :param squash_output: Whether to squash the output using a tanh function,\n this allows to ensure boundaries when using gSDE.\n :param features_extractor_class: Features extractor to use.\n :param features_extractor_kwargs: Keyword arguments\n to pass to the features extractor.\n :param normalize_images: Whether to normalize images or not,\n dividing by 255.0 (True by default)\n :param optimizer_class: The optimizer to use,\n ``th.optim.Adam`` by default\n :param optimizer_kwargs: Additional keyword arguments,\n excluding the learning rate, to pass to the optimizer\n ",
|
| 7 |
+
"__init__": "<function ActorCriticPolicy.__init__ at 0x7f04948f4ca0>",
|
| 8 |
+
"_get_constructor_parameters": "<function ActorCriticPolicy._get_constructor_parameters at 0x7f04948f4d30>",
|
| 9 |
+
"reset_noise": "<function ActorCriticPolicy.reset_noise at 0x7f04948f4dc0>",
|
| 10 |
+
"_build_mlp_extractor": "<function ActorCriticPolicy._build_mlp_extractor at 0x7f04948f4e50>",
|
| 11 |
+
"_build": "<function ActorCriticPolicy._build at 0x7f04948f4ee0>",
|
| 12 |
+
"forward": "<function ActorCriticPolicy.forward at 0x7f04948f4f70>",
|
| 13 |
+
"_get_action_dist_from_latent": "<function ActorCriticPolicy._get_action_dist_from_latent at 0x7f04948f1040>",
|
| 14 |
+
"_predict": "<function ActorCriticPolicy._predict at 0x7f04948f10d0>",
|
| 15 |
+
"evaluate_actions": "<function ActorCriticPolicy.evaluate_actions at 0x7f04948f1160>",
|
| 16 |
+
"get_distribution": "<function ActorCriticPolicy.get_distribution at 0x7f04948f11f0>",
|
| 17 |
+
"predict_values": "<function ActorCriticPolicy.predict_values at 0x7f04948f1280>",
|
| 18 |
"__abstractmethods__": "frozenset()",
|
| 19 |
+
"_abc_impl": "<_abc_data object at 0x7f04948f8090>"
|
| 20 |
},
|
| 21 |
"verbose": 1,
|
| 22 |
"policy_kwargs": {
|
|
|
|
| 68 |
"_num_timesteps_at_start": 0,
|
| 69 |
"seed": 0,
|
| 70 |
"action_noise": null,
|
| 71 |
+
"start_time": 1658845755.784661,
|
| 72 |
"learning_rate": {
|
| 73 |
":type:": "<class 'function'>",
|
| 74 |
":serialized:": "gAWVhwIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwOX21ha2VfZnVuY3Rpb26Uk5QoaACMDV9idWlsdGluX3R5cGWUk5SMCENvZGVUeXBllIWUUpQoSwFLAEsASwFLAUsTQwSIAFMAlE6FlCmMAV+UhZSMVS9ob21lL21heGltaWxpYW4vLmxvY2FsL2xpYi9weXRob24zLjgvc2l0ZS1wYWNrYWdlcy9zdGFibGVfYmFzZWxpbmVzMy9jb21tb24vdXRpbHMucHmUjARmdW5jlEuAQwIAAZSMA3ZhbJSFlCl0lFKUfZQojAtfX3BhY2thZ2VfX5SMGHN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbpSMCF9fbmFtZV9flIwec3RhYmxlX2Jhc2VsaW5lczMuY29tbW9uLnV0aWxzlIwIX19maWxlX1+UaAx1Tk5oAIwQX21ha2VfZW1wdHlfY2VsbJSTlClSlIWUdJRSlIwcY2xvdWRwaWNrbGUuY2xvdWRwaWNrbGVfZmFzdJSMEl9mdW5jdGlvbl9zZXRzdGF0ZZSTlGgefZR9lChoFmgNjAxfX3F1YWxuYW1lX1+UjBljb25zdGFudF9mbi48bG9jYWxzPi5mdW5jlIwPX19hbm5vdGF0aW9uc19flH2UjA5fX2t3ZGVmYXVsdHNfX5ROjAxfX2RlZmF1bHRzX1+UTowKX19tb2R1bGVfX5RoF4wHX19kb2NfX5ROjAtfX2Nsb3N1cmVfX5RoAIwKX21ha2VfY2VsbJSTlEc/OZceI3Gfy4WUUpSFlIwXX2Nsb3VkcGlja2xlX3N1Ym1vZHVsZXOUXZSMC19fZ2xvYmFsc19flH2UdYaUhlIwLg=="
|
| 75 |
},
|
| 76 |
+
"tensorboard_log": "runs/seals/Hopper-v0__ppo__3__1658845752/seals-Hopper-v0",
|
| 77 |
"lr_schedule": {
|
| 78 |
":type:": "<class 'function'>",
|
| 79 |
":serialized:": "gAWVhwIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwOX21ha2VfZnVuY3Rpb26Uk5QoaACMDV9idWlsdGluX3R5cGWUk5SMCENvZGVUeXBllIWUUpQoSwFLAEsASwFLAUsTQwSIAFMAlE6FlCmMAV+UhZSMVS9ob21lL21heGltaWxpYW4vLmxvY2FsL2xpYi9weXRob24zLjgvc2l0ZS1wYWNrYWdlcy9zdGFibGVfYmFzZWxpbmVzMy9jb21tb24vdXRpbHMucHmUjARmdW5jlEuAQwIAAZSMA3ZhbJSFlCl0lFKUfZQojAtfX3BhY2thZ2VfX5SMGHN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbpSMCF9fbmFtZV9flIwec3RhYmxlX2Jhc2VsaW5lczMuY29tbW9uLnV0aWxzlIwIX19maWxlX1+UaAx1Tk5oAIwQX21ha2VfZW1wdHlfY2VsbJSTlClSlIWUdJRSlIwcY2xvdWRwaWNrbGUuY2xvdWRwaWNrbGVfZmFzdJSMEl9mdW5jdGlvbl9zZXRzdGF0ZZSTlGgefZR9lChoFmgNjAxfX3F1YWxuYW1lX1+UjBljb25zdGFudF9mbi48bG9jYWxzPi5mdW5jlIwPX19hbm5vdGF0aW9uc19flH2UjA5fX2t3ZGVmYXVsdHNfX5ROjAxfX2RlZmF1bHRzX1+UTowKX19tb2R1bGVfX5RoF4wHX19kb2NfX5ROjAtfX2Nsb3N1cmVfX5RoAIwKX21ha2VfY2VsbJSTlEc/OZceI3Gfy4WUUpSFlIwXX2Nsb3VkcGlja2xlX3N1Ym1vZHVsZXOUXZSMC19fZ2xvYmFsc19flH2UdYaUhlIwLg=="
|
|
|
|
| 93 |
"_current_progress_remaining": -0.0014719999999999178,
|
| 94 |
"ep_info_buffer": {
|
| 95 |
":type:": "<class 'collections.deque'>",
|
| 96 |
+
":serialized:": "gAWVgRAAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKUKH2UKIwBcpSMFW51bXB5LmNvcmUubXVsdGlhcnJheZSMBnNjYWxhcpSTlIwFbnVtcHmUjAVkdHlwZZSTlIwCZjiUiYiHlFKUKEsDjAE8lE5OTkr/////Sv////9LAHSUYkMIg09z8rKmrkCUhpRSlIwBbJRN6AOMAXSUR0CZOwcYqG1ydX2UKGgGaAloD0MIytx8IzoSrUCUhpRSlGgVTegDaBZHQJlAxQl8gIR1fZQoaAZoCWgPQwiqYFRSH+GrQJSGlFKUaBVN6ANoFkdAmUhMVclgMXV9lChoBmgJaA9DCPvnacCYD6pAlIaUUpRoFU3oA2gWR0CZTeKgZjx1dX2UKGgGaAloD0MIti3KbEDVpUCUhpRSlGgVTegDaBZHQJlVZyIYWLx1fZQoaAZoCWgPQwhnuWx0bpesQJSGlFKUaBVN6ANoFkdAmVsO9alk6XV9lChoBmgJaA9DCOm2RC5g86lAlIaUUpRoFU3oA2gWR0CZYrqqfe1sdX2UKGgGaAloD0MIIcms3mEQrkCUhpRSlGgVTegDaBZHQJloR+CsfaJ1fZQoaAZoCWgPQwhEqFKzf9KlQJSGlFKUaBVN6ANoFkdAmXAn2h7E53V9lChoBmgJaA9DCBhd3hyOAKpAlIaUUpRoFU3oA2gWR0CZdZwm3OObdX2UKGgGaAloD0MI+Um1T0diq0CUhpRSlGgVTegDaBZHQJl9NmTTvy91fZQoaAZoCWgPQwhsfCb796esQJSGlFKUaBVN6ANoFkdAmYLBCUornXV9lChoBmgJaA9DCKfn3Vjw8qVAlIaUUpRoFU3oA2gWR0CZiht/nW8RdX2UKGgGaAloD0MIXMZNDQz5qUCUhpRSlGgVTegDaBZHQJmPqivgWJt1fZQoaAZoCWgPQwjJy5pY4KWoQJSGlFKUaBVN6ANoFkdAmZcHYHxBmnV9lChoBmgJaA9DCPtYwW+7Ha1AlIaUUpRoFU3oA2gWR0CZnHaiKziTdX2UKGgGaAloD0MI7YLBNXdpq0CUhpRSlGgVTegDaBZHQJmj/pNbkfd1fZQoaAZoCWgPQwhbsirCBRqqQJSGlFKUaBVN6ANoFkdAmalpDmbLEHV9lChoBmgJaA9DCLRWtDmu56pAlIaUUpRoFU3oA2gWR0CZsOA5Jbt7dX2UKGgGaAloD0MI9diWAd/+qkCUhpRSlGgVTegDaBZHQJm2R5le4Td1fZQoaAZoCWgPQwj+fjFbypikQJSGlFKUaBVN6ANoFkdAmb35e7cwg3V9lChoBmgJaA9DCAn+t5J9G6RAlIaUUpRoFU3oA2gWR0CZwy+BYmsvdX2UKGgGaAloD0MImMEYkXh8q0CUhpRSlGgVTegDaBZHQJnKj/sE7nx1fZQoaAZoCWgPQwhnnlxTiHWoQJSGlFKUaBVN6ANoFkdAmc9yOearm3V9lChoBmgJaA9DCCXJc33XtKxAlIaUUpRoFU3oA2gWR0CZ7DITGo73dX2UKGgGaAloD0MI0T3rGs2tm0CUhpRSlGgVTegDaBZHQJnxyf+S8rZ1fZQoaAZoCWgPQwiXHk319NKoQJSGlFKUaBVN6ANoFkdAmfkp04iosXV9lChoBmgJaA9DCHgN+tLDZalAlIaUUpRoFU3oA2gWR0CZ/lIFvAGjdX2UKGgGaAloD0MIz4QmiZ1/p0CUhpRSlGgVTegDaBZHQJoFfsqril11fZQoaAZoCWgPQwjd6jnpLTGsQJSGlFKUaBVN6ANoFkdAmgsD/uLJjnV9lChoBmgJaA9DCPJfIAgICqtAlIaUUpRoFU3oA2gWR0CaEn+2E0zkdX2UKGgGaAloD0MINV1PdNUzrECUhpRSlGgVTegDaBZHQJoX8l/pdKN1fZQoaAZoCWgPQwh8YwgAnoirQJSGlFKUaBVN6ANoFkdAmh9caGYa53V9lChoBmgJaA9DCAtD5PSNB6VAlIaUUpRoFU3oA2gWR0CaJLsLORkmdX2UKGgGaAloD0MIECBDx84Rq0CUhpRSlGgVTegDaBZHQJosQH8jzI51fZQoaAZoCWgPQwiHxD2WXu2qQJSGlFKUaBVN6ANoFkdAmjIVXRw6yXV9lChoBmgJaA9DCJHSbB7Xb61AlIaUUpRoFU3oA2gWR0CaOfEqlP8AdX2UKGgGaAloD0MIzEOmfKgoq0CUhpRSlGgVTegDaBZHQJo/SucMEzR1fZQoaAZoCWgPQwj7WwLw1ySqQJSGlFKUaBVN6ANoFkdAmkSbPD50sHV9lChoBmgJaA9DCIHptG475qpAlIaUUpRoFU3oA2gWR0CaTEZvkzXSdX2UKGgGaAloD0MIdM5PcaTap0CUhpRSlGgVTegDaBZHQJpSTuNPxhF1fZQoaAZoCWgPQwggls0camGpQJSGlFKUaBVN6ANoFkdAmlnUSIxgzHV9lChoBmgJaA9DCE9cjlfIIq5AlIaUUpRoFU3oA2gWR0CaX0VFQVKxdX2UKGgGaAloD0MIIPDAAILPp0CUhpRSlGgVTegDaBZHQJpml49ovi91fZQoaAZoCWgPQwgOLbKd3yKrQJSGlFKUaBVN6ANoFkdAmmwZo4+8oXV9lChoBmgJaA9DCDoGZK+PxKpAlIaUUpRoFU3oA2gWR0Cac8aLGaQWdX2UKGgGaAloD0MIy7kUV61EqkCUhpRSlGgVTegDaBZHQJp5aW+oLoh1fZQoaAZoCWgPQwjuPsdH69mtQJSGlFKUaBVN6ANoFkdAmoC7rPdEcHV9lChoBmgJaA9DCOjZrPrEj61AlIaUUpRoFU3oA2gWR0CahigQHzH0dX2UKGgGaAloD0MINpNvtmGkrUCUhpRSlGgVTegDaBZHQJqjJQzk6tF1fZQoaAZoCWgPQwjGo1TCa5SrQJSGlFKUaBVN6ANoFkdAmqiFPFefI3V9lChoBmgJaA9DCB+6oL79SK9AlIaUUpRoFU3oA2gWR0CasAyq+8GtdX2UKGgGaAloD0MIvko+dodfoECUhpRSlGgVTegDaBZHQJq10D2alUJ1fZQoaAZoCWgPQwhfCDnvP6mqQJSGlFKUaBVN6ANoFkdAmr1r3K0UoXV9lChoBmgJaA9DCN9t3jix+6JAlIaUUpRoFU3oA2gWR0CawvqTr3TNdX2UKGgGaAloD0MIjURoBGN3rECUhpRSlGgVTegDaBZHQJrKSbhFVkt1fZQoaAZoCWgPQwiRRZp4vz6uQJSGlFKUaBVN6ANoFkdAms+f029+PXV9lChoBmgJaA9DCOUJhJ1CtKpAlIaUUpRoFU3oA2gWR0Ca1xfwZwXJdX2UKGgGaAloD0MIsRafAlCprUCUhpRSlGgVTegDaBZHQJrcRzltCRh1fZQoaAZoCWgPQwjhQh7BzbyrQJSGlFKUaBVN6ANoFkdAmuM47/4qPXV9lChoBmgJaA9DCH8zMV0IOplAlIaUUpRoFU3oA2gWR0Ca6Id5Y5ktdX2UKGgGaAloD0MIXDl7Z6T0rUCUhpRSlGgVTegDaBZHQJrvlnscABF1fZQoaAZoCWgPQwibHhSUulqtQJSGlFKUaBVN6ANoFkdAmvTlZkkKNXV9lChoBmgJaA9DCKa4quzDMaZAlIaUUpRoFU3oA2gWR0Ca/DvBrN4adX2UKGgGaAloD0MI4sgDkU0JpUCUhpRSlGgVTegDaBZHQJsBn1xsEaF1fZQoaAZoCWgPQwh6NUBpiOiqQJSGlFKUaBVN6ANoFkdAmwjQr+YMOXV9lChoBmgJaA9DCG+5+rGJqqlAlIaUUpRoFU3oA2gWR0CbDg/d69kCdX2UKGgGaAloD0MI2PM1y8UKp0CUhpRSlGgVTegDaBZHQJsVlQbdadN1fZQoaAZoCWgPQwgHeT2Y/BaoQJSGlFKUaBVN6ANoFkdAmxsKi48U23V9lChoBmgJaA9DCN47akxgT6tAlIaUUpRoFU3oA2gWR0CbIjK2rn1WdX2UKGgGaAloD0MIrG9gcvsBq0CUhpRSlGgVTegDaBZHQJsnkksz2vl1fZQoaAZoCWgPQwj1g7pIOcapQJSGlFKUaBVN6ANoFkdAmy8OY6XBxnV9lChoBmgJaA9DCJRoyePBkqRAlIaUUpRoFU3oA2gWR0CbNLep4rz5dX2UKGgGaAloD0MIvk9VoRlyrUCUhpRSlGgVTegDaBZHQJs8LjwQUYd1fZQoaAZoCWgPQwhM+nspdLaqQJSGlFKUaBVN6ANoFkdAm1gfGhmGunV9lChoBmgJaA9DCG0f8pbbpqxAlIaUUpRoFU3oA2gWR0CbX6t8eCCjdX2UKGgGaAloD0MI3NWryMiVqkCUhpRSlGgVTegDaBZHQJtlIQ2/BWR1fZQoaAZoCWgPQwgDYDyDtpiqQJSGlFKUaBVN6ANoFkdAm2yjR6Ww/3V9lChoBmgJaA9DCJmdRe/MkqJAlIaUUpRoFU3oA2gWR0Cbclfr8iwCdX2UKGgGaAloD0MIFcWrrGWrrECUhpRSlGgVTegDaBZHQJt6FfTkQwt1fZQoaAZoCWgPQwjdIjDWh4KpQJSGlFKUaBVN6ANoFkdAm3+n5JsfrHV9lChoBmgJaA9DCL8Mxojs4adAlIaUUpRoFU3oA2gWR0CbhUFZgXuWdX2UKGgGaAloD0MINe1imgkcrECUhpRSlGgVTegDaBZHQJuMfDCP6sR1fZQoaAZoCWgPQwixMEROz/CrQJSGlFKUaBVN6ANoFkdAm5HXskY4yXV9lChoBmgJaA9DCDeo/dauOKpAlIaUUpRoFU3oA2gWR0CbmaOBUaQ4dX2UKGgGaAloD0MIGqN1VEUMp0CUhpRSlGgVTegDaBZHQJufLES/TLJ1fZQoaAZoCWgPQwhVwD3Pd0msQJSGlFKUaBVN6ANoFkdAm6bDRMN+b3V9lChoBmgJaA9DCKuy74oIuKlAlIaUUpRoFU3oA2gWR0CbrEXXAdn1dX2UKGgGaAloD0MIZapgVALIqUCUhpRSlGgVTegDaBZHQJuz7laKUFB1fZQoaAZoCWgPQwj+fFuwhHunQJSGlFKUaBVN6ANoFkdAm7mVj3Ehq3V9lChoBmgJaA9DCL9EvHXePq5AlIaUUpRoFU3oA2gWR0CbwSim2sq8dX2UKGgGaAloD0MI071O6nvpqkCUhpRSlGgVTegDaBZHQJvGqji4rjJ1fZQoaAZoCWgPQwjYLQJjfaWrQJSGlFKUaBVN6ANoFkdAm85DfWMCLnV9lChoBmgJaA9DCGk6Oxl0eqxAlIaUUpRoFU3oA2gWR0Cb0/pPRArydX2UKGgGaAloD0MI+BxYjjDiqUCUhpRSlGgVTegDaBZHQJvbnGff4yp1fZQoaAZoCWgPQwieCyO9MPCqQJSGlFKUaBVN6ANoFkdAm+FJLh73PHV9lChoBmgJaA9DCGk3+pin/6tAlIaUUpRoFU3oA2gWR0Cb6OCkXUH6dX2UKGgGaAloD0MIlpf8T05RqECUhpRSlGgVTegDaBZHQJvulMYdhiN1fZQoaAZoCWgPQwiI2jaMipGsQJSGlFKUaBVN6ANoFkdAm/Z3vDxb0XV9lChoBmgJaA9DCDvEP2whHqxAlIaUUpRoFU3oA2gWR0CcEXKvmozfdWUu"
|
| 97 |
},
|
| 98 |
"ep_success_buffer": {
|
| 99 |
":type:": "<class 'collections.deque'>",
|
results.json
CHANGED
|
@@ -1 +1 @@
|
|
| 1 |
-
{"mean_reward": 3416.0004174, "std_reward": 328.5927750661263, "is_deterministic": true, "n_eval_episodes": 10, "eval_datetime": "2022-07-
|
|
|
|
| 1 |
+
{"mean_reward": 3416.0004174, "std_reward": 328.5927750661263, "is_deterministic": true, "n_eval_episodes": 10, "eval_datetime": "2022-07-27T16:53:37.036060"}
|
train_eval_metrics.zip
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1b65c63408034d01239b3d278015626572d4ec32690500ee2fb7ac69f197d12b
|
| 3 |
+
size 33658
|