Initial commit
Browse files- README.md +19 -7
- args.yml +11 -5
- config.yml +12 -2
- ppo-seals-HalfCheetah-v0.zip +2 -2
- ppo-seals-HalfCheetah-v0/_stable_baselines3_version +1 -1
- ppo-seals-HalfCheetah-v0/data +24 -23
- ppo-seals-HalfCheetah-v0/policy.optimizer.pth +2 -2
- ppo-seals-HalfCheetah-v0/policy.pth +2 -2
- ppo-seals-HalfCheetah-v0/system_info.txt +2 -2
- replay.mp4 +2 -2
- results.json +1 -1
- train_eval_metrics.zip +2 -2
- vec_normalize.pkl +3 -0
README.md
CHANGED
|
@@ -10,7 +10,7 @@ model-index:
|
|
| 10 |
results:
|
| 11 |
- metrics:
|
| 12 |
- type: mean_reward
|
| 13 |
-
value:
|
| 14 |
name: mean_reward
|
| 15 |
task:
|
| 16 |
type: reinforcement-learning
|
|
@@ -37,15 +37,21 @@ SB3 Contrib: https://github.com/Stable-Baselines-Team/stable-baselines3-contrib
|
|
| 37 |
|
| 38 |
```
|
| 39 |
# Download model and save it into the logs/ folder
|
| 40 |
-
python -m
|
| 41 |
python enjoy.py --algo ppo --env seals/HalfCheetah-v0 -f logs/
|
| 42 |
```
|
| 43 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 44 |
## Training (with the RL Zoo)
|
| 45 |
```
|
| 46 |
python train.py --algo ppo --env seals/HalfCheetah-v0 -f logs/
|
| 47 |
# Upload the model and generate video (when possible)
|
| 48 |
-
python -m
|
| 49 |
```
|
| 50 |
|
| 51 |
## Hyperparameters
|
|
@@ -61,11 +67,17 @@ OrderedDict([('batch_size', 64),
|
|
| 61 |
('n_epochs', 5),
|
| 62 |
('n_steps', 512),
|
| 63 |
('n_timesteps', 1000000.0),
|
| 64 |
-
('normalize',
|
|
|
|
| 65 |
('policy', 'MlpPolicy'),
|
| 66 |
('policy_kwargs',
|
| 67 |
-
'
|
| 68 |
-
|
|
|
|
| 69 |
('vf_coef', 0.11483689492120866),
|
| 70 |
-
('normalize_kwargs',
|
|
|
|
|
|
|
|
|
|
|
|
|
| 71 |
```
|
|
|
|
| 10 |
results:
|
| 11 |
- metrics:
|
| 12 |
- type: mean_reward
|
| 13 |
+
value: 1755.78 +/- 45.04
|
| 14 |
name: mean_reward
|
| 15 |
task:
|
| 16 |
type: reinforcement-learning
|
|
|
|
| 37 |
|
| 38 |
```
|
| 39 |
# Download model and save it into the logs/ folder
|
| 40 |
+
python -m rl_zoo3.load_from_hub --algo ppo --env seals/HalfCheetah-v0 -orga HumanCompatibleAI -f logs/
|
| 41 |
python enjoy.py --algo ppo --env seals/HalfCheetah-v0 -f logs/
|
| 42 |
```
|
| 43 |
|
| 44 |
+
If you installed the RL Zoo3 via pip (`pip install rl_zoo3`), from anywhere you can do:
|
| 45 |
+
```
|
| 46 |
+
python -m rl_zoo3.load_from_hub --algo ppo --env seals/HalfCheetah-v0 -orga HumanCompatibleAI -f logs/
|
| 47 |
+
rl_zoo3 enjoy --algo ppo --env seals/HalfCheetah-v0 -f logs/
|
| 48 |
+
```
|
| 49 |
+
|
| 50 |
## Training (with the RL Zoo)
|
| 51 |
```
|
| 52 |
python train.py --algo ppo --env seals/HalfCheetah-v0 -f logs/
|
| 53 |
# Upload the model and generate video (when possible)
|
| 54 |
+
python -m rl_zoo3.push_to_hub --algo ppo --env seals/HalfCheetah-v0 -f logs/ -orga HumanCompatibleAI
|
| 55 |
```
|
| 56 |
|
| 57 |
## Hyperparameters
|
|
|
|
| 67 |
('n_epochs', 5),
|
| 68 |
('n_steps', 512),
|
| 69 |
('n_timesteps', 1000000.0),
|
| 70 |
+
('normalize',
|
| 71 |
+
{'gamma': 0.95, 'norm_obs': False, 'norm_reward': True}),
|
| 72 |
('policy', 'MlpPolicy'),
|
| 73 |
('policy_kwargs',
|
| 74 |
+
{'activation_fn': <class 'torch.nn.modules.activation.Tanh'>,
|
| 75 |
+
'features_extractor_class': <class 'imitation.policies.base.NormalizeFeaturesExtractor'>,
|
| 76 |
+
'net_arch': [{'pi': [64, 64], 'vf': [64, 64]}]}),
|
| 77 |
('vf_coef', 0.11483689492120866),
|
| 78 |
+
('normalize_kwargs',
|
| 79 |
+
{'norm_obs': {'gamma': 0.95,
|
| 80 |
+
'norm_obs': False,
|
| 81 |
+
'norm_reward': True},
|
| 82 |
+
'norm_reward': False})])
|
| 83 |
```
|
args.yml
CHANGED
|
@@ -1,6 +1,8 @@
|
|
| 1 |
!!python/object/apply:collections.OrderedDict
|
| 2 |
- - - algo
|
| 3 |
- ppo
|
|
|
|
|
|
|
| 4 |
- - device
|
| 5 |
- cpu
|
| 6 |
- - env
|
|
@@ -16,7 +18,7 @@
|
|
| 16 |
- - hyperparams
|
| 17 |
- null
|
| 18 |
- - log_folder
|
| 19 |
-
-
|
| 20 |
- - log_interval
|
| 21 |
- -1
|
| 22 |
- - max_total_trials
|
|
@@ -41,6 +43,8 @@
|
|
| 41 |
- null
|
| 42 |
- - optimize_hyperparameters
|
| 43 |
- false
|
|
|
|
|
|
|
| 44 |
- - pruner
|
| 45 |
- median
|
| 46 |
- - sampler
|
|
@@ -50,13 +54,13 @@
|
|
| 50 |
- - save_replay_buffer
|
| 51 |
- false
|
| 52 |
- - seed
|
| 53 |
-
-
|
| 54 |
- - storage
|
| 55 |
- null
|
| 56 |
- - study_name
|
| 57 |
- null
|
| 58 |
- - tensorboard_log
|
| 59 |
-
- runs/seals/HalfCheetah-
|
| 60 |
- - track
|
| 61 |
- true
|
| 62 |
- - trained_agent
|
|
@@ -70,6 +74,8 @@
|
|
| 70 |
- - verbose
|
| 71 |
- 1
|
| 72 |
- - wandb_entity
|
| 73 |
-
-
|
| 74 |
- - wandb_project_name
|
| 75 |
-
- seals-experts-
|
|
|
|
|
|
|
|
|
| 1 |
!!python/object/apply:collections.OrderedDict
|
| 2 |
- - - algo
|
| 3 |
- ppo
|
| 4 |
+
- - conf_file
|
| 5 |
+
- hyperparams/python/ppo.py
|
| 6 |
- - device
|
| 7 |
- cpu
|
| 8 |
- - env
|
|
|
|
| 18 |
- - hyperparams
|
| 19 |
- null
|
| 20 |
- - log_folder
|
| 21 |
+
- logs
|
| 22 |
- - log_interval
|
| 23 |
- -1
|
| 24 |
- - max_total_trials
|
|
|
|
| 43 |
- null
|
| 44 |
- - optimize_hyperparameters
|
| 45 |
- false
|
| 46 |
+
- - progress
|
| 47 |
+
- false
|
| 48 |
- - pruner
|
| 49 |
- median
|
| 50 |
- - sampler
|
|
|
|
| 54 |
- - save_replay_buffer
|
| 55 |
- false
|
| 56 |
- - seed
|
| 57 |
+
- 8
|
| 58 |
- - storage
|
| 59 |
- null
|
| 60 |
- - study_name
|
| 61 |
- null
|
| 62 |
- - tensorboard_log
|
| 63 |
+
- runs/seals/HalfCheetah-v0__ppo__8__1672325185
|
| 64 |
- - track
|
| 65 |
- true
|
| 66 |
- - trained_agent
|
|
|
|
| 74 |
- - verbose
|
| 75 |
- 1
|
| 76 |
- - wandb_entity
|
| 77 |
+
- ernestum
|
| 78 |
- - wandb_project_name
|
| 79 |
+
- seals-experts-normalized
|
| 80 |
+
- - yaml_file
|
| 81 |
+
- null
|
config.yml
CHANGED
|
@@ -22,10 +22,20 @@
|
|
| 22 |
- - n_timesteps
|
| 23 |
- 1000000.0
|
| 24 |
- - normalize
|
| 25 |
-
-
|
|
|
|
|
|
|
| 26 |
- - policy
|
| 27 |
- MlpPolicy
|
| 28 |
- - policy_kwargs
|
| 29 |
-
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 30 |
- - vf_coef
|
| 31 |
- 0.11483689492120866
|
|
|
|
| 22 |
- - n_timesteps
|
| 23 |
- 1000000.0
|
| 24 |
- - normalize
|
| 25 |
+
- gamma: 0.95
|
| 26 |
+
norm_obs: false
|
| 27 |
+
norm_reward: true
|
| 28 |
- - policy
|
| 29 |
- MlpPolicy
|
| 30 |
- - policy_kwargs
|
| 31 |
+
- activation_fn: !!python/name:torch.nn.modules.activation.Tanh ''
|
| 32 |
+
features_extractor_class: !!python/name:imitation.policies.base.NormalizeFeaturesExtractor ''
|
| 33 |
+
net_arch:
|
| 34 |
+
- pi:
|
| 35 |
+
- 64
|
| 36 |
+
- 64
|
| 37 |
+
vf:
|
| 38 |
+
- 64
|
| 39 |
+
- 64
|
| 40 |
- - vf_coef
|
| 41 |
- 0.11483689492120866
|
ppo-seals-HalfCheetah-v0.zip
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c576fa87c8f55328a7b83943438ff11ece009f095fa11cddd322285baffa063d
|
| 3 |
+
size 171669
|
ppo-seals-HalfCheetah-v0/_stable_baselines3_version
CHANGED
|
@@ -1 +1 @@
|
|
| 1 |
-
1.6.
|
|
|
|
| 1 |
+
1.6.2
|
ppo-seals-HalfCheetah-v0/data
CHANGED
|
@@ -4,24 +4,24 @@
|
|
| 4 |
":serialized:": "gAWVOwAAAAAAAACMIXN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi5wb2xpY2llc5SMEUFjdG9yQ3JpdGljUG9saWN5lJOULg==",
|
| 5 |
"__module__": "stable_baselines3.common.policies",
|
| 6 |
"__doc__": "\n Policy class for actor-critic algorithms (has both policy and value prediction).\n Used by A2C, PPO and the likes.\n\n :param observation_space: Observation space\n :param action_space: Action space\n :param lr_schedule: Learning rate schedule (could be constant)\n :param net_arch: The specification of the policy and value networks.\n :param activation_fn: Activation function\n :param ortho_init: Whether to use or not orthogonal initialization\n :param use_sde: Whether to use State Dependent Exploration or not\n :param log_std_init: Initial value for the log standard deviation\n :param full_std: Whether to use (n_features x n_actions) parameters\n for the std instead of only (n_features,) when using gSDE\n :param sde_net_arch: Network architecture for extracting features\n when using gSDE. If None, the latent features from the policy will be used.\n Pass an empty list to use the states as features.\n :param use_expln: Use ``expln()`` function instead of ``exp()`` to ensure\n a positive standard deviation (cf paper). It allows to keep variance\n above zero and prevent it from growing too fast. In practice, ``exp()`` is usually enough.\n :param squash_output: Whether to squash the output using a tanh function,\n this allows to ensure boundaries when using gSDE.\n :param features_extractor_class: Features extractor to use.\n :param features_extractor_kwargs: Keyword arguments\n to pass to the features extractor.\n :param normalize_images: Whether to normalize images or not,\n dividing by 255.0 (True by default)\n :param optimizer_class: The optimizer to use,\n ``th.optim.Adam`` by default\n :param optimizer_kwargs: Additional keyword arguments,\n excluding the learning rate, to pass to the optimizer\n ",
|
| 7 |
-
"__init__": "<function ActorCriticPolicy.__init__ at
|
| 8 |
-
"_get_constructor_parameters": "<function ActorCriticPolicy._get_constructor_parameters at
|
| 9 |
-
"reset_noise": "<function ActorCriticPolicy.reset_noise at
|
| 10 |
-
"_build_mlp_extractor": "<function ActorCriticPolicy._build_mlp_extractor at
|
| 11 |
-
"_build": "<function ActorCriticPolicy._build at
|
| 12 |
-
"forward": "<function ActorCriticPolicy.forward at
|
| 13 |
-
"_get_action_dist_from_latent": "<function ActorCriticPolicy._get_action_dist_from_latent at
|
| 14 |
-
"_predict": "<function ActorCriticPolicy._predict at
|
| 15 |
-
"evaluate_actions": "<function ActorCriticPolicy.evaluate_actions at
|
| 16 |
-
"get_distribution": "<function ActorCriticPolicy.get_distribution at
|
| 17 |
-
"predict_values": "<function ActorCriticPolicy.predict_values at
|
| 18 |
"__abstractmethods__": "frozenset()",
|
| 19 |
-
"_abc_impl": "<_abc_data object at
|
| 20 |
},
|
| 21 |
"verbose": 1,
|
| 22 |
"policy_kwargs": {
|
| 23 |
":type:": "<class 'dict'>",
|
| 24 |
-
":serialized:": "
|
| 25 |
"activation_fn": "<class 'torch.nn.modules.activation.Tanh'>",
|
| 26 |
"net_arch": [
|
| 27 |
{
|
|
@@ -34,7 +34,8 @@
|
|
| 34 |
64
|
| 35 |
]
|
| 36 |
}
|
| 37 |
-
]
|
|
|
|
| 38 |
},
|
| 39 |
"observation_space": {
|
| 40 |
":type:": "<class 'gym.spaces.box.Box'>",
|
|
@@ -51,7 +52,7 @@
|
|
| 51 |
},
|
| 52 |
"action_space": {
|
| 53 |
":type:": "<class 'gym.spaces.box.Box'>",
|
| 54 |
-
":serialized:": "gAWVEwwAAAAAAACMDmd5bS5zcGFjZXMuYm94lIwDQm94lJOUKYGUfZQojAVkdHlwZZSMBW51bXB5lGgFk5SMAmY0lImIh5RSlChLA4wBPJROTk5K/////0r/////SwB0lGKMBl9zaGFwZZRLBoWUjANsb3eUjBJudW1weS5jb3JlLm51bWVyaWOUjAtfZnJvbWJ1ZmZlcpSTlCiWGAAAAAAAAAAAAIC/AACAvwAAgL8AAIC/AACAvwAAgL+UaApLBoWUjAFDlHSUUpSMBGhpZ2iUaBIolhgAAAAAAAAAAACAPwAAgD8AAIA/AACAPwAAgD8AAIA/lGgKSwaFlGgVdJRSlIwNYm91bmRlZF9iZWxvd5RoEiiWBgAAAAAAAAABAQEBAQGUaAeMAmIxlImIh5RSlChLA4wBfJROTk5K/////0r/////
|
| 55 |
"dtype": "float32",
|
| 56 |
"_shape": [
|
| 57 |
6
|
|
@@ -66,17 +67,17 @@
|
|
| 66 |
"num_timesteps": 1000448,
|
| 67 |
"_total_timesteps": 1000000,
|
| 68 |
"_num_timesteps_at_start": 0,
|
| 69 |
-
"seed":
|
| 70 |
"action_noise": null,
|
| 71 |
-
"start_time":
|
| 72 |
"learning_rate": {
|
| 73 |
":type:": "<class 'function'>",
|
| 74 |
-
":serialized:": "
|
| 75 |
},
|
| 76 |
-
"tensorboard_log": "runs/seals/HalfCheetah-
|
| 77 |
"lr_schedule": {
|
| 78 |
":type:": "<class 'function'>",
|
| 79 |
-
":serialized:": "
|
| 80 |
},
|
| 81 |
"_last_obs": null,
|
| 82 |
"_last_episode_starts": {
|
|
@@ -85,7 +86,7 @@
|
|
| 85 |
},
|
| 86 |
"_last_original_obs": {
|
| 87 |
":type:": "<class 'numpy.ndarray'>",
|
| 88 |
-
":serialized:": "
|
| 89 |
},
|
| 90 |
"_episode_num": 0,
|
| 91 |
"use_sde": false,
|
|
@@ -93,7 +94,7 @@
|
|
| 93 |
"_current_progress_remaining": -0.00044800000000000395,
|
| 94 |
"ep_info_buffer": {
|
| 95 |
":type:": "<class 'collections.deque'>",
|
| 96 |
-
":serialized:": "gAWVgRAAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKUKH2UKIwBcpSMFW51bXB5LmNvcmUubXVsdGlhcnJheZSMBnNjYWxhcpSTlIwFbnVtcHmUjAVkdHlwZZSTlIwCZjiUiYiHlFKUKEsDjAE8lE5OTkr/////Sv////
|
| 97 |
},
|
| 98 |
"ep_success_buffer": {
|
| 99 |
":type:": "<class 'collections.deque'>",
|
|
@@ -110,7 +111,7 @@
|
|
| 110 |
"n_epochs": 5,
|
| 111 |
"clip_range": {
|
| 112 |
":type:": "<class 'function'>",
|
| 113 |
-
":serialized:": "
|
| 114 |
},
|
| 115 |
"clip_range_vf": null,
|
| 116 |
"normalize_advantage": true,
|
|
|
|
| 4 |
":serialized:": "gAWVOwAAAAAAAACMIXN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi5wb2xpY2llc5SMEUFjdG9yQ3JpdGljUG9saWN5lJOULg==",
|
| 5 |
"__module__": "stable_baselines3.common.policies",
|
| 6 |
"__doc__": "\n Policy class for actor-critic algorithms (has both policy and value prediction).\n Used by A2C, PPO and the likes.\n\n :param observation_space: Observation space\n :param action_space: Action space\n :param lr_schedule: Learning rate schedule (could be constant)\n :param net_arch: The specification of the policy and value networks.\n :param activation_fn: Activation function\n :param ortho_init: Whether to use or not orthogonal initialization\n :param use_sde: Whether to use State Dependent Exploration or not\n :param log_std_init: Initial value for the log standard deviation\n :param full_std: Whether to use (n_features x n_actions) parameters\n for the std instead of only (n_features,) when using gSDE\n :param sde_net_arch: Network architecture for extracting features\n when using gSDE. If None, the latent features from the policy will be used.\n Pass an empty list to use the states as features.\n :param use_expln: Use ``expln()`` function instead of ``exp()`` to ensure\n a positive standard deviation (cf paper). It allows to keep variance\n above zero and prevent it from growing too fast. In practice, ``exp()`` is usually enough.\n :param squash_output: Whether to squash the output using a tanh function,\n this allows to ensure boundaries when using gSDE.\n :param features_extractor_class: Features extractor to use.\n :param features_extractor_kwargs: Keyword arguments\n to pass to the features extractor.\n :param normalize_images: Whether to normalize images or not,\n dividing by 255.0 (True by default)\n :param optimizer_class: The optimizer to use,\n ``th.optim.Adam`` by default\n :param optimizer_kwargs: Additional keyword arguments,\n excluding the learning rate, to pass to the optimizer\n ",
|
| 7 |
+
"__init__": "<function ActorCriticPolicy.__init__ at 0x7fad92e96790>",
|
| 8 |
+
"_get_constructor_parameters": "<function ActorCriticPolicy._get_constructor_parameters at 0x7fad92e96820>",
|
| 9 |
+
"reset_noise": "<function ActorCriticPolicy.reset_noise at 0x7fad92e968b0>",
|
| 10 |
+
"_build_mlp_extractor": "<function ActorCriticPolicy._build_mlp_extractor at 0x7fad92e96940>",
|
| 11 |
+
"_build": "<function ActorCriticPolicy._build at 0x7fad92e969d0>",
|
| 12 |
+
"forward": "<function ActorCriticPolicy.forward at 0x7fad92e96a60>",
|
| 13 |
+
"_get_action_dist_from_latent": "<function ActorCriticPolicy._get_action_dist_from_latent at 0x7fad92e96af0>",
|
| 14 |
+
"_predict": "<function ActorCriticPolicy._predict at 0x7fad92e96b80>",
|
| 15 |
+
"evaluate_actions": "<function ActorCriticPolicy.evaluate_actions at 0x7fad92e96c10>",
|
| 16 |
+
"get_distribution": "<function ActorCriticPolicy.get_distribution at 0x7fad92e96ca0>",
|
| 17 |
+
"predict_values": "<function ActorCriticPolicy.predict_values at 0x7fad92e96d30>",
|
| 18 |
"__abstractmethods__": "frozenset()",
|
| 19 |
+
"_abc_impl": "<_abc_data object at 0x7fad92e8cb40>"
|
| 20 |
},
|
| 21 |
"verbose": 1,
|
| 22 |
"policy_kwargs": {
|
| 23 |
":type:": "<class 'dict'>",
|
| 24 |
+
":serialized:": "gAWVvAAAAAAAAAB9lCiMDWFjdGl2YXRpb25fZm6UjBt0b3JjaC5ubi5tb2R1bGVzLmFjdGl2YXRpb26UjARUYW5olJOUjAhuZXRfYXJjaJRdlH2UKIwCcGmUXZQoS0BLQGWMAnZmlF2UKEtAS0BldWGMGGZlYXR1cmVzX2V4dHJhY3Rvcl9jbGFzc5SMF2ltaXRhdGlvbi5wb2xpY2llcy5iYXNllIwaTm9ybWFsaXplRmVhdHVyZXNFeHRyYWN0b3KUk5R1Lg==",
|
| 25 |
"activation_fn": "<class 'torch.nn.modules.activation.Tanh'>",
|
| 26 |
"net_arch": [
|
| 27 |
{
|
|
|
|
| 34 |
64
|
| 35 |
]
|
| 36 |
}
|
| 37 |
+
],
|
| 38 |
+
"features_extractor_class": "<class 'imitation.policies.base.NormalizeFeaturesExtractor'>"
|
| 39 |
},
|
| 40 |
"observation_space": {
|
| 41 |
":type:": "<class 'gym.spaces.box.Box'>",
|
|
|
|
| 52 |
},
|
| 53 |
"action_space": {
|
| 54 |
":type:": "<class 'gym.spaces.box.Box'>",
|
| 55 |
+
":serialized:": "gAWVEwwAAAAAAACMDmd5bS5zcGFjZXMuYm94lIwDQm94lJOUKYGUfZQojAVkdHlwZZSMBW51bXB5lGgFk5SMAmY0lImIh5RSlChLA4wBPJROTk5K/////0r/////SwB0lGKMBl9zaGFwZZRLBoWUjANsb3eUjBJudW1weS5jb3JlLm51bWVyaWOUjAtfZnJvbWJ1ZmZlcpSTlCiWGAAAAAAAAAAAAIC/AACAvwAAgL8AAIC/AACAvwAAgL+UaApLBoWUjAFDlHSUUpSMBGhpZ2iUaBIolhgAAAAAAAAAAACAPwAAgD8AAIA/AACAPwAAgD8AAIA/lGgKSwaFlGgVdJRSlIwNYm91bmRlZF9iZWxvd5RoEiiWBgAAAAAAAAABAQEBAQGUaAeMAmIxlImIh5RSlChLA4wBfJROTk5K/////0r/////SwB0lGJLBoWUaBV0lFKUjA1ib3VuZGVkX2Fib3ZllGgSKJYGAAAAAAAAAAEBAQEBAZRoIUsGhZRoFXSUUpSMCl9ucF9yYW5kb22UjBRudW1weS5yYW5kb20uX3BpY2tsZZSMEl9fcmFuZG9tc3RhdGVfY3RvcpSTlIwHTVQxOTkzN5SFlFKUfZQojA1iaXRfZ2VuZXJhdG9ylGgwjAVzdGF0ZZR9lCiMA2tleZRoEiiWwAkAAAAAAAAAAACAXTYspE1QxanlRjSsWje2jremKvC3xUx9Z6aIk9tk5/bvbWs1m2usKUQk15IFBg+XGdbf5B2B6HWheHLVPzlCvunZEhAYAuRNx9JfqoPzg1XO5yNxKgCiGRxMGejVZrd90rKhuwE9awn00hWEoSmCxlJ/DekpK1kj0pZfJ+nbcJJNE3Ks3QLlHroFJW82nWcu43/9U1r4iMo34AkttQR12Vk0EvOG34kyQhmaETVqoqKNK273eDhYtcojPSpzwM3SUbJzk5628+tslzFtbgcSL9Qnd2w7CRogsbgzZYhhh+hxYTXhsGuhmKEHjme6gSUNt+cSdjS1qdQYgZ1jwb1chyhic6SQYrVvVXhQd/RUgmkjSKbFkfs1by5RK728jEzzdbNtgV0PglbkyTPjEMWyBAvLHNPlX7AFcaajYf6DEn0wEO4MEkzWanNIdyfqDF39pm7oRUVaRDqn4D119O0s2kmZsmHJi4525oUOzL/lsLAFdrc8ABVlvbCz3RfdYl3zs+I+FD3/PWl3kjk3czjurn5XFuurkv0Fh1AVnQRpWFDo1o29uN80IeDjAKEpLMpN2fD57hKekc+qDO1iW+qOJI8pZP65dSk464s0maGVPRW3sNPvreBgI8nmBCwB9NQYnyff3MOc1YPHoaWE/UW+yd5v4v6nb7DEOuQ+JRu4vxsOLCzGhOhi15sLiBNaJWMz/YNweMQLkrH/jqExtWZzBz2uouAUj6cf3yr0qY8jxOouffDNe3woWJx5ELCIp0IetPxLM2/lg+njmVXvHKEd1Y7dew9Qs1EPNtXsstDFdlm79I5uAYKqzuhZUbHhY1HV4rJidbylX3tpsdaoQ8O9n2ODME5Zi2OzcqxDMS5LBni30uDPw52AdXme1MBlgAO3UwSK27j187w+yzVN3qcstZKm30sPhQsz5rKYeH89QNQ3koN9ZgvjF/ybyRPbf4PJwOY+IWaTeowXipVdvEtadpy9NUCivq+qSq5JiElN/Jmkh0WbhW6tyyc/Let9ZD79I5tJLTO2fjWoXsKJ6nyDYuG6VTkAV30bT23kxy/MIvXHPm+DJZwXLSXRgVV1xoBU0Zc3VFA4bAnwAI08sfMJSm/MwG2xirbGVoCHM2F9Lm/1HO5PI7yMlCd4K61WDodnLhU02MToUrRoKy2J6yoGLdXO0yC4TqGh82fLl/YoQclWI9wEvsGu7X0Bsjz8T+aTiBLgeT0vUGnNxnG3XjTmvwvwVegD39dm0jbIelamIUjYwp2nnHgbx++U5wLSx8N7JXIOFmz17HazisWQxI6RxNFyCrvBzeY0N1P9xZ93oe3zdHSJVtJoqwb2PB4nrxnFWCSE88SmuP6nqMZBt4Me2eIhldSBQvtbsVkg2tOH+Zf4mRdJZwxfp14IPJIusqN/SgtxF3fLVKUbth5rPErSu4ivhx8lf+3X/YAoMmEIj4346w9uSA65Z5v68eMKAAFSLT6+7qsfu5NJ724IFJxfY7X61BBhNOORV73lz+sqiUlu1em5VsMzvAYogqRkaW1NUeQC6JEyuKQ8ctgajrEQ3bjQE2v9jYdeEirOPo/E37aMuMvhyeT58fg0nTtzdRrWk6U2zMhMwPkoXTGEfHE7ThaVHuIiDE6ktk1MlDhZPhNAYJElqVzGSTq1hnoplRRjCr88lUDDzTHSq9bv5RWDEwWYGUd1kmGx9JalKXBiXS/RjFTpjV1xQEsitGoUZhjY0nhjIWQxiLZX+ChTtxNeVm++JNxWIhxLGrxxlMveoW0XAF4FFiMevsGHXdh2II5FKWxKBxsWTentnMcRWMgz1k25dVsypZe/huTupX1LNT7v0QS0P5vo/KSaBrDuoSw5UvJdsdwjYcBu/5g0qR3tuEkcW782k4QL70gRHktFaYHjrxACDLk0+b5aDwZZ0tXnskHsST3j6sWZJKZPnkp0SsjVlpiWqvh63jja6jnBFoYm2bSUzn9eB+GiP85BiuNtwAfV9tiYZRvJ24xIVlS4prizcjyhSN88+BWFtt5tyLtobrOI4jxRb7wELiJ7gv/Kuyfhm23TF3FeWLSjRIlCFL+txD2WpNzdXGeF/1kXvRUOrR3xO7+G7+tdzlhyA8hLYI4UiAP7sOVXKOQw++STFz+4mAP9LwU4K/J0OWhO2URSL3rmjA953scHImTRN8rSAulKgoxVqs9yR9P78THjmZ/hISZ8IQFyCnc5EaEp45Rx5RiL8FkvJU3/N/7xUbVVU0aGbQdxmTdTc4wfqzGQxiR0++OgHOQpM085/YZsrNTtp+plPybJ105jW+51R+wLG6tZQfF8NdNorYz0uG8mcUpdlcuWfVRMp7YKwTP6dRmw1g3A13iDoQjaieZ0NPuZ6TX+/rT2DaLmGq4/fYZFK5e5ZYa5HAeM49lfuUrq4WlJhnLPR+2FhdNZDiBqe4nAp0FIS/jn9lRoRqtiLalzGyFaaL1uryqFFZqIp8vEn7uPMCV7NU1rbfyplpFIPByLc5TVVnD1AbeXNJyMB6PORoBlqSSFlJVz7NLxB6pafSkmg+TrBK0/SOGiYzZ8oTENdKh2Xkd4+oeX50lKC1F6qey1uagex7L7MAvgL+EgjBsg8d6AtCOjM/46Ws1/CxbBBKjltbBxGH4t+GO6qsCgRXXDZfmC22z/CQ5duxl2HGXWxfV7ks9zyBbqghobsB9wD8ZClXFq6rcHxttcG7MPDQpFiY5MCewFf1MOyVKLxMGWlEuXLC+80tDbHlRrmPqUkskltCMu8z414o6m2idF+relhiMHpLHFqL4zqIBZ0F/6FpigMkVZHffloXq2UULDKaDXA7kt08lmsfP4QS1VD9ITlnVpb3OPvIAzvi2H5Nk5FQLePOTMe55qUc51KlN46ERd/bVb1yTj08Jm/Q/hHkf+uActX/w/EnLXO5FK+lySlMSZkd1Cz/Y0DvQZ59rXNKEbD+6R3LpyDIQOe2D1vAhaSsEJBl/fuOGzKYOL1NmrTuXxgeeVhSts3b6dfZtLhLAmmQf3T4/zV0Um6heSyZIMat+mcjsaTnGgHlhWmsDKxCqpDywwzrl0rkl0b3HBB8ik/hA5jPHZWdYTAMII+OY9qRqUF0O10OOO05qaxo6pnrlQfFugZHuLGFQ6OZDztnORuIZ8X3eubfn+vguzLZ10+t9ywsMrzBvI5RjqI860YJcODnWtPcxH5mdzkJw7HMM9JshHfM3i32KXfpKrWKnD3qwz5JO2P+cyw61ClqH2anw4H7vFKi5DVNFn6pACRy7F8opmGZkmxUlAuooL2GESx9gUSPTGFdCtTUoCgMpB/BcSFo9dwBLu2m2UaAeMAnU0lImIh5RSlChLA2gLTk5OSv////9K/////0sAdJRiTXAChZRoFXSUUpSMA3Bvc5RNcAJ1jAloYXNfZ2F1c3OUSwCMBWdhdXNzlEcAAAAAAAAAAHVidWIu",
|
| 56 |
"dtype": "float32",
|
| 57 |
"_shape": [
|
| 58 |
6
|
|
|
|
| 67 |
"num_timesteps": 1000448,
|
| 68 |
"_total_timesteps": 1000000,
|
| 69 |
"_num_timesteps_at_start": 0,
|
| 70 |
+
"seed": 4,
|
| 71 |
"action_noise": null,
|
| 72 |
+
"start_time": 1672325190790677991,
|
| 73 |
"learning_rate": {
|
| 74 |
":type:": "<class 'function'>",
|
| 75 |
+
":serialized:": "gAWVhQIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwOX21ha2VfZnVuY3Rpb26Uk5QoaACMDV9idWlsdGluX3R5cGWUk5SMCENvZGVUeXBllIWUUpQoSwFLAEsASwFLAUsTQwSIAFMAlE6FlCmMAV+UhZSMUy9ob21lL21heGltaWxpYW4vdmVudi9saWIvcHl0aG9uMy44L3NpdGUtcGFja2FnZXMvc3RhYmxlX2Jhc2VsaW5lczMvY29tbW9uL3V0aWxzLnB5lIwEZnVuY5RLgEMCAAGUjAN2YWyUhZQpdJRSlH2UKIwLX19wYWNrYWdlX1+UjBhzdGFibGVfYmFzZWxpbmVzMy5jb21tb26UjAhfX25hbWVfX5SMHnN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi51dGlsc5SMCF9fZmlsZV9flGgMdU5OaACMEF9tYWtlX2VtcHR5X2NlbGyUk5QpUpSFlHSUUpSMHGNsb3VkcGlja2xlLmNsb3VkcGlja2xlX2Zhc3SUjBJfZnVuY3Rpb25fc2V0c3RhdGWUk5RoHn2UfZQoaBZoDYwMX19xdWFsbmFtZV9flIwZY29uc3RhbnRfZm4uPGxvY2Fscz4uZnVuY5SMD19fYW5ub3RhdGlvbnNfX5R9lIwOX19rd2RlZmF1bHRzX1+UTowMX19kZWZhdWx0c19flE6MCl9fbW9kdWxlX1+UaBeMB19fZG9jX1+UTowLX19jbG9zdXJlX1+UaACMCl9tYWtlX2NlbGyUk5RHPzWKdLD3DSmFlFKUhZSMF19jbG91ZHBpY2tsZV9zdWJtb2R1bGVzlF2UjAtfX2dsb2JhbHNfX5R9lHWGlIZSMC4="
|
| 76 |
},
|
| 77 |
+
"tensorboard_log": "runs/seals/HalfCheetah-v0__ppo__8__1672325185/seals-HalfCheetah-v0",
|
| 78 |
"lr_schedule": {
|
| 79 |
":type:": "<class 'function'>",
|
| 80 |
+
":serialized:": "gAWVhQIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwOX21ha2VfZnVuY3Rpb26Uk5QoaACMDV9idWlsdGluX3R5cGWUk5SMCENvZGVUeXBllIWUUpQoSwFLAEsASwFLAUsTQwSIAFMAlE6FlCmMAV+UhZSMUy9ob21lL21heGltaWxpYW4vdmVudi9saWIvcHl0aG9uMy44L3NpdGUtcGFja2FnZXMvc3RhYmxlX2Jhc2VsaW5lczMvY29tbW9uL3V0aWxzLnB5lIwEZnVuY5RLgEMCAAGUjAN2YWyUhZQpdJRSlH2UKIwLX19wYWNrYWdlX1+UjBhzdGFibGVfYmFzZWxpbmVzMy5jb21tb26UjAhfX25hbWVfX5SMHnN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi51dGlsc5SMCF9fZmlsZV9flGgMdU5OaACMEF9tYWtlX2VtcHR5X2NlbGyUk5QpUpSFlHSUUpSMHGNsb3VkcGlja2xlLmNsb3VkcGlja2xlX2Zhc3SUjBJfZnVuY3Rpb25fc2V0c3RhdGWUk5RoHn2UfZQoaBZoDYwMX19xdWFsbmFtZV9flIwZY29uc3RhbnRfZm4uPGxvY2Fscz4uZnVuY5SMD19fYW5ub3RhdGlvbnNfX5R9lIwOX19rd2RlZmF1bHRzX1+UTowMX19kZWZhdWx0c19flE6MCl9fbW9kdWxlX1+UaBeMB19fZG9jX1+UTowLX19jbG9zdXJlX1+UaACMCl9tYWtlX2NlbGyUk5RHPzWKdLD3DSmFlFKUhZSMF19jbG91ZHBpY2tsZV9zdWJtb2R1bGVzlF2UjAtfX2dsb2JhbHNfX5R9lHWGlIZSMC4="
|
| 81 |
},
|
| 82 |
"_last_obs": null,
|
| 83 |
"_last_episode_starts": {
|
|
|
|
| 86 |
},
|
| 87 |
"_last_original_obs": {
|
| 88 |
":type:": "<class 'numpy.ndarray'>",
|
| 89 |
+
":serialized:": "gAWVBQEAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJaQAAAAAAAAAEg2JxioDaA//AHDaK7jqj8YP1y+CRqMP1i27lnvD42/ZGbV3s28pL9clJGRPyWfv1zsQywjEqU/yBk/stMXsj8NGzEBjcKtv2cIt7b5n7g/73YLavTbzz9YY/P3HOu2PzhlL124x5u/wVkiyTDVpj8UJpJfdsiXP7i13WluZ7E/2PRffFnowD84ujvr2H2tv5SMBW51bXB5lIwFZHR5cGWUk5SMAmY4lImIh5RSlChLA4wBPJROTk5K/////0r/////SwB0lGJLAUsShpSMAUOUdJRSlC4="
|
| 90 |
},
|
| 91 |
"_episode_num": 0,
|
| 92 |
"use_sde": false,
|
|
|
|
| 94 |
"_current_progress_remaining": -0.00044800000000000395,
|
| 95 |
"ep_info_buffer": {
|
| 96 |
":type:": "<class 'collections.deque'>",
|
| 97 |
+
":serialized:": "gAWVgRAAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKUKH2UKIwBcpSMFW51bXB5LmNvcmUubXVsdGlhcnJheZSMBnNjYWxhcpSTlIwFbnVtcHmUjAVkdHlwZZSTlIwCZjiUiYiHlFKUKEsDjAE8lE5OTkr/////Sv////9LAHSUYkMI2ln0TrV/mkCUhpRSlIwBbJRN6AOMAXSUR0CdMy/BWPtEdX2UKGgGaAloD0MIvodLjotkmkCUhpRSlGgVTegDaBZHQJ05mxA0Kqp1fZQoaAZoCWgPQwi1iCgmn6+ZQJSGlFKUaBVN6ANoFkdAnT/7aEi+tnV9lChoBmgJaA9DCNHN/kCpsZlAlIaUUpRoFU3oA2gWR0CdRqpYLb5/dX2UKGgGaAloD0MIMshdhMmrmUCUhpRSlGgVTegDaBZHQJ1NKneizs11fZQoaAZoCWgPQwi693DJ4WmaQJSGlFKUaBVN6ANoFkdAnVQRcVxjrnV9lChoBmgJaA9DCCQO2UCa5ppAlIaUUpRoFU3oA2gWR0CdWd7jkuHvdX2UKGgGaAloD0MIkzXqIer2mUCUhpRSlGgVTegDaBZHQJ1iQ4Nqgyx1fZQoaAZoCWgPQwjacFgamAWZQJSGlFKUaBVN6ANoFkdAnWqqd+Xqq3V9lChoBmgJaA9DCK1p3nHql5lAlIaUUpRoFU3oA2gWR0CdcQnJDE3sdX2UKGgGaAloD0MIg6W6gPeCmUCUhpRSlGgVTegDaBZHQJ12uPbO/tZ1fZQoaAZoCWgPQwgK9l/nhlSZQJSGlFKUaBVN6ANoFkdAnXxpAQg9vHV9lChoBmgJaA9DCPa3BOBfEnbAlIaUUpRoFU3oA2gWR0Cdg9kwN9YwdX2UKGgGaAloD0MIoYMu4eDemUCUhpRSlGgVTegDaBZHQJ2MD7P6bfB1fZQoaAZoCWgPQwhAFMyYcuqZQJSGlFKUaBVN6ANoFkdAnZMlJcxCY3V9lChoBmgJaA9DCCy4H/AAHJpAlIaUUpRoFU3oA2gWR0CdmoeyAxzrdX2UKGgGaAloD0MInYL8bMQ3mkCUhpRSlGgVTegDaBZHQJ2i/L5hz/91fZQoaAZoCWgPQwgv205bwyuZQJSGlFKUaBVN6ANoFkdAnalz50r9VHV9lChoBmgJaA9DCPCmW3bYuJlAlIaUUpRoFU3oA2gWR0CdsPm7J4jbdX2UKGgGaAloD0MI2A3bFpWwmUCUhpRSlGgVTegDaBZHQJ23KksSTQp1fZQoaAZoCWgPQwinPLoRhjiaQJSGlFKUaBVN6ANoFkdAnby4hQm/nHV9lChoBmgJaA9DCNsWZTYIzZpAlIaUUpRoFU3oA2gWR0Cdw6aCcwxndX2UKGgGaAloD0MId4apLTWcmkCUhpRSlGgVTegDaBZHQJ3LN+UhV2l1fZQoaAZoCWgPQwiFlQoq6vyZQJSGlFKUaBVN6ANoFkdAndOcXvYvnXV9lChoBmgJaA9DCHHJcadUI5pAlIaUUpRoFU3oA2gWR0Cd2x6xxDLKdX2UKGgGaAloD0MIIZIhx5Y8mUCUhpRSlGgVTegDaBZHQJ30pvZRKpV1fZQoaAZoCWgPQwiIghlTMMl0wJSGlFKUaBVN6ANoFkdAnfsuJP69CnV9lChoBmgJaA9DCAcmN4pMGZlAlIaUUpRoFU3oA2gWR0CeAOb961LKdX2UKGgGaAloD0MIYviImFKJmkCUhpRSlGgVTegDaBZHQJ4G4hxHXmN1fZQoaAZoCWgPQwgAA0GAXNmZQJSGlFKUaBVN6ANoFkdAng5uD3/PxHV9lChoBmgJaA9DCLCryVOW1ZlAlIaUUpRoFU3oA2gWR0CeFR5D7ZWadX2UKGgGaAloD0MIzc03ojszdcCUhpRSlGgVTegDaBZHQJ4bRu89Oh11fZQoaAZoCWgPQwhQAMXIQryZQJSGlFKUaBVN6ANoFkdAniFkS7GvOnV9lChoBmgJaA9DCJzc71DUoplAlIaUUpRoFU3oA2gWR0CeKRZ/CqIadX2UKGgGaAloD0MI6nb2lYf4mUCUhpRSlGgVTegDaBZHQJ4wqC+UQkJ1fZQoaAZoCWgPQwjb96i/3u2ZQJSGlFKUaBVN6ANoFkdAnjZj3Ehq03V9lChoBmgJaA9DCMGopE7gRppAlIaUUpRoFU3oA2gWR0CePJYjSofkdX2UKGgGaAloD0MIWDuKcxTVmkCUhpRSlGgVTegDaBZHQJ5DvgYP5Hp1fZQoaAZoCWgPQwjrkJvhljaaQJSGlFKUaBVN6ANoFkdAnkl2S2Yv4HV9lChoBmgJaA9DCFA25Qofu5pAlIaUUpRoFU3oA2gWR0CeUCraufVadX2UKGgGaAloD0MIbQA2IDJJmkCUhpRSlGgVTegDaBZHQJ5XJbkfcN91fZQoaAZoCWgPQwg2kZkLTCSaQJSGlFKUaBVN6ANoFkdAnl0g1ivxIHV9lChoBmgJaA9DCC4dc55x2HTAlIaUUpRoFU3oA2gWR0CeZMvoNd7fdX2UKGgGaAloD0MIUps4uc9YmUCUhpRSlGgVTegDaBZHQJ5rh+MIeHV1fZQoaAZoCWgPQwh+calKS22aQJSGlFKUaBVN6ANoFkdAnnEdAxBVuXV9lChoBmgJaA9DCPyKNVw0dZpAlIaUUpRoFU3oA2gWR0Cedtm16Vt5dX2UKGgGaAloD0MI8MAAwvfzmUCUhpRSlGgVTegDaBZHQJ5/K508vEl1fZQoaAZoCWgPQwhwlpLldGWaQJSGlFKUaBVN6ANoFkdAnocOzQeFL3V9lChoBmgJaA9DCChJ10z+7ZhAlIaUUpRoFU3oA2gWR0CejZG+9Jz1dX2UKGgGaAloD0MIVwqBXGLfmkCUhpRSlGgVTegDaBZHQJ6TbxI8QqZ1fZQoaAZoCWgPQwjJdyl1aeKaQJSGlFKUaBVN6ANoFkdAnqgpRbbDdnV9lChoBmgJaA9DCMZq8/9q45pAlIaUUpRoFU3oA2gWR0Cerl/7BO58dX2UKGgGaAloD0MIhe0nYxxsmkCUhpRSlGgVTegDaBZHQJ62OFg2Ift1fZQoaAZoCWgPQwhH5SZqCQ6ZQJSGlFKUaBVN6ANoFkdAnr5BUR3/xXV9lChoBmgJaA9DCEGd8ui2m5pAlIaUUpRoFU3oA2gWR0CexcOUdJardX2UKGgGaAloD0MIIEJcOYsumkCUhpRSlGgVTegDaBZHQJ7L3Ai3XqZ1fZQoaAZoCWgPQwiMnfASPAabQJSGlFKUaBVN6ANoFkdAntIezt1IRXV9lChoBmgJaA9DCI80uK3d7JlAlIaUUpRoFU3oA2gWR0Ce1/vitJWedX2UKGgGaAloD0MIuw7VlITYmUCUhpRSlGgVTegDaBZHQJ7e8m/nGKh1fZQoaAZoCWgPQwiE1VjCipaaQJSGlFKUaBVN6ANoFkdAnuRfBN21UnV9lChoBmgJaA9DCDLMCdpkxZpAlIaUUpRoFU3oA2gWR0Ce6dnrpqyodX2UKGgGaAloD0MIIenTKjoydsCUhpRSlGgVTegDaBZHQJ7xHTAnDzl1fZQoaAZoCWgPQwj0/j9OGEaaQJSGlFKUaBVN6ANoFkdAnvjzBAOav3V9lChoBmgJaA9DCMPUljp4D5tAlIaUUpRoFU3oA2gWR0CfASHiWE9MdX2UKGgGaAloD0MIlrIMcZzhmUCUhpRSlGgVTegDaBZHQJ8JGxdIGyJ1fZQoaAZoCWgPQwiSW5Nu25GZQJSGlFKUaBVN6ANoFkdAnxBqXBxgiXV9lChoBmgJaA9DCDf8brqVZZpAlIaUUpRoFU3oA2gWR0CfGDbEP1+RdX2UKGgGaAloD0MIYeP6d42YmUCUhpRSlGgVTegDaBZHQJ8dv3dsSCh1fZQoaAZoCWgPQwhprWhznCqaQJSGlFKUaBVN6ANoFkdAnySSXt0FKXV9lChoBmgJaA9DCJrN4zDo9plAlIaUUpRoFU3oA2gWR0CfLWBOpKjBdX2UKGgGaAloD0MIZFjFG9kXmkCUhpRSlGgVTegDaBZHQJ824TXarWB1fZQoaAZoCWgPQwjgoL36KHuaQJSGlFKUaBVN6ANoFkdAnz3/NA1NxnV9lChoBmgJaA9DCNkiaTe6EZlAlIaUUpRoFU3oA2gWR0CfRJc3l0YCdX2UKGgGaAloD0MIsaVHU80DmkCUhpRSlGgVTegDaBZHQJ9KfCO3lS11fZQoaAZoCWgPQwgUsvM2VnKZQJSGlFKUaBVN6ANoFkdAn1NmGucME3V9lChoBmgJaA9DCErusIl8f5pAlIaUUpRoFU3oA2gWR0CfbewjdHlPdX2UKGgGaAloD0MIv51EhM/imUCUhpRSlGgVTegDaBZHQJ92Lz7MxGl1fZQoaAZoCWgPQwi2EyUhYbuZQJSGlFKUaBVN6ANoFkdAn32JR8+ianV9lChoBmgJaA9DCIVBmUYjzJpAlIaUUpRoFU3oA2gWR0Cfgwj5KvmpdX2UKGgGaAloD0MImx9/afHnmUCUhpRSlGgVTegDaBZHQJ+IoB6rvLJ1fZQoaAZoCWgPQwj44/bLRzSZQJSGlFKUaBVN6ANoFkdAn4/V0o0ALnV9lChoBmgJaA9DCBpQb0bNQppAlIaUUpRoFU3oA2gWR0CflQYoRZlndX2UKGgGaAloD0MIG7gDddpumkCUhpRSlGgVTegDaBZHQJ+adDRc/t91fZQoaAZoCWgPQwhwXpz4inGZQJSGlFKUaBVN6ANoFkdAn6KK8DjioHV9lChoBmgJaA9DCLO3lPNFAppAlIaUUpRoFU3oA2gWR0CfqnZjx0+1dX2UKGgGaAloD0MI2uGvyUocmUCUhpRSlGgVTegDaBZHQJ+yoY8+zMR1fZQoaAZoCWgPQwiZgjXOlmKaQJSGlFKUaBVN6ANoFkdAn7qNrCWNWHV9lChoBmgJaA9DCFhzgGDOXppAlIaUUpRoFU3oA2gWR0CfwvL+xW1ddX2UKGgGaAloD0MIbAn5oEcXmkCUhpRSlGgVTegDaBZHQJ/LUntv4ud1fZQoaAZoCWgPQwgcQL/vH2eaQJSGlFKUaBVN6ANoFkdAn9OVeF+NLnV9lChoBmgJaA9DCGnEzD6/hplAlIaUUpRoFU3oA2gWR0Cf2SFsYVIqdX2UKGgGaAloD0MIglZgyCoqmkCUhpRSlGgVTegDaBZHQJ/enS3LFGZ1fZQoaAZoCWgPQwigw3x5YcyZQJSGlFKUaBVN6ANoFkdAn+P1yJbdJ3V9lChoBmgJaA9DCLMpV3jnk5lAlIaUUpRoFU3oA2gWR0Cf6UubZvkzdX2UKGgGaAloD0MI/8u1aPH4mUCUhpRSlGgVTegDaBZHQJ/vW3solUp1fZQoaAZoCWgPQwgF24gne6maQJSGlFKUaBVN6ANoFkdAn/cnNX5nDnV9lChoBmgJaA9DCEUNpmFYlplAlIaUUpRoFU3oA2gWR0Cf//5d4VyndX2UKGgGaAloD0MIqrhxi/kFmkCUhpRSlGgVTegDaBZHQKAD+ivgWJt1fZQoaAZoCWgPQwi/gF64k0SaQJSGlFKUaBVN6ANoFkdAoAbYPPLPlnV9lChoBmgJaA9DCD9vKlIRP5lAlIaUUpRoFU3oA2gWR0CgCiA5Jbt7dWUu"
|
| 98 |
},
|
| 99 |
"ep_success_buffer": {
|
| 100 |
":type:": "<class 'collections.deque'>",
|
|
|
|
| 111 |
"n_epochs": 5,
|
| 112 |
"clip_range": {
|
| 113 |
":type:": "<class 'function'>",
|
| 114 |
+
":serialized:": "gAWVhQIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwOX21ha2VfZnVuY3Rpb26Uk5QoaACMDV9idWlsdGluX3R5cGWUk5SMCENvZGVUeXBllIWUUpQoSwFLAEsASwFLAUsTQwSIAFMAlE6FlCmMAV+UhZSMUy9ob21lL21heGltaWxpYW4vdmVudi9saWIvcHl0aG9uMy44L3NpdGUtcGFja2FnZXMvc3RhYmxlX2Jhc2VsaW5lczMvY29tbW9uL3V0aWxzLnB5lIwEZnVuY5RLgEMCAAGUjAN2YWyUhZQpdJRSlH2UKIwLX19wYWNrYWdlX1+UjBhzdGFibGVfYmFzZWxpbmVzMy5jb21tb26UjAhfX25hbWVfX5SMHnN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi51dGlsc5SMCF9fZmlsZV9flGgMdU5OaACMEF9tYWtlX2VtcHR5X2NlbGyUk5QpUpSFlHSUUpSMHGNsb3VkcGlja2xlLmNsb3VkcGlja2xlX2Zhc3SUjBJfZnVuY3Rpb25fc2V0c3RhdGWUk5RoHn2UfZQoaBZoDYwMX19xdWFsbmFtZV9flIwZY29uc3RhbnRfZm4uPGxvY2Fscz4uZnVuY5SMD19fYW5ub3RhdGlvbnNfX5R9lIwOX19rd2RlZmF1bHRzX1+UTowMX19kZWZhdWx0c19flE6MCl9fbW9kdWxlX1+UaBeMB19fZG9jX1+UTowLX19jbG9zdXJlX1+UaACMCl9tYWtlX2NlbGyUk5RHP7mZmZmZmZqFlFKUhZSMF19jbG91ZHBpY2tsZV9zdWJtb2R1bGVzlF2UjAtfX2dsb2JhbHNfX5R9lHWGlIZSMC4="
|
| 115 |
},
|
| 116 |
"clip_range_vf": null,
|
| 117 |
"normalize_advantage": true,
|
ppo-seals-HalfCheetah-v0/policy.optimizer.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f212469eeee5074b8816c5847ea4f3aefb53e33f41d1eac0e6e90951df87228b
|
| 3 |
+
size 99568
|
ppo-seals-HalfCheetah-v0/policy.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:307937f6af1a80ab71fb167c948683725431f97d2c9782d7a952ff0e30fb65fe
|
| 3 |
+
size 50165
|
ppo-seals-HalfCheetah-v0/system_info.txt
CHANGED
|
@@ -1,6 +1,6 @@
|
|
| 1 |
-
OS: Linux-5.4.0-
|
| 2 |
Python: 3.8.10
|
| 3 |
-
Stable-Baselines3: 1.6.
|
| 4 |
PyTorch: 1.11.0+cu102
|
| 5 |
GPU Enabled: False
|
| 6 |
Numpy: 1.22.3
|
|
|
|
| 1 |
+
OS: Linux-5.4.0-125-generic-x86_64-with-glibc2.29 #141-Ubuntu SMP Wed Aug 10 13:42:03 UTC 2022
|
| 2 |
Python: 3.8.10
|
| 3 |
+
Stable-Baselines3: 1.6.2
|
| 4 |
PyTorch: 1.11.0+cu102
|
| 5 |
GPU Enabled: False
|
| 6 |
Numpy: 1.22.3
|
replay.mp4
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7e2b7a8c3647fcbe65efea07a20d3619f7b6a0012aa29317f89482960513271f
|
| 3 |
+
size 1968711
|
results.json
CHANGED
|
@@ -1 +1 @@
|
|
| 1 |
-
{"mean_reward":
|
|
|
|
| 1 |
+
{"mean_reward": 1755.7794036, "std_reward": 45.035864951952206, "is_deterministic": true, "n_eval_episodes": 10, "eval_datetime": "2022-12-31T18:37:45.761755"}
|
train_eval_metrics.zip
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:82809eaa34e023f429e8a1d235850136bc38706a350ac1f8ae844e1974f03521
|
| 3 |
+
size 33656
|
vec_normalize.pkl
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4697be4d0613e7c9f394cb307138fb7f306d9386bcfc1c605d8dc6c18cb5fe1f
|
| 3 |
+
size 4373
|