Initial commit
Browse files- README.md +19 -7
- args.yml +11 -5
- config.yml +12 -2
- ppo-seals-Hopper-v0.zip +2 -2
- ppo-seals-Hopper-v0/_stable_baselines3_version +1 -1
- ppo-seals-Hopper-v0/data +24 -23
- ppo-seals-Hopper-v0/policy.optimizer.pth +2 -2
- ppo-seals-Hopper-v0/policy.pth +2 -2
- ppo-seals-Hopper-v0/system_info.txt +2 -2
- replay.mp4 +2 -2
- results.json +1 -1
- train_eval_metrics.zip +2 -2
- vec_normalize.pkl +3 -0
README.md
CHANGED
|
@@ -10,7 +10,7 @@ model-index:
|
|
| 10 |
results:
|
| 11 |
- metrics:
|
| 12 |
- type: mean_reward
|
| 13 |
-
value:
|
| 14 |
name: mean_reward
|
| 15 |
task:
|
| 16 |
type: reinforcement-learning
|
|
@@ -37,15 +37,21 @@ SB3 Contrib: https://github.com/Stable-Baselines-Team/stable-baselines3-contrib
|
|
| 37 |
|
| 38 |
```
|
| 39 |
# Download model and save it into the logs/ folder
|
| 40 |
-
python -m
|
| 41 |
python enjoy.py --algo ppo --env seals/Hopper-v0 -f logs/
|
| 42 |
```
|
| 43 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 44 |
## Training (with the RL Zoo)
|
| 45 |
```
|
| 46 |
python train.py --algo ppo --env seals/Hopper-v0 -f logs/
|
| 47 |
# Upload the model and generate video (when possible)
|
| 48 |
-
python -m
|
| 49 |
```
|
| 50 |
|
| 51 |
## Hyperparameters
|
|
@@ -61,11 +67,17 @@ OrderedDict([('batch_size', 512),
|
|
| 61 |
('n_epochs', 20),
|
| 62 |
('n_steps', 2048),
|
| 63 |
('n_timesteps', 1000000.0),
|
| 64 |
-
('normalize',
|
|
|
|
| 65 |
('policy', 'MlpPolicy'),
|
| 66 |
('policy_kwargs',
|
| 67 |
-
'
|
| 68 |
-
|
|
|
|
| 69 |
('vf_coef', 0.20315938606555833),
|
| 70 |
-
('normalize_kwargs',
|
|
|
|
|
|
|
|
|
|
|
|
|
| 71 |
```
|
|
|
|
| 10 |
results:
|
| 11 |
- metrics:
|
| 12 |
- type: mean_reward
|
| 13 |
+
value: 2862.69 +/- 127.20
|
| 14 |
name: mean_reward
|
| 15 |
task:
|
| 16 |
type: reinforcement-learning
|
|
|
|
| 37 |
|
| 38 |
```
|
| 39 |
# Download model and save it into the logs/ folder
|
| 40 |
+
python -m rl_zoo3.load_from_hub --algo ppo --env seals/Hopper-v0 -orga HumanCompatibleAI -f logs/
|
| 41 |
python enjoy.py --algo ppo --env seals/Hopper-v0 -f logs/
|
| 42 |
```
|
| 43 |
|
| 44 |
+
If you installed the RL Zoo3 via pip (`pip install rl_zoo3`), from anywhere you can do:
|
| 45 |
+
```
|
| 46 |
+
python -m rl_zoo3.load_from_hub --algo ppo --env seals/Hopper-v0 -orga HumanCompatibleAI -f logs/
|
| 47 |
+
rl_zoo3 enjoy --algo ppo --env seals/Hopper-v0 -f logs/
|
| 48 |
+
```
|
| 49 |
+
|
| 50 |
## Training (with the RL Zoo)
|
| 51 |
```
|
| 52 |
python train.py --algo ppo --env seals/Hopper-v0 -f logs/
|
| 53 |
# Upload the model and generate video (when possible)
|
| 54 |
+
python -m rl_zoo3.push_to_hub --algo ppo --env seals/Hopper-v0 -f logs/ -orga HumanCompatibleAI
|
| 55 |
```
|
| 56 |
|
| 57 |
## Hyperparameters
|
|
|
|
| 67 |
('n_epochs', 20),
|
| 68 |
('n_steps', 2048),
|
| 69 |
('n_timesteps', 1000000.0),
|
| 70 |
+
('normalize',
|
| 71 |
+
{'gamma': 0.995, 'norm_obs': False, 'norm_reward': True}),
|
| 72 |
('policy', 'MlpPolicy'),
|
| 73 |
('policy_kwargs',
|
| 74 |
+
{'activation_fn': <class 'torch.nn.modules.activation.ReLU'>,
|
| 75 |
+
'features_extractor_class': <class 'imitation.policies.base.NormalizeFeaturesExtractor'>,
|
| 76 |
+
'net_arch': [{'pi': [64, 64], 'vf': [64, 64]}]}),
|
| 77 |
('vf_coef', 0.20315938606555833),
|
| 78 |
+
('normalize_kwargs',
|
| 79 |
+
{'norm_obs': {'gamma': 0.995,
|
| 80 |
+
'norm_obs': False,
|
| 81 |
+
'norm_reward': True},
|
| 82 |
+
'norm_reward': False})])
|
| 83 |
```
|
args.yml
CHANGED
|
@@ -1,6 +1,8 @@
|
|
| 1 |
!!python/object/apply:collections.OrderedDict
|
| 2 |
- - - algo
|
| 3 |
- ppo
|
|
|
|
|
|
|
| 4 |
- - device
|
| 5 |
- cpu
|
| 6 |
- - env
|
|
@@ -16,7 +18,7 @@
|
|
| 16 |
- - hyperparams
|
| 17 |
- null
|
| 18 |
- - log_folder
|
| 19 |
-
-
|
| 20 |
- - log_interval
|
| 21 |
- -1
|
| 22 |
- - max_total_trials
|
|
@@ -41,6 +43,8 @@
|
|
| 41 |
- null
|
| 42 |
- - optimize_hyperparameters
|
| 43 |
- false
|
|
|
|
|
|
|
| 44 |
- - pruner
|
| 45 |
- median
|
| 46 |
- - sampler
|
|
@@ -50,13 +54,13 @@
|
|
| 50 |
- - save_replay_buffer
|
| 51 |
- false
|
| 52 |
- - seed
|
| 53 |
-
-
|
| 54 |
- - storage
|
| 55 |
- null
|
| 56 |
- - study_name
|
| 57 |
- null
|
| 58 |
- - tensorboard_log
|
| 59 |
-
- runs/seals/Hopper-
|
| 60 |
- - track
|
| 61 |
- true
|
| 62 |
- - trained_agent
|
|
@@ -70,6 +74,8 @@
|
|
| 70 |
- - verbose
|
| 71 |
- 1
|
| 72 |
- - wandb_entity
|
| 73 |
-
-
|
| 74 |
- - wandb_project_name
|
| 75 |
-
- seals-experts-
|
|
|
|
|
|
|
|
|
| 1 |
!!python/object/apply:collections.OrderedDict
|
| 2 |
- - - algo
|
| 3 |
- ppo
|
| 4 |
+
- - conf_file
|
| 5 |
+
- hyperparams/python/ppo.py
|
| 6 |
- - device
|
| 7 |
- cpu
|
| 8 |
- - env
|
|
|
|
| 18 |
- - hyperparams
|
| 19 |
- null
|
| 20 |
- - log_folder
|
| 21 |
+
- logs
|
| 22 |
- - log_interval
|
| 23 |
- -1
|
| 24 |
- - max_total_trials
|
|
|
|
| 43 |
- null
|
| 44 |
- - optimize_hyperparameters
|
| 45 |
- false
|
| 46 |
+
- - progress
|
| 47 |
+
- false
|
| 48 |
- - pruner
|
| 49 |
- median
|
| 50 |
- - sampler
|
|
|
|
| 54 |
- - save_replay_buffer
|
| 55 |
- false
|
| 56 |
- - seed
|
| 57 |
+
- 1
|
| 58 |
- - storage
|
| 59 |
- null
|
| 60 |
- - study_name
|
| 61 |
- null
|
| 62 |
- - tensorboard_log
|
| 63 |
+
- runs/seals/Hopper-v0__ppo__1__1672325330
|
| 64 |
- - track
|
| 65 |
- true
|
| 66 |
- - trained_agent
|
|
|
|
| 74 |
- - verbose
|
| 75 |
- 1
|
| 76 |
- - wandb_entity
|
| 77 |
+
- ernestum
|
| 78 |
- - wandb_project_name
|
| 79 |
+
- seals-experts-normalized
|
| 80 |
+
- - yaml_file
|
| 81 |
+
- null
|
config.yml
CHANGED
|
@@ -22,10 +22,20 @@
|
|
| 22 |
- - n_timesteps
|
| 23 |
- 1000000.0
|
| 24 |
- - normalize
|
| 25 |
-
-
|
|
|
|
|
|
|
| 26 |
- - policy
|
| 27 |
- MlpPolicy
|
| 28 |
- - policy_kwargs
|
| 29 |
-
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 30 |
- - vf_coef
|
| 31 |
- 0.20315938606555833
|
|
|
|
| 22 |
- - n_timesteps
|
| 23 |
- 1000000.0
|
| 24 |
- - normalize
|
| 25 |
+
- gamma: 0.995
|
| 26 |
+
norm_obs: false
|
| 27 |
+
norm_reward: true
|
| 28 |
- - policy
|
| 29 |
- MlpPolicy
|
| 30 |
- - policy_kwargs
|
| 31 |
+
- activation_fn: !!python/name:torch.nn.modules.activation.ReLU ''
|
| 32 |
+
features_extractor_class: !!python/name:imitation.policies.base.NormalizeFeaturesExtractor ''
|
| 33 |
+
net_arch:
|
| 34 |
+
- pi:
|
| 35 |
+
- 64
|
| 36 |
+
- 64
|
| 37 |
+
vf:
|
| 38 |
+
- 64
|
| 39 |
+
- 64
|
| 40 |
- - vf_coef
|
| 41 |
- 0.20315938606555833
|
ppo-seals-Hopper-v0.zip
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d5feefbff8e2d3a04ed48e9f111bd311dc4efa03b70d3dd80f99ce4869ebb832
|
| 3 |
+
size 159577
|
ppo-seals-Hopper-v0/_stable_baselines3_version
CHANGED
|
@@ -1 +1 @@
|
|
| 1 |
-
1.6.
|
|
|
|
| 1 |
+
1.6.2
|
ppo-seals-Hopper-v0/data
CHANGED
|
@@ -4,24 +4,24 @@
|
|
| 4 |
":serialized:": "gAWVOwAAAAAAAACMIXN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi5wb2xpY2llc5SMEUFjdG9yQ3JpdGljUG9saWN5lJOULg==",
|
| 5 |
"__module__": "stable_baselines3.common.policies",
|
| 6 |
"__doc__": "\n Policy class for actor-critic algorithms (has both policy and value prediction).\n Used by A2C, PPO and the likes.\n\n :param observation_space: Observation space\n :param action_space: Action space\n :param lr_schedule: Learning rate schedule (could be constant)\n :param net_arch: The specification of the policy and value networks.\n :param activation_fn: Activation function\n :param ortho_init: Whether to use or not orthogonal initialization\n :param use_sde: Whether to use State Dependent Exploration or not\n :param log_std_init: Initial value for the log standard deviation\n :param full_std: Whether to use (n_features x n_actions) parameters\n for the std instead of only (n_features,) when using gSDE\n :param sde_net_arch: Network architecture for extracting features\n when using gSDE. If None, the latent features from the policy will be used.\n Pass an empty list to use the states as features.\n :param use_expln: Use ``expln()`` function instead of ``exp()`` to ensure\n a positive standard deviation (cf paper). It allows to keep variance\n above zero and prevent it from growing too fast. In practice, ``exp()`` is usually enough.\n :param squash_output: Whether to squash the output using a tanh function,\n this allows to ensure boundaries when using gSDE.\n :param features_extractor_class: Features extractor to use.\n :param features_extractor_kwargs: Keyword arguments\n to pass to the features extractor.\n :param normalize_images: Whether to normalize images or not,\n dividing by 255.0 (True by default)\n :param optimizer_class: The optimizer to use,\n ``th.optim.Adam`` by default\n :param optimizer_kwargs: Additional keyword arguments,\n excluding the learning rate, to pass to the optimizer\n ",
|
| 7 |
-
"__init__": "<function ActorCriticPolicy.__init__ at
|
| 8 |
-
"_get_constructor_parameters": "<function ActorCriticPolicy._get_constructor_parameters at
|
| 9 |
-
"reset_noise": "<function ActorCriticPolicy.reset_noise at
|
| 10 |
-
"_build_mlp_extractor": "<function ActorCriticPolicy._build_mlp_extractor at
|
| 11 |
-
"_build": "<function ActorCriticPolicy._build at
|
| 12 |
-
"forward": "<function ActorCriticPolicy.forward at
|
| 13 |
-
"_get_action_dist_from_latent": "<function ActorCriticPolicy._get_action_dist_from_latent at
|
| 14 |
-
"_predict": "<function ActorCriticPolicy._predict at
|
| 15 |
-
"evaluate_actions": "<function ActorCriticPolicy.evaluate_actions at
|
| 16 |
-
"get_distribution": "<function ActorCriticPolicy.get_distribution at
|
| 17 |
-
"predict_values": "<function ActorCriticPolicy.predict_values at
|
| 18 |
"__abstractmethods__": "frozenset()",
|
| 19 |
-
"_abc_impl": "<_abc_data object at
|
| 20 |
},
|
| 21 |
"verbose": 1,
|
| 22 |
"policy_kwargs": {
|
| 23 |
":type:": "<class 'dict'>",
|
| 24 |
-
":serialized:": "
|
| 25 |
"activation_fn": "<class 'torch.nn.modules.activation.ReLU'>",
|
| 26 |
"net_arch": [
|
| 27 |
{
|
|
@@ -34,7 +34,8 @@
|
|
| 34 |
64
|
| 35 |
]
|
| 36 |
}
|
| 37 |
-
]
|
|
|
|
| 38 |
},
|
| 39 |
"observation_space": {
|
| 40 |
":type:": "<class 'gym.spaces.box.Box'>",
|
|
@@ -51,7 +52,7 @@
|
|
| 51 |
},
|
| 52 |
"action_space": {
|
| 53 |
":type:": "<class 'gym.spaces.box.Box'>",
|
| 54 |
-
":serialized:": "gAWV9QsAAAAAAACMDmd5bS5zcGFjZXMuYm94lIwDQm94lJOUKYGUfZQojAVkdHlwZZSMBW51bXB5lGgFk5SMAmY0lImIh5RSlChLA4wBPJROTk5K/////0r/////SwB0lGKMBl9zaGFwZZRLA4WUjANsb3eUjBJudW1weS5jb3JlLm51bWVyaWOUjAtfZnJvbWJ1ZmZlcpSTlCiWDAAAAAAAAAAAAIC/AACAvwAAgL+UaApLA4WUjAFDlHSUUpSMBGhpZ2iUaBIolgwAAAAAAAAAAACAPwAAgD8AAIA/lGgKSwOFlGgVdJRSlIwNYm91bmRlZF9iZWxvd5RoEiiWAwAAAAAAAAABAQGUaAeMAmIxlImIh5RSlChLA4wBfJROTk5K/////0r/////
|
| 55 |
"dtype": "float32",
|
| 56 |
"_shape": [
|
| 57 |
3
|
|
@@ -66,17 +67,17 @@
|
|
| 66 |
"num_timesteps": 1001472,
|
| 67 |
"_total_timesteps": 1000000,
|
| 68 |
"_num_timesteps_at_start": 0,
|
| 69 |
-
"seed":
|
| 70 |
"action_noise": null,
|
| 71 |
-
"start_time":
|
| 72 |
"learning_rate": {
|
| 73 |
":type:": "<class 'function'>",
|
| 74 |
-
":serialized:": "
|
| 75 |
},
|
| 76 |
-
"tensorboard_log": "runs/seals/Hopper-
|
| 77 |
"lr_schedule": {
|
| 78 |
":type:": "<class 'function'>",
|
| 79 |
-
":serialized:": "
|
| 80 |
},
|
| 81 |
"_last_obs": null,
|
| 82 |
"_last_episode_starts": {
|
|
@@ -85,7 +86,7 @@
|
|
| 85 |
},
|
| 86 |
"_last_original_obs": {
|
| 87 |
":type:": "<class 'numpy.ndarray'>",
|
| 88 |
-
":serialized:": "
|
| 89 |
},
|
| 90 |
"_episode_num": 0,
|
| 91 |
"use_sde": false,
|
|
@@ -93,7 +94,7 @@
|
|
| 93 |
"_current_progress_remaining": -0.0014719999999999178,
|
| 94 |
"ep_info_buffer": {
|
| 95 |
":type:": "<class 'collections.deque'>",
|
| 96 |
-
":serialized:": "gAWVgRAAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKUKH2UKIwBcpSMFW51bXB5LmNvcmUubXVsdGlhcnJheZSMBnNjYWxhcpSTlIwFbnVtcHmUjAVkdHlwZZSTlIwCZjiUiYiHlFKUKEsDjAE8lE5OTkr/////Sv////
|
| 97 |
},
|
| 98 |
"ep_success_buffer": {
|
| 99 |
":type:": "<class 'collections.deque'>",
|
|
@@ -110,7 +111,7 @@
|
|
| 110 |
"n_epochs": 20,
|
| 111 |
"clip_range": {
|
| 112 |
":type:": "<class 'function'>",
|
| 113 |
-
":serialized:": "
|
| 114 |
},
|
| 115 |
"clip_range_vf": null,
|
| 116 |
"normalize_advantage": true,
|
|
|
|
| 4 |
":serialized:": "gAWVOwAAAAAAAACMIXN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi5wb2xpY2llc5SMEUFjdG9yQ3JpdGljUG9saWN5lJOULg==",
|
| 5 |
"__module__": "stable_baselines3.common.policies",
|
| 6 |
"__doc__": "\n Policy class for actor-critic algorithms (has both policy and value prediction).\n Used by A2C, PPO and the likes.\n\n :param observation_space: Observation space\n :param action_space: Action space\n :param lr_schedule: Learning rate schedule (could be constant)\n :param net_arch: The specification of the policy and value networks.\n :param activation_fn: Activation function\n :param ortho_init: Whether to use or not orthogonal initialization\n :param use_sde: Whether to use State Dependent Exploration or not\n :param log_std_init: Initial value for the log standard deviation\n :param full_std: Whether to use (n_features x n_actions) parameters\n for the std instead of only (n_features,) when using gSDE\n :param sde_net_arch: Network architecture for extracting features\n when using gSDE. If None, the latent features from the policy will be used.\n Pass an empty list to use the states as features.\n :param use_expln: Use ``expln()`` function instead of ``exp()`` to ensure\n a positive standard deviation (cf paper). It allows to keep variance\n above zero and prevent it from growing too fast. In practice, ``exp()`` is usually enough.\n :param squash_output: Whether to squash the output using a tanh function,\n this allows to ensure boundaries when using gSDE.\n :param features_extractor_class: Features extractor to use.\n :param features_extractor_kwargs: Keyword arguments\n to pass to the features extractor.\n :param normalize_images: Whether to normalize images or not,\n dividing by 255.0 (True by default)\n :param optimizer_class: The optimizer to use,\n ``th.optim.Adam`` by default\n :param optimizer_kwargs: Additional keyword arguments,\n excluding the learning rate, to pass to the optimizer\n ",
|
| 7 |
+
"__init__": "<function ActorCriticPolicy.__init__ at 0x7f9554b75790>",
|
| 8 |
+
"_get_constructor_parameters": "<function ActorCriticPolicy._get_constructor_parameters at 0x7f9554b75820>",
|
| 9 |
+
"reset_noise": "<function ActorCriticPolicy.reset_noise at 0x7f9554b758b0>",
|
| 10 |
+
"_build_mlp_extractor": "<function ActorCriticPolicy._build_mlp_extractor at 0x7f9554b75940>",
|
| 11 |
+
"_build": "<function ActorCriticPolicy._build at 0x7f9554b759d0>",
|
| 12 |
+
"forward": "<function ActorCriticPolicy.forward at 0x7f9554b75a60>",
|
| 13 |
+
"_get_action_dist_from_latent": "<function ActorCriticPolicy._get_action_dist_from_latent at 0x7f9554b75af0>",
|
| 14 |
+
"_predict": "<function ActorCriticPolicy._predict at 0x7f9554b75b80>",
|
| 15 |
+
"evaluate_actions": "<function ActorCriticPolicy.evaluate_actions at 0x7f9554b75c10>",
|
| 16 |
+
"get_distribution": "<function ActorCriticPolicy.get_distribution at 0x7f9554b75ca0>",
|
| 17 |
+
"predict_values": "<function ActorCriticPolicy.predict_values at 0x7f9554b75d30>",
|
| 18 |
"__abstractmethods__": "frozenset()",
|
| 19 |
+
"_abc_impl": "<_abc_data object at 0x7f9554b6cc00>"
|
| 20 |
},
|
| 21 |
"verbose": 1,
|
| 22 |
"policy_kwargs": {
|
| 23 |
":type:": "<class 'dict'>",
|
| 24 |
+
":serialized:": "gAWVvAAAAAAAAAB9lCiMDWFjdGl2YXRpb25fZm6UjBt0b3JjaC5ubi5tb2R1bGVzLmFjdGl2YXRpb26UjARSZUxVlJOUjAhuZXRfYXJjaJRdlH2UKIwCcGmUXZQoS0BLQGWMAnZmlF2UKEtAS0BldWGMGGZlYXR1cmVzX2V4dHJhY3Rvcl9jbGFzc5SMF2ltaXRhdGlvbi5wb2xpY2llcy5iYXNllIwaTm9ybWFsaXplRmVhdHVyZXNFeHRyYWN0b3KUk5R1Lg==",
|
| 25 |
"activation_fn": "<class 'torch.nn.modules.activation.ReLU'>",
|
| 26 |
"net_arch": [
|
| 27 |
{
|
|
|
|
| 34 |
64
|
| 35 |
]
|
| 36 |
}
|
| 37 |
+
],
|
| 38 |
+
"features_extractor_class": "<class 'imitation.policies.base.NormalizeFeaturesExtractor'>"
|
| 39 |
},
|
| 40 |
"observation_space": {
|
| 41 |
":type:": "<class 'gym.spaces.box.Box'>",
|
|
|
|
| 52 |
},
|
| 53 |
"action_space": {
|
| 54 |
":type:": "<class 'gym.spaces.box.Box'>",
|
| 55 |
+
":serialized:": "gAWV9QsAAAAAAACMDmd5bS5zcGFjZXMuYm94lIwDQm94lJOUKYGUfZQojAVkdHlwZZSMBW51bXB5lGgFk5SMAmY0lImIh5RSlChLA4wBPJROTk5K/////0r/////SwB0lGKMBl9zaGFwZZRLA4WUjANsb3eUjBJudW1weS5jb3JlLm51bWVyaWOUjAtfZnJvbWJ1ZmZlcpSTlCiWDAAAAAAAAAAAAIC/AACAvwAAgL+UaApLA4WUjAFDlHSUUpSMBGhpZ2iUaBIolgwAAAAAAAAAAACAPwAAgD8AAIA/lGgKSwOFlGgVdJRSlIwNYm91bmRlZF9iZWxvd5RoEiiWAwAAAAAAAAABAQGUaAeMAmIxlImIh5RSlChLA4wBfJROTk5K/////0r/////SwB0lGJLA4WUaBV0lFKUjA1ib3VuZGVkX2Fib3ZllGgSKJYDAAAAAAAAAAEBAZRoIUsDhZRoFXSUUpSMCl9ucF9yYW5kb22UjBRudW1weS5yYW5kb20uX3BpY2tsZZSMEl9fcmFuZG9tc3RhdGVfY3RvcpSTlIwHTVQxOTkzN5SFlFKUfZQojA1iaXRfZ2VuZXJhdG9ylGgwjAVzdGF0ZZR9lCiMA2tleZRoEiiWwAkAAAAAAAAAAACAXTYspE1QxanlRjSsWje2jremKvC3xUx9Z6aIk9tk5/bvbWs1m2usKUQk15IFBg+XGdbf5B2B6HWheHLVPzlCvunZEhAYAuRNx9JfqoPzg1XO5yNxKgCiGRxMGejVZrd90rKhuwE9awn00hWEoSmCxlJ/DekpK1kj0pZfJ+nbcJJNE3Ks3QLlHroFJW82nWcu43/9U1r4iMo34AkttQR12Vk0EvOG34kyQhmaETVqoqKNK273eDhYtcojPSpzwM3SUbJzk5628+tslzFtbgcSL9Qnd2w7CRogsbgzZYhhh+hxYTXhsGuhmKEHjme6gSUNt+cSdjS1qdQYgZ1jwb1chyhic6SQYrVvVXhQd/RUgmkjSKbFkfs1by5RK728jEzzdbNtgV0PglbkyTPjEMWyBAvLHNPlX7AFcaajYf6DEn0wEO4MEkzWanNIdyfqDF39pm7oRUVaRDqn4D119O0s2kmZsmHJi4525oUOzL/lsLAFdrc8ABVlvbCz3RfdYl3zs+I+FD3/PWl3kjk3czjurn5XFuurkv0Fh1AVnQRpWFDo1o29uN80IeDjAKEpLMpN2fD57hKekc+qDO1iW+qOJI8pZP65dSk464s0maGVPRW3sNPvreBgI8nmBCwB9NQYnyff3MOc1YPHoaWE/UW+yd5v4v6nb7DEOuQ+JRu4vxsOLCzGhOhi15sLiBNaJWMz/YNweMQLkrH/jqExtWZzBz2uouAUj6cf3yr0qY8jxOouffDNe3woWJx5ELCIp0IetPxLM2/lg+njmVXvHKEd1Y7dew9Qs1EPNtXsstDFdlm79I5uAYKqzuhZUbHhY1HV4rJidbylX3tpsdaoQ8O9n2ODME5Zi2OzcqxDMS5LBni30uDPw52AdXme1MBlgAO3UwSK27j187w+yzVN3qcstZKm30sPhQsz5rKYeH89QNQ3koN9ZgvjF/ybyRPbf4PJwOY+IWaTeowXipVdvEtadpy9NUCivq+qSq5JiElN/Jmkh0WbhW6tyyc/Let9ZD79I5tJLTO2fjWoXsKJ6nyDYuG6VTkAV30bT23kxy/MIvXHPm+DJZwXLSXRgVV1xoBU0Zc3VFA4bAnwAI08sfMJSm/MwG2xirbGVoCHM2F9Lm/1HO5PI7yMlCd4K61WDodnLhU02MToUrRoKy2J6yoGLdXO0yC4TqGh82fLl/YoQclWI9wEvsGu7X0Bsjz8T+aTiBLgeT0vUGnNxnG3XjTmvwvwVegD39dm0jbIelamIUjYwp2nnHgbx++U5wLSx8N7JXIOFmz17HazisWQxI6RxNFyCrvBzeY0N1P9xZ93oe3zdHSJVtJoqwb2PB4nrxnFWCSE88SmuP6nqMZBt4Me2eIhldSBQvtbsVkg2tOH+Zf4mRdJZwxfp14IPJIusqN/SgtxF3fLVKUbth5rPErSu4ivhx8lf+3X/YAoMmEIj4346w9uSA65Z5v68eMKAAFSLT6+7qsfu5NJ724IFJxfY7X61BBhNOORV73lz+sqiUlu1em5VsMzvAYogqRkaW1NUeQC6JEyuKQ8ctgajrEQ3bjQE2v9jYdeEirOPo/E37aMuMvhyeT58fg0nTtzdRrWk6U2zMhMwPkoXTGEfHE7ThaVHuIiDE6ktk1MlDhZPhNAYJElqVzGSTq1hnoplRRjCr88lUDDzTHSq9bv5RWDEwWYGUd1kmGx9JalKXBiXS/RjFTpjV1xQEsitGoUZhjY0nhjIWQxiLZX+ChTtxNeVm++JNxWIhxLGrxxlMveoW0XAF4FFiMevsGHXdh2II5FKWxKBxsWTentnMcRWMgz1k25dVsypZe/huTupX1LNT7v0QS0P5vo/KSaBrDuoSw5UvJdsdwjYcBu/5g0qR3tuEkcW782k4QL70gRHktFaYHjrxACDLk0+b5aDwZZ0tXnskHsST3j6sWZJKZPnkp0SsjVlpiWqvh63jja6jnBFoYm2bSUzn9eB+GiP85BiuNtwAfV9tiYZRvJ24xIVlS4prizcjyhSN88+BWFtt5tyLtobrOI4jxRb7wELiJ7gv/Kuyfhm23TF3FeWLSjRIlCFL+txD2WpNzdXGeF/1kXvRUOrR3xO7+G7+tdzlhyA8hLYI4UiAP7sOVXKOQw++STFz+4mAP9LwU4K/J0OWhO2URSL3rmjA953scHImTRN8rSAulKgoxVqs9yR9P78THjmZ/hISZ8IQFyCnc5EaEp45Rx5RiL8FkvJU3/N/7xUbVVU0aGbQdxmTdTc4wfqzGQxiR0++OgHOQpM085/YZsrNTtp+plPybJ105jW+51R+wLG6tZQfF8NdNorYz0uG8mcUpdlcuWfVRMp7YKwTP6dRmw1g3A13iDoQjaieZ0NPuZ6TX+/rT2DaLmGq4/fYZFK5e5ZYa5HAeM49lfuUrq4WlJhnLPR+2FhdNZDiBqe4nAp0FIS/jn9lRoRqtiLalzGyFaaL1uryqFFZqIp8vEn7uPMCV7NU1rbfyplpFIPByLc5TVVnD1AbeXNJyMB6PORoBlqSSFlJVz7NLxB6pafSkmg+TrBK0/SOGiYzZ8oTENdKh2Xkd4+oeX50lKC1F6qey1uagex7L7MAvgL+EgjBsg8d6AtCOjM/46Ws1/CxbBBKjltbBxGH4t+GO6qsCgRXXDZfmC22z/CQ5duxl2HGXWxfV7ks9zyBbqghobsB9wD8ZClXFq6rcHxttcG7MPDQpFiY5MCewFf1MOyVKLxMGWlEuXLC+80tDbHlRrmPqUkskltCMu8z414o6m2idF+relhiMHpLHFqL4zqIBZ0F/6FpigMkVZHffloXq2UULDKaDXA7kt08lmsfP4QS1VD9ITlnVpb3OPvIAzvi2H5Nk5FQLePOTMe55qUc51KlN46ERd/bVb1yTj08Jm/Q/hHkf+uActX/w/EnLXO5FK+lySlMSZkd1Cz/Y0DvQZ59rXNKEbD+6R3LpyDIQOe2D1vAhaSsEJBl/fuOGzKYOL1NmrTuXxgeeVhSts3b6dfZtLhLAmmQf3T4/zV0Um6heSyZIMat+mcjsaTnGgHlhWmsDKxCqpDywwzrl0rkl0b3HBB8ik/hA5jPHZWdYTAMII+OY9qRqUF0O10OOO05qaxo6pnrlQfFugZHuLGFQ6OZDztnORuIZ8X3eubfn+vguzLZ10+t9ywsMrzBvI5RjqI860YJcODnWtPcxH5mdzkJw7HMM9JshHfM3i32KXfpKrWKnD3qwz5JO2P+cyw61ClqH2anw4H7vFKi5DVNFn6pACRy7F8opmGZkmxUlAuooL2GESx9gUSPTGFdCtTUoCgMpB/BcSFo9dwBLu2m2UaAeMAnU0lImIh5RSlChLA2gLTk5OSv////9K/////0sAdJRiTXAChZRoFXSUUpSMA3Bvc5RNcAJ1jAloYXNfZ2F1c3OUSwCMBWdhdXNzlEcAAAAAAAAAAHVidWIu",
|
| 56 |
"dtype": "float32",
|
| 57 |
"_shape": [
|
| 58 |
3
|
|
|
|
| 67 |
"num_timesteps": 1001472,
|
| 68 |
"_total_timesteps": 1000000,
|
| 69 |
"_num_timesteps_at_start": 0,
|
| 70 |
+
"seed": 4,
|
| 71 |
"action_noise": null,
|
| 72 |
+
"start_time": 1672325335203760040,
|
| 73 |
"learning_rate": {
|
| 74 |
":type:": "<class 'function'>",
|
| 75 |
+
":serialized:": "gAWVhQIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwOX21ha2VfZnVuY3Rpb26Uk5QoaACMDV9idWlsdGluX3R5cGWUk5SMCENvZGVUeXBllIWUUpQoSwFLAEsASwFLAUsTQwSIAFMAlE6FlCmMAV+UhZSMUy9ob21lL21heGltaWxpYW4vdmVudi9saWIvcHl0aG9uMy44L3NpdGUtcGFja2FnZXMvc3RhYmxlX2Jhc2VsaW5lczMvY29tbW9uL3V0aWxzLnB5lIwEZnVuY5RLgEMCAAGUjAN2YWyUhZQpdJRSlH2UKIwLX19wYWNrYWdlX1+UjBhzdGFibGVfYmFzZWxpbmVzMy5jb21tb26UjAhfX25hbWVfX5SMHnN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi51dGlsc5SMCF9fZmlsZV9flGgMdU5OaACMEF9tYWtlX2VtcHR5X2NlbGyUk5QpUpSFlHSUUpSMHGNsb3VkcGlja2xlLmNsb3VkcGlja2xlX2Zhc3SUjBJfZnVuY3Rpb25fc2V0c3RhdGWUk5RoHn2UfZQoaBZoDYwMX19xdWFsbmFtZV9flIwZY29uc3RhbnRfZm4uPGxvY2Fscz4uZnVuY5SMD19fYW5ub3RhdGlvbnNfX5R9lIwOX19rd2RlZmF1bHRzX1+UTowMX19kZWZhdWx0c19flE6MCl9fbW9kdWxlX1+UaBeMB19fZG9jX1+UTowLX19jbG9zdXJlX1+UaACMCl9tYWtlX2NlbGyUk5RHPzmXHiNxn8uFlFKUhZSMF19jbG91ZHBpY2tsZV9zdWJtb2R1bGVzlF2UjAtfX2dsb2JhbHNfX5R9lHWGlIZSMC4="
|
| 76 |
},
|
| 77 |
+
"tensorboard_log": "runs/seals/Hopper-v0__ppo__1__1672325330/seals-Hopper-v0",
|
| 78 |
"lr_schedule": {
|
| 79 |
":type:": "<class 'function'>",
|
| 80 |
+
":serialized:": "gAWVhQIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwOX21ha2VfZnVuY3Rpb26Uk5QoaACMDV9idWlsdGluX3R5cGWUk5SMCENvZGVUeXBllIWUUpQoSwFLAEsASwFLAUsTQwSIAFMAlE6FlCmMAV+UhZSMUy9ob21lL21heGltaWxpYW4vdmVudi9saWIvcHl0aG9uMy44L3NpdGUtcGFja2FnZXMvc3RhYmxlX2Jhc2VsaW5lczMvY29tbW9uL3V0aWxzLnB5lIwEZnVuY5RLgEMCAAGUjAN2YWyUhZQpdJRSlH2UKIwLX19wYWNrYWdlX1+UjBhzdGFibGVfYmFzZWxpbmVzMy5jb21tb26UjAhfX25hbWVfX5SMHnN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi51dGlsc5SMCF9fZmlsZV9flGgMdU5OaACMEF9tYWtlX2VtcHR5X2NlbGyUk5QpUpSFlHSUUpSMHGNsb3VkcGlja2xlLmNsb3VkcGlja2xlX2Zhc3SUjBJfZnVuY3Rpb25fc2V0c3RhdGWUk5RoHn2UfZQoaBZoDYwMX19xdWFsbmFtZV9flIwZY29uc3RhbnRfZm4uPGxvY2Fscz4uZnVuY5SMD19fYW5ub3RhdGlvbnNfX5R9lIwOX19rd2RlZmF1bHRzX1+UTowMX19kZWZhdWx0c19flE6MCl9fbW9kdWxlX1+UaBeMB19fZG9jX1+UTowLX19jbG9zdXJlX1+UaACMCl9tYWtlX2NlbGyUk5RHPzmXHiNxn8uFlFKUhZSMF19jbG91ZHBpY2tsZV9zdWJtb2R1bGVzlF2UjAtfX2dsb2JhbHNfX5R9lHWGlIZSMC4="
|
| 81 |
},
|
| 82 |
"_last_obs": null,
|
| 83 |
"_last_episode_starts": {
|
|
|
|
| 86 |
},
|
| 87 |
"_last_original_obs": {
|
| 88 |
":type:": "<class 'numpy.ndarray'>",
|
| 89 |
+
":serialized:": "gAWV1QAAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJZgAAAAAAAAAMJV8NVxLmk/V5hVC5gA9D+SMYTMuE1pv5dpCTe2pmm/OBYZZTFTZj8wkQHYbFtyv3Ak0UjYoXG/oAXRDO3jIz9crOH7F+RePyCRoR9IdTq/5ZujsWR3cD/QIT3CxoxmP5SMBW51bXB5lIwFZHR5cGWUk5SMAmY4lImIh5RSlChLA4wBPJROTk5K/////0r/////SwB0lGJLAUsMhpSMAUOUdJRSlC4="
|
| 90 |
},
|
| 91 |
"_episode_num": 0,
|
| 92 |
"use_sde": false,
|
|
|
|
| 94 |
"_current_progress_remaining": -0.0014719999999999178,
|
| 95 |
"ep_info_buffer": {
|
| 96 |
":type:": "<class 'collections.deque'>",
|
| 97 |
+
":serialized:": "gAWVgRAAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKUKH2UKIwBcpSMFW51bXB5LmNvcmUubXVsdGlhcnJheZSMBnNjYWxhcpSTlIwFbnVtcHmUjAVkdHlwZZSTlIwCZjiUiYiHlFKUKEsDjAE8lE5OTkr/////Sv////9LAHSUYkMIa+9TVWD3pECUhpRSlIwBbJRN6AOMAXSUR0CcO+jesPrfdX2UKGgGaAloD0MIN24xP+edpECUhpRSlGgVTegDaBZHQJxAZLBbfP51fZQoaAZoCWgPQwic4QZ8hsKkQJSGlFKUaBVN6ANoFkdAnEdfMGHHm3V9lChoBmgJaA9DCKBSJcpOu6RAlIaUUpRoFU3oA2gWR0CcS/nWattAdX2UKGgGaAloD0MInz2XqQltpUCUhpRSlGgVTegDaBZHQJxS3s9jgAJ1fZQoaAZoCWgPQwhjQswl3aulQJSGlFKUaBVN6ANoFkdAnFfZ/smfG3V9lChoBmgJaA9DCHIYzF85LqVAlIaUUpRoFU3oA2gWR0CcXhKJ2t+1dX2UKGgGaAloD0MI2ekHdYncokCUhpRSlGgVTegDaBZHQJxilhhH9WJ1fZQoaAZoCWgPQwhSnnk57D6lQJSGlFKUaBVN6ANoFkdAnGk2UOd5IHV9lChoBmgJaA9DCHIZNzVY4qVAlIaUUpRoFU3oA2gWR0CcbaqlP8AJdX2UKGgGaAloD0MI1PAtrJNhpECUhpRSlGgVTegDaBZHQJx0FSBK+SN1fZQoaAZoCWgPQwgbDeAtwMGiQJSGlFKUaBVN6ANoFkdAnHhmsRxtHnV9lChoBmgJaA9DCMK9Mm/1x6RAlIaUUpRoFU3oA2gWR0CcfshL5AQhdX2UKGgGaAloD0MIxOv6BUOOpUCUhpRSlGgVTegDaBZHQJyDQaqCHyp1fZQoaAZoCWgPQwh4DI/91N6kQJSGlFKUaBVN6ANoFkdAnIoEqYqoZXV9lChoBmgJaA9DCIDwoURL5adAlIaUUpRoFU3oA2gWR0Ccjmxfv4M4dX2UKGgGaAloD0MIQS0GD/N1pUCUhpRSlGgVTegDaBZHQJyU6W4Vh1F1fZQoaAZoCWgPQwig3oyar+SkQJSGlFKUaBVN6ANoFkdAnJmAFX7tRnV9lChoBmgJaA9DCKmluRXSdaVAlIaUUpRoFU3oA2gWR0CcoH/TLGJfdX2UKGgGaAloD0MIKLou/Fi1pUCUhpRSlGgVTegDaBZHQJylLoKUmlZ1fZQoaAZoCWgPQwizXgzlJGCnQJSGlFKUaBVN6ANoFkdAnKuuyu6mO3V9lChoBmgJaA9DCACrI0eyT6ZAlIaUUpRoFU3oA2gWR0CcsBbjcVQAdX2UKGgGaAloD0MIHO+OjK0npECUhpRSlGgVTegDaBZHQJy2SnO0LMN1fZQoaAZoCWgPQwjcZFQZ1jalQJSGlFKUaBVN6ANoFkdAnLrLk8zQ/3V9lChoBmgJaA9DCHxHjQkhgqZAlIaUUpRoFU3oA2gWR0Cc0ikVvddndX2UKGgGaAloD0MI7dYyGeYXpkCUhpRSlGgVTegDaBZHQJzWxO45Lh91fZQoaAZoCWgPQwitiQW+CkylQJSGlFKUaBVN6ANoFkdAnN1DJMg2ZXV9lChoBmgJaA9DCAeZZOQ0l6NAlIaUUpRoFU3oA2gWR0Cc4Z2PT5O8dX2UKGgGaAloD0MI+pgPCMzFpUCUhpRSlGgVTegDaBZHQJznxNfw7T51fZQoaAZoCWgPQwjIef8fT+CjQJSGlFKUaBVN6ANoFkdAnOxrVWjoIXV9lChoBmgJaA9DCLCMDd3UL6ZAlIaUUpRoFU3oA2gWR0Cc8sgLJCBxdX2UKGgGaAloD0MINxjqsKIepkCUhpRSlGgVTegDaBZHQJz3cNXo1UF1fZQoaAZoCWgPQwh3hNOCrwakQJSGlFKUaBVN6ANoFkdAnP3ZEDyOJnV9lChoBmgJaA9DCKxT5Xv+b6ZAlIaUUpRoFU3oA2gWR0CdAkRF7UobdX2UKGgGaAloD0MINPJ5xYOKpkCUhpRSlGgVTegDaBZHQJ0IntQbdad1fZQoaAZoCWgPQwgsKuJ0OmGlQJSGlFKUaBVN6ANoFkdAnQ0U2LpA2XV9lChoBmgJaA9DCEsC1NQ6R6ZAlIaUUpRoFU3oA2gWR0CdE4ieNDMNdX2UKGgGaAloD0MIRpbMsRS+pUCUhpRSlGgVTegDaBZHQJ0YDM3ZPEd1fZQoaAZoCWgPQwimK9hG3K2kQJSGlFKUaBVN6ANoFkdAnRx/GEPDpHV9lChoBmgJaA9DCEeQSrGjV6RAlIaUUpRoFU3oA2gWR0CdIvF0PpY+dX2UKGgGaAloD0MIbqXXZvt+p0CUhpRSlGgVTegDaBZHQJ0n5QhwEQp1fZQoaAZoCWgPQwiwdakRcrekQJSGlFKUaBVN6ANoFkdAnS469GqgiHV9lChoBmgJaA9DCPGD86kzmKdAlIaUUpRoFU3oA2gWR0CdMvQLNOdodX2UKGgGaAloD0MId9fZkK/LpkCUhpRSlGgVTegDaBZHQJ05Xnied091fZQoaAZoCWgPQwhTA83n9GSlQJSGlFKUaBVN6ANoFkdAnT3V7dBSk3V9lChoBmgJaA9DCMmrcwxw7aNAlIaUUpRoFU3oA2gWR0CdRJLLZBcBdX2UKGgGaAloD0MIUtZvJoZipECUhpRSlGgVTegDaBZHQJ1I7vsqril1fZQoaAZoCWgPQwiTpkHRbHulQJSGlFKUaBVN6ANoFkdAnVBGUOd5IHV9lChoBmgJaA9DCOUNMPPdZKVAlIaUUpRoFU3oA2gWR0CdVNNOuaF3dX2UKGgGaAloD0MI7s1vmBhcpECUhpRSlGgVTegDaBZHQJ1s8UoKD011fZQoaAZoCWgPQwg9DRgkDTikQJSGlFKUaBVN6ANoFkdAnXFdrj5sTHV9lChoBmgJaA9DCFOzB1oJLqRAlIaUUpRoFU3oA2gWR0Cdd8nr6ciGdX2UKGgGaAloD0MIZysv+V/fpECUhpRSlGgVTegDaBZHQJ18KFwkxAV1fZQoaAZoCWgPQwj7BiY3ikKkQJSGlFKUaBVN6ANoFkdAnYLZxeb/fnV9lChoBmgJaA9DCIf6Xdi6BKRAlIaUUpRoFU3oA2gWR0CdhznDziCKdX2UKGgGaAloD0MIDAIrhwaTo0CUhpRSlGgVTegDaBZHQJ2Nmb7TDwZ1fZQoaAZoCWgPQwgZyol2veumQJSGlFKUaBVN6ANoFkdAnZKGsmv4d3V9lChoBmgJaA9DCM5Q3PFWd6VAlIaUUpRoFU3oA2gWR0CdmNUc4o7WdX2UKGgGaAloD0MIrkUL0K5mpkCUhpRSlGgVTegDaBZHQJ2dPd2xIJ91fZQoaAZoCWgPQwhRSghWNfSlQJSGlFKUaBVN6ANoFkdAnaONK28Zk3V9lChoBmgJaA9DCN3temlCcKRAlIaUUpRoFU3oA2gWR0CdqI/20zCUdX2UKGgGaAloD0MIptO6DbqOokCUhpRSlGgVTegDaBZHQJ2vJiTdLxt1fZQoaAZoCWgPQwjqeMxAvY+kQJSGlFKUaBVN6ANoFkdAnbOYRIz3y3V9lChoBmgJaA9DCEEo7+PIV6VAlIaUUpRoFU3oA2gWR0CdufnGbTc7dX2UKGgGaAloD0MI4L4OnPP8pECUhpRSlGgVTegDaBZHQJ2+gQOFxn51fZQoaAZoCWgPQwgUsYhhRwKkQJSGlFKUaBVN6ANoFkdAncSqoESuhnV9lChoBmgJaA9DCDCBW3eLI6VAlIaUUpRoFU3oA2gWR0CdyQ7lJYkndX2UKGgGaAloD0MI53Pudj0bpECUhpRSlGgVTegDaBZHQJ3Pavq1PWR1fZQoaAZoCWgPQwgRj8TLq5ekQJSGlFKUaBVN6ANoFkdAndPRr8BMjHV9lChoBmgJaA9DCKPLm8M9w6RAlIaUUpRoFU3oA2gWR0Cd2jzvqkdndX2UKGgGaAloD0MI2nQEcItjpECUhpRSlGgVTegDaBZHQJ3eqJ3xFy91fZQoaAZoCWgPQwijO4idgeSmQJSGlFKUaBVN6ANoFkdAneUZ+MIeHXV9lChoBmgJaA9DCKEuUihL2aRAlIaUUpRoFU3oA2gWR0Cd6Yq1w5vMdX2UKGgGaAloD0MI98391VvjpECUhpRSlGgVTegDaBZHQJ3vzUb1h9d1fZQoaAZoCWgPQwjeWibDeUGoQJSGlFKUaBVN6ANoFkdAngVRVZLZjHV9lChoBmgJaA9DCDOny2J6x6RAlIaUUpRoFU3oA2gWR0CeC2/7zkIYdX2UKGgGaAloD0MIfHvXoB8Dp0CUhpRSlGgVTegDaBZHQJ4PgETxoZh1fZQoaAZoCWgPQwhmTSzw9Y+lQJSGlFKUaBVN6ANoFkdAnhWOYx+KCXV9lChoBmgJaA9DCI4j1uLTxaNAlIaUUpRoFU3oA2gWR0CeGdqM3qA0dX2UKGgGaAloD0MI/wdYq24ipkCUhpRSlGgVTegDaBZHQJ4gD2OAAhl1fZQoaAZoCWgPQwhNvW4RQNGkQJSGlFKUaBVN6ANoFkdAniS9xMnJDHV9lChoBmgJaA9DCBAiGXKc4aVAlIaUUpRoFU3oA2gWR0CeKQGipNsWdX2UKGgGaAloD0MI4DDRII0XpECUhpRSlGgVTegDaBZHQJ4vaKMvRJF1fZQoaAZoCWgPQwh8REyJlA6lQJSGlFKUaBVN6ANoFkdAnjOlRgqmTHV9lChoBmgJaA9DCJmAXyMpSqRAlIaUUpRoFU3oA2gWR0CeOieLNwBHdX2UKGgGaAloD0MITRWMStp3o0CUhpRSlGgVTegDaBZHQJ4+sDfWMCN1fZQoaAZoCWgPQwhSYWwhoH6kQJSGlFKUaBVN6ANoFkdAnkWXerMkhXV9lChoBmgJaA9DCJPH0/KjdqZAlIaUUpRoFU3oA2gWR0CeSiNDMNc4dX2UKGgGaAloD0MIWafK95Smo0CUhpRSlGgVTegDaBZHQJ5QUw22oeh1fZQoaAZoCWgPQwgRVmMJ+2ulQJSGlFKUaBVN6ANoFkdAnlS8C9ytFXV9lChoBmgJaA9DCDfEeM0rnaRAlIaUUpRoFU3oA2gWR0CeWvMNMGordX2UKGgGaAloD0MIWcAEbi2bpECUhpRSlGgVTegDaBZHQJ5fT8CPp6h1fZQoaAZoCWgPQwg25+CZsCSmQJSGlFKUaBVN6ANoFkdAnmWZo9LYgHV9lChoBmgJaA9DCAiqRq8WX6dAlIaUUpRoFU3oA2gWR0Ceae46fapQdX2UKGgGaAloD0MIYfvJGC9bo0CUhpRSlGgVTegDaBZHQJ5wKOT7l7t1fZQoaAZoCWgPQwi0Imqih5qlQJSGlFKUaBVN6ANoFkdAnnSBDb8FZHV9lChoBmgJaA9DCKFMo8m9/aRAlIaUUpRoFU3oA2gWR0CeesQfIS13dX2UKGgGaAloD0MI8Il1qgSRo0CUhpRSlGgVTegDaBZHQJ5+9g/keZJ1fZQoaAZoCWgPQwhQVaGBOLCkQJSGlFKUaBVN6ANoFkdAnoUx6F/QSnV9lChoBmgJaA9DCDo8hPEzJqRAlIaUUpRoFU3oA2gWR0CempSv1UVBdWUu"
|
| 98 |
},
|
| 99 |
"ep_success_buffer": {
|
| 100 |
":type:": "<class 'collections.deque'>",
|
|
|
|
| 111 |
"n_epochs": 20,
|
| 112 |
"clip_range": {
|
| 113 |
":type:": "<class 'function'>",
|
| 114 |
+
":serialized:": "gAWVhQIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwOX21ha2VfZnVuY3Rpb26Uk5QoaACMDV9idWlsdGluX3R5cGWUk5SMCENvZGVUeXBllIWUUpQoSwFLAEsASwFLAUsTQwSIAFMAlE6FlCmMAV+UhZSMUy9ob21lL21heGltaWxpYW4vdmVudi9saWIvcHl0aG9uMy44L3NpdGUtcGFja2FnZXMvc3RhYmxlX2Jhc2VsaW5lczMvY29tbW9uL3V0aWxzLnB5lIwEZnVuY5RLgEMCAAGUjAN2YWyUhZQpdJRSlH2UKIwLX19wYWNrYWdlX1+UjBhzdGFibGVfYmFzZWxpbmVzMy5jb21tb26UjAhfX25hbWVfX5SMHnN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi51dGlsc5SMCF9fZmlsZV9flGgMdU5OaACMEF9tYWtlX2VtcHR5X2NlbGyUk5QpUpSFlHSUUpSMHGNsb3VkcGlja2xlLmNsb3VkcGlja2xlX2Zhc3SUjBJfZnVuY3Rpb25fc2V0c3RhdGWUk5RoHn2UfZQoaBZoDYwMX19xdWFsbmFtZV9flIwZY29uc3RhbnRfZm4uPGxvY2Fscz4uZnVuY5SMD19fYW5ub3RhdGlvbnNfX5R9lIwOX19rd2RlZmF1bHRzX1+UTowMX19kZWZhdWx0c19flE6MCl9fbW9kdWxlX1+UaBeMB19fZG9jX1+UTowLX19jbG9zdXJlX1+UaACMCl9tYWtlX2NlbGyUk5RHP7mZmZmZmZqFlFKUhZSMF19jbG91ZHBpY2tsZV9zdWJtb2R1bGVzlF2UjAtfX2dsb2JhbHNfX5R9lHWGlIZSMC4="
|
| 115 |
},
|
| 116 |
"clip_range_vf": null,
|
| 117 |
"normalize_advantage": true,
|
ppo-seals-Hopper-v0/policy.optimizer.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5a2dc0cc6c01c40a8988c5d504579a880768118797e6c3ee8ad84dc793d84c4d
|
| 3 |
+
size 91888
|
ppo-seals-Hopper-v0/policy.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:261c63aa4b0af15f7cb4479087188f837eb673d5d7181e2ff01132de1a634554
|
| 3 |
+
size 46197
|
ppo-seals-Hopper-v0/system_info.txt
CHANGED
|
@@ -1,6 +1,6 @@
|
|
| 1 |
-
OS: Linux-5.4.0-
|
| 2 |
Python: 3.8.10
|
| 3 |
-
Stable-Baselines3: 1.6.
|
| 4 |
PyTorch: 1.11.0+cu102
|
| 5 |
GPU Enabled: False
|
| 6 |
Numpy: 1.22.3
|
|
|
|
| 1 |
+
OS: Linux-5.4.0-125-generic-x86_64-with-glibc2.29 #141-Ubuntu SMP Wed Aug 10 13:42:03 UTC 2022
|
| 2 |
Python: 3.8.10
|
| 3 |
+
Stable-Baselines3: 1.6.2
|
| 4 |
PyTorch: 1.11.0+cu102
|
| 5 |
GPU Enabled: False
|
| 6 |
Numpy: 1.22.3
|
replay.mp4
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b22dd39a8e88d2fbe0fff4d6937ebe78f47db1e68d5e4070eb6958c22d5d4616
|
| 3 |
+
size 1618187
|
results.json
CHANGED
|
@@ -1 +1 @@
|
|
| 1 |
-
{"mean_reward":
|
|
|
|
| 1 |
+
{"mean_reward": 2862.6938039999995, "std_reward": 127.19838412290773, "is_deterministic": true, "n_eval_episodes": 10, "eval_datetime": "2022-12-31T18:49:30.926136"}
|
train_eval_metrics.zip
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:24ac15be2dd841cfd73cc316ae9ca3719311585733344f1b7bbf58399fc6a23d
|
| 3 |
+
size 33790
|
vec_normalize.pkl
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:cd2b2cc76deac20e7252c9b52f1bf0ef9e08658624d159027a1d8015d03a3ae2
|
| 3 |
+
size 4187
|