Initial commit
Browse files- README.md +19 -7
- args.yml +11 -5
- config.yml +12 -2
- ppo-seals-Swimmer-v0.zip +2 -2
- ppo-seals-Swimmer-v0/_stable_baselines3_version +1 -1
- ppo-seals-Swimmer-v0/data +24 -23
- ppo-seals-Swimmer-v0/policy.optimizer.pth +2 -2
- ppo-seals-Swimmer-v0/policy.pth +2 -2
- ppo-seals-Swimmer-v0/system_info.txt +2 -2
- replay.mp4 +2 -2
- results.json +1 -1
- train_eval_metrics.zip +2 -2
- vec_normalize.pkl +3 -0
README.md
CHANGED
|
@@ -10,7 +10,7 @@ model-index:
|
|
| 10 |
results:
|
| 11 |
- metrics:
|
| 12 |
- type: mean_reward
|
| 13 |
-
value:
|
| 14 |
name: mean_reward
|
| 15 |
task:
|
| 16 |
type: reinforcement-learning
|
|
@@ -37,15 +37,21 @@ SB3 Contrib: https://github.com/Stable-Baselines-Team/stable-baselines3-contrib
|
|
| 37 |
|
| 38 |
```
|
| 39 |
# Download model and save it into the logs/ folder
|
| 40 |
-
python -m
|
| 41 |
python enjoy.py --algo ppo --env seals/Swimmer-v0 -f logs/
|
| 42 |
```
|
| 43 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 44 |
## Training (with the RL Zoo)
|
| 45 |
```
|
| 46 |
python train.py --algo ppo --env seals/Swimmer-v0 -f logs/
|
| 47 |
# Upload the model and generate video (when possible)
|
| 48 |
-
python -m
|
| 49 |
```
|
| 50 |
|
| 51 |
## Hyperparameters
|
|
@@ -60,11 +66,17 @@ OrderedDict([('batch_size', 8),
|
|
| 60 |
('n_epochs', 20),
|
| 61 |
('n_steps', 2048),
|
| 62 |
('n_timesteps', 1000000.0),
|
| 63 |
-
('normalize',
|
|
|
|
| 64 |
('policy', 'MlpPolicy'),
|
| 65 |
('policy_kwargs',
|
| 66 |
-
'
|
| 67 |
-
|
|
|
|
| 68 |
('vf_coef', 0.6162112311062333),
|
| 69 |
-
('normalize_kwargs',
|
|
|
|
|
|
|
|
|
|
|
|
|
| 70 |
```
|
|
|
|
| 10 |
results:
|
| 11 |
- metrics:
|
| 12 |
- type: mean_reward
|
| 13 |
+
value: 347.70 +/- 5.88
|
| 14 |
name: mean_reward
|
| 15 |
task:
|
| 16 |
type: reinforcement-learning
|
|
|
|
| 37 |
|
| 38 |
```
|
| 39 |
# Download model and save it into the logs/ folder
|
| 40 |
+
python -m rl_zoo3.load_from_hub --algo ppo --env seals/Swimmer-v0 -orga HumanCompatibleAI -f logs/
|
| 41 |
python enjoy.py --algo ppo --env seals/Swimmer-v0 -f logs/
|
| 42 |
```
|
| 43 |
|
| 44 |
+
If you installed the RL Zoo3 via pip (`pip install rl_zoo3`), from anywhere you can do:
|
| 45 |
+
```
|
| 46 |
+
python -m rl_zoo3.load_from_hub --algo ppo --env seals/Swimmer-v0 -orga HumanCompatibleAI -f logs/
|
| 47 |
+
rl_zoo3 enjoy --algo ppo --env seals/Swimmer-v0 -f logs/
|
| 48 |
+
```
|
| 49 |
+
|
| 50 |
## Training (with the RL Zoo)
|
| 51 |
```
|
| 52 |
python train.py --algo ppo --env seals/Swimmer-v0 -f logs/
|
| 53 |
# Upload the model and generate video (when possible)
|
| 54 |
+
python -m rl_zoo3.push_to_hub --algo ppo --env seals/Swimmer-v0 -f logs/ -orga HumanCompatibleAI
|
| 55 |
```
|
| 56 |
|
| 57 |
## Hyperparameters
|
|
|
|
| 66 |
('n_epochs', 20),
|
| 67 |
('n_steps', 2048),
|
| 68 |
('n_timesteps', 1000000.0),
|
| 69 |
+
('normalize',
|
| 70 |
+
{'gamma': 0.999, 'norm_obs': False, 'norm_reward': True}),
|
| 71 |
('policy', 'MlpPolicy'),
|
| 72 |
('policy_kwargs',
|
| 73 |
+
{'activation_fn': <class 'torch.nn.modules.activation.Tanh'>,
|
| 74 |
+
'features_extractor_class': <class 'imitation.policies.base.NormalizeFeaturesExtractor'>,
|
| 75 |
+
'net_arch': [{'pi': [64, 64], 'vf': [64, 64]}]}),
|
| 76 |
('vf_coef', 0.6162112311062333),
|
| 77 |
+
('normalize_kwargs',
|
| 78 |
+
{'norm_obs': {'gamma': 0.999,
|
| 79 |
+
'norm_obs': False,
|
| 80 |
+
'norm_reward': True},
|
| 81 |
+
'norm_reward': False})])
|
| 82 |
```
|
args.yml
CHANGED
|
@@ -1,6 +1,8 @@
|
|
| 1 |
!!python/object/apply:collections.OrderedDict
|
| 2 |
- - - algo
|
| 3 |
- ppo
|
|
|
|
|
|
|
| 4 |
- - device
|
| 5 |
- cpu
|
| 6 |
- - env
|
|
@@ -16,7 +18,7 @@
|
|
| 16 |
- - hyperparams
|
| 17 |
- null
|
| 18 |
- - log_folder
|
| 19 |
-
-
|
| 20 |
- - log_interval
|
| 21 |
- -1
|
| 22 |
- - max_total_trials
|
|
@@ -41,6 +43,8 @@
|
|
| 41 |
- null
|
| 42 |
- - optimize_hyperparameters
|
| 43 |
- false
|
|
|
|
|
|
|
| 44 |
- - pruner
|
| 45 |
- median
|
| 46 |
- - sampler
|
|
@@ -50,13 +54,13 @@
|
|
| 50 |
- - save_replay_buffer
|
| 51 |
- false
|
| 52 |
- - seed
|
| 53 |
-
-
|
| 54 |
- - storage
|
| 55 |
- null
|
| 56 |
- - study_name
|
| 57 |
- null
|
| 58 |
- - tensorboard_log
|
| 59 |
-
- runs/seals/Swimmer-
|
| 60 |
- - track
|
| 61 |
- true
|
| 62 |
- - trained_agent
|
|
@@ -70,6 +74,8 @@
|
|
| 70 |
- - verbose
|
| 71 |
- 1
|
| 72 |
- - wandb_entity
|
| 73 |
-
-
|
| 74 |
- - wandb_project_name
|
| 75 |
-
- seals-experts-
|
|
|
|
|
|
|
|
|
| 1 |
!!python/object/apply:collections.OrderedDict
|
| 2 |
- - - algo
|
| 3 |
- ppo
|
| 4 |
+
- - conf_file
|
| 5 |
+
- hyperparams/python/ppo.py
|
| 6 |
- - device
|
| 7 |
- cpu
|
| 8 |
- - env
|
|
|
|
| 18 |
- - hyperparams
|
| 19 |
- null
|
| 20 |
- - log_folder
|
| 21 |
+
- logs
|
| 22 |
- - log_interval
|
| 23 |
- -1
|
| 24 |
- - max_total_trials
|
|
|
|
| 43 |
- null
|
| 44 |
- - optimize_hyperparameters
|
| 45 |
- false
|
| 46 |
+
- - progress
|
| 47 |
+
- false
|
| 48 |
- - pruner
|
| 49 |
- median
|
| 50 |
- - sampler
|
|
|
|
| 54 |
- - save_replay_buffer
|
| 55 |
- false
|
| 56 |
- - seed
|
| 57 |
+
- 6
|
| 58 |
- - storage
|
| 59 |
- null
|
| 60 |
- - study_name
|
| 61 |
- null
|
| 62 |
- - tensorboard_log
|
| 63 |
+
- runs/seals/Swimmer-v0__ppo__6__1670518602
|
| 64 |
- - track
|
| 65 |
- true
|
| 66 |
- - trained_agent
|
|
|
|
| 74 |
- - verbose
|
| 75 |
- 1
|
| 76 |
- - wandb_entity
|
| 77 |
+
- ernestum
|
| 78 |
- - wandb_project_name
|
| 79 |
+
- seals-experts-normalized
|
| 80 |
+
- - yaml_file
|
| 81 |
+
- null
|
config.yml
CHANGED
|
@@ -20,10 +20,20 @@
|
|
| 20 |
- - n_timesteps
|
| 21 |
- 1000000.0
|
| 22 |
- - normalize
|
| 23 |
-
-
|
|
|
|
|
|
|
| 24 |
- - policy
|
| 25 |
- MlpPolicy
|
| 26 |
- - policy_kwargs
|
| 27 |
-
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 28 |
- - vf_coef
|
| 29 |
- 0.6162112311062333
|
|
|
|
| 20 |
- - n_timesteps
|
| 21 |
- 1000000.0
|
| 22 |
- - normalize
|
| 23 |
+
- gamma: 0.999
|
| 24 |
+
norm_obs: false
|
| 25 |
+
norm_reward: true
|
| 26 |
- - policy
|
| 27 |
- MlpPolicy
|
| 28 |
- - policy_kwargs
|
| 29 |
+
- activation_fn: !!python/name:torch.nn.modules.activation.Tanh ''
|
| 30 |
+
features_extractor_class: !!python/name:imitation.policies.base.NormalizeFeaturesExtractor ''
|
| 31 |
+
net_arch:
|
| 32 |
+
- pi:
|
| 33 |
+
- 64
|
| 34 |
+
- 64
|
| 35 |
+
vf:
|
| 36 |
+
- 64
|
| 37 |
+
- 64
|
| 38 |
- - vf_coef
|
| 39 |
- 0.6162112311062333
|
ppo-seals-Swimmer-v0.zip
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:bd472334868d047da1f2dbe26bc47cbfd9374248978285b1045a865e4a775a65
|
| 3 |
+
size 155593
|
ppo-seals-Swimmer-v0/_stable_baselines3_version
CHANGED
|
@@ -1 +1 @@
|
|
| 1 |
-
1.6.
|
|
|
|
| 1 |
+
1.6.2
|
ppo-seals-Swimmer-v0/data
CHANGED
|
@@ -4,24 +4,24 @@
|
|
| 4 |
":serialized:": "gAWVOwAAAAAAAACMIXN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi5wb2xpY2llc5SMEUFjdG9yQ3JpdGljUG9saWN5lJOULg==",
|
| 5 |
"__module__": "stable_baselines3.common.policies",
|
| 6 |
"__doc__": "\n Policy class for actor-critic algorithms (has both policy and value prediction).\n Used by A2C, PPO and the likes.\n\n :param observation_space: Observation space\n :param action_space: Action space\n :param lr_schedule: Learning rate schedule (could be constant)\n :param net_arch: The specification of the policy and value networks.\n :param activation_fn: Activation function\n :param ortho_init: Whether to use or not orthogonal initialization\n :param use_sde: Whether to use State Dependent Exploration or not\n :param log_std_init: Initial value for the log standard deviation\n :param full_std: Whether to use (n_features x n_actions) parameters\n for the std instead of only (n_features,) when using gSDE\n :param sde_net_arch: Network architecture for extracting features\n when using gSDE. If None, the latent features from the policy will be used.\n Pass an empty list to use the states as features.\n :param use_expln: Use ``expln()`` function instead of ``exp()`` to ensure\n a positive standard deviation (cf paper). It allows to keep variance\n above zero and prevent it from growing too fast. In practice, ``exp()`` is usually enough.\n :param squash_output: Whether to squash the output using a tanh function,\n this allows to ensure boundaries when using gSDE.\n :param features_extractor_class: Features extractor to use.\n :param features_extractor_kwargs: Keyword arguments\n to pass to the features extractor.\n :param normalize_images: Whether to normalize images or not,\n dividing by 255.0 (True by default)\n :param optimizer_class: The optimizer to use,\n ``th.optim.Adam`` by default\n :param optimizer_kwargs: Additional keyword arguments,\n excluding the learning rate, to pass to the optimizer\n ",
|
| 7 |
-
"__init__": "<function ActorCriticPolicy.__init__ at
|
| 8 |
-
"_get_constructor_parameters": "<function ActorCriticPolicy._get_constructor_parameters at
|
| 9 |
-
"reset_noise": "<function ActorCriticPolicy.reset_noise at
|
| 10 |
-
"_build_mlp_extractor": "<function ActorCriticPolicy._build_mlp_extractor at
|
| 11 |
-
"_build": "<function ActorCriticPolicy._build at
|
| 12 |
-
"forward": "<function ActorCriticPolicy.forward at
|
| 13 |
-
"_get_action_dist_from_latent": "<function ActorCriticPolicy._get_action_dist_from_latent at
|
| 14 |
-
"_predict": "<function ActorCriticPolicy._predict at
|
| 15 |
-
"evaluate_actions": "<function ActorCriticPolicy.evaluate_actions at
|
| 16 |
-
"get_distribution": "<function ActorCriticPolicy.get_distribution at
|
| 17 |
-
"predict_values": "<function ActorCriticPolicy.predict_values at
|
| 18 |
"__abstractmethods__": "frozenset()",
|
| 19 |
-
"_abc_impl": "<_abc_data object at
|
| 20 |
},
|
| 21 |
"verbose": 1,
|
| 22 |
"policy_kwargs": {
|
| 23 |
":type:": "<class 'dict'>",
|
| 24 |
-
":serialized:": "
|
| 25 |
"activation_fn": "<class 'torch.nn.modules.activation.Tanh'>",
|
| 26 |
"net_arch": [
|
| 27 |
{
|
|
@@ -34,7 +34,8 @@
|
|
| 34 |
64
|
| 35 |
]
|
| 36 |
}
|
| 37 |
-
]
|
|
|
|
| 38 |
},
|
| 39 |
"observation_space": {
|
| 40 |
":type:": "<class 'gym.spaces.box.Box'>",
|
|
@@ -51,7 +52,7 @@
|
|
| 51 |
},
|
| 52 |
"action_space": {
|
| 53 |
":type:": "<class 'gym.spaces.box.Box'>",
|
| 54 |
-
":serialized:": "gAWV6wsAAAAAAACMDmd5bS5zcGFjZXMuYm94lIwDQm94lJOUKYGUfZQojAVkdHlwZZSMBW51bXB5lGgFk5SMAmY0lImIh5RSlChLA4wBPJROTk5K/////0r/////SwB0lGKMBl9zaGFwZZRLAoWUjANsb3eUjBJudW1weS5jb3JlLm51bWVyaWOUjAtfZnJvbWJ1ZmZlcpSTlCiWCAAAAAAAAAAAAIC/AACAv5RoCksChZSMAUOUdJRSlIwEaGlnaJRoEiiWCAAAAAAAAAAAAIA/AACAP5RoCksChZRoFXSUUpSMDWJvdW5kZWRfYmVsb3eUaBIolgIAAAAAAAAAAQGUaAeMAmIxlImIh5RSlChLA4wBfJROTk5K/////0r/////
|
| 55 |
"dtype": "float32",
|
| 56 |
"_shape": [
|
| 57 |
2
|
|
@@ -66,17 +67,17 @@
|
|
| 66 |
"num_timesteps": 1001472,
|
| 67 |
"_total_timesteps": 1000000,
|
| 68 |
"_num_timesteps_at_start": 0,
|
| 69 |
-
"seed":
|
| 70 |
"action_noise": null,
|
| 71 |
-
"start_time":
|
| 72 |
"learning_rate": {
|
| 73 |
":type:": "<class 'function'>",
|
| 74 |
-
":serialized:": "
|
| 75 |
},
|
| 76 |
-
"tensorboard_log": "runs/seals/Swimmer-
|
| 77 |
"lr_schedule": {
|
| 78 |
":type:": "<class 'function'>",
|
| 79 |
-
":serialized:": "
|
| 80 |
},
|
| 81 |
"_last_obs": null,
|
| 82 |
"_last_episode_starts": {
|
|
@@ -85,7 +86,7 @@
|
|
| 85 |
},
|
| 86 |
"_last_original_obs": {
|
| 87 |
":type:": "<class 'numpy.ndarray'>",
|
| 88 |
-
":serialized:": "
|
| 89 |
},
|
| 90 |
"_episode_num": 0,
|
| 91 |
"use_sde": false,
|
|
@@ -93,7 +94,7 @@
|
|
| 93 |
"_current_progress_remaining": -0.0014719999999999178,
|
| 94 |
"ep_info_buffer": {
|
| 95 |
":type:": "<class 'collections.deque'>",
|
| 96 |
-
":serialized:": "gAWVgRAAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKUKH2UKIwBcpSMFW51bXB5LmNvcmUubXVsdGlhcnJheZSMBnNjYWxhcpSTlIwFbnVtcHmUjAVkdHlwZZSTlIwCZjiUiYiHlFKUKEsDjAE8lE5OTkr/////Sv////
|
| 97 |
},
|
| 98 |
"ep_success_buffer": {
|
| 99 |
":type:": "<class 'collections.deque'>",
|
|
@@ -110,7 +111,7 @@
|
|
| 110 |
"n_epochs": 20,
|
| 111 |
"clip_range": {
|
| 112 |
":type:": "<class 'function'>",
|
| 113 |
-
":serialized:": "
|
| 114 |
},
|
| 115 |
"clip_range_vf": null,
|
| 116 |
"normalize_advantage": true,
|
|
|
|
| 4 |
":serialized:": "gAWVOwAAAAAAAACMIXN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi5wb2xpY2llc5SMEUFjdG9yQ3JpdGljUG9saWN5lJOULg==",
|
| 5 |
"__module__": "stable_baselines3.common.policies",
|
| 6 |
"__doc__": "\n Policy class for actor-critic algorithms (has both policy and value prediction).\n Used by A2C, PPO and the likes.\n\n :param observation_space: Observation space\n :param action_space: Action space\n :param lr_schedule: Learning rate schedule (could be constant)\n :param net_arch: The specification of the policy and value networks.\n :param activation_fn: Activation function\n :param ortho_init: Whether to use or not orthogonal initialization\n :param use_sde: Whether to use State Dependent Exploration or not\n :param log_std_init: Initial value for the log standard deviation\n :param full_std: Whether to use (n_features x n_actions) parameters\n for the std instead of only (n_features,) when using gSDE\n :param sde_net_arch: Network architecture for extracting features\n when using gSDE. If None, the latent features from the policy will be used.\n Pass an empty list to use the states as features.\n :param use_expln: Use ``expln()`` function instead of ``exp()`` to ensure\n a positive standard deviation (cf paper). It allows to keep variance\n above zero and prevent it from growing too fast. In practice, ``exp()`` is usually enough.\n :param squash_output: Whether to squash the output using a tanh function,\n this allows to ensure boundaries when using gSDE.\n :param features_extractor_class: Features extractor to use.\n :param features_extractor_kwargs: Keyword arguments\n to pass to the features extractor.\n :param normalize_images: Whether to normalize images or not,\n dividing by 255.0 (True by default)\n :param optimizer_class: The optimizer to use,\n ``th.optim.Adam`` by default\n :param optimizer_kwargs: Additional keyword arguments,\n excluding the learning rate, to pass to the optimizer\n ",
|
| 7 |
+
"__init__": "<function ActorCriticPolicy.__init__ at 0x7fded74c2790>",
|
| 8 |
+
"_get_constructor_parameters": "<function ActorCriticPolicy._get_constructor_parameters at 0x7fded74c2820>",
|
| 9 |
+
"reset_noise": "<function ActorCriticPolicy.reset_noise at 0x7fded74c28b0>",
|
| 10 |
+
"_build_mlp_extractor": "<function ActorCriticPolicy._build_mlp_extractor at 0x7fded74c2940>",
|
| 11 |
+
"_build": "<function ActorCriticPolicy._build at 0x7fded74c29d0>",
|
| 12 |
+
"forward": "<function ActorCriticPolicy.forward at 0x7fded74c2a60>",
|
| 13 |
+
"_get_action_dist_from_latent": "<function ActorCriticPolicy._get_action_dist_from_latent at 0x7fded74c2af0>",
|
| 14 |
+
"_predict": "<function ActorCriticPolicy._predict at 0x7fded74c2b80>",
|
| 15 |
+
"evaluate_actions": "<function ActorCriticPolicy.evaluate_actions at 0x7fded74c2c10>",
|
| 16 |
+
"get_distribution": "<function ActorCriticPolicy.get_distribution at 0x7fded74c2ca0>",
|
| 17 |
+
"predict_values": "<function ActorCriticPolicy.predict_values at 0x7fded74c2d30>",
|
| 18 |
"__abstractmethods__": "frozenset()",
|
| 19 |
+
"_abc_impl": "<_abc_data object at 0x7fded74b9c00>"
|
| 20 |
},
|
| 21 |
"verbose": 1,
|
| 22 |
"policy_kwargs": {
|
| 23 |
":type:": "<class 'dict'>",
|
| 24 |
+
":serialized:": "gAWVvAAAAAAAAAB9lCiMDWFjdGl2YXRpb25fZm6UjBt0b3JjaC5ubi5tb2R1bGVzLmFjdGl2YXRpb26UjARUYW5olJOUjAhuZXRfYXJjaJRdlH2UKIwCcGmUXZQoS0BLQGWMAnZmlF2UKEtAS0BldWGMGGZlYXR1cmVzX2V4dHJhY3Rvcl9jbGFzc5SMF2ltaXRhdGlvbi5wb2xpY2llcy5iYXNllIwaTm9ybWFsaXplRmVhdHVyZXNFeHRyYWN0b3KUk5R1Lg==",
|
| 25 |
"activation_fn": "<class 'torch.nn.modules.activation.Tanh'>",
|
| 26 |
"net_arch": [
|
| 27 |
{
|
|
|
|
| 34 |
64
|
| 35 |
]
|
| 36 |
}
|
| 37 |
+
],
|
| 38 |
+
"features_extractor_class": "<class 'imitation.policies.base.NormalizeFeaturesExtractor'>"
|
| 39 |
},
|
| 40 |
"observation_space": {
|
| 41 |
":type:": "<class 'gym.spaces.box.Box'>",
|
|
|
|
| 52 |
},
|
| 53 |
"action_space": {
|
| 54 |
":type:": "<class 'gym.spaces.box.Box'>",
|
| 55 |
+
":serialized:": "gAWV6wsAAAAAAACMDmd5bS5zcGFjZXMuYm94lIwDQm94lJOUKYGUfZQojAVkdHlwZZSMBW51bXB5lGgFk5SMAmY0lImIh5RSlChLA4wBPJROTk5K/////0r/////SwB0lGKMBl9zaGFwZZRLAoWUjANsb3eUjBJudW1weS5jb3JlLm51bWVyaWOUjAtfZnJvbWJ1ZmZlcpSTlCiWCAAAAAAAAAAAAIC/AACAv5RoCksChZSMAUOUdJRSlIwEaGlnaJRoEiiWCAAAAAAAAAAAAIA/AACAP5RoCksChZRoFXSUUpSMDWJvdW5kZWRfYmVsb3eUaBIolgIAAAAAAAAAAQGUaAeMAmIxlImIh5RSlChLA4wBfJROTk5K/////0r/////SwB0lGJLAoWUaBV0lFKUjA1ib3VuZGVkX2Fib3ZllGgSKJYCAAAAAAAAAAEBlGghSwKFlGgVdJRSlIwKX25wX3JhbmRvbZSMFG51bXB5LnJhbmRvbS5fcGlja2xllIwSX19yYW5kb21zdGF0ZV9jdG9ylJOUjAdNVDE5OTM3lIWUUpR9lCiMDWJpdF9nZW5lcmF0b3KUaDCMBXN0YXRllH2UKIwDa2V5lGgSKJbACQAAAAAAAAAAAIBdNiykTVDFqeVGNKxaN7aOt6Yq8LfFTH1npoiT22Tn9u9tazWba6wpRCTXkgUGD5cZ1t/kHYHodaF4ctU/OUK+6dkSEBgC5E3H0l+qg/ODVc7nI3EqAKIZHEwZ6NVmt33SsqG7AT1rCfTSFYShKYLGUn8N6SkrWSPSll8n6dtwkk0TcqzdAuUeugUlbzadZy7jf/1TWviIyjfgCS21BHXZWTQS84bfiTJCGZoRNWqioo0rbvd4OFi1yiM9KnPAzdJRsnOTnrbz62yXMW1uBxIv1Cd3bDsJGiCxuDNliGGH6HFhNeGwa6GYoQeOZ7qBJQ235xJ2NLWp1BiBnWPBvVyHKGJzpJBitW9VeFB39FSCaSNIpsWR+zVvLlErvbyMTPN1s22BXQ+CVuTJM+MQxbIEC8sc0+VfsAVxpqNh/oMSfTAQ7gwSTNZqc0h3J+oMXf2mbuhFRVpEOqfgPXX07SzaSZmyYcmLjnbmhQ7Mv+WwsAV2tzwAFWW9sLPdF91iXfOz4j4UPf89aXeSOTdzOO6uflcW66uS/QWHUBWdBGlYUOjWjb243zQh4OMAoSksyk3Z8PnuEp6Rz6oM7WJb6o4kjylk/rl1KTjrizSZoZU9Fbew0++t4GAjyeYELAH01BifJ9/cw5zVg8ehpYT9Rb7J3m/i/qdvsMQ65D4lG7i/Gw4sLMaE6GLXmwuIE1olYzP9g3B4xAuSsf+OoTG1ZnMHPa6i4BSPpx/fKvSpjyPE6i598M17fChYnHkQsIinQh60/Eszb+WD6eOZVe8coR3Vjt17D1CzUQ821eyy0MV2Wbv0jm4BgqrO6FlRseFjUdXismJ1vKVfe2mx1qhDw72fY4MwTlmLY7NyrEMxLksGeLfS4M/DnYB1eZ7UwGWAA7dTBIrbuPXzvD7LNU3epyy1kqbfSw+FCzPmsph4fz1A1DeSg31mC+MX/JvJE9t/g8nA5j4hZpN6jBeKlV28S1p2nL01QKK+r6pKrkmISU38maSHRZuFbq3LJz8t631kPv0jm0ktM7Z+NahewonqfINi4bpVOQBXfRtPbeTHL8wi9cc+b4MlnBctJdGBVXXGgFTRlzdUUDhsCfAAjTyx8wlKb8zAbbGKtsZWgIczYX0ub/Uc7k8jvIyUJ3grrVYOh2cuFTTYxOhStGgrLYnrKgYt1c7TILhOoaHzZ8uX9ihByVYj3AS+wa7tfQGyPPxP5pOIEuB5PS9Qac3GcbdeNOa/C/BV6APf12bSNsh6VqYhSNjCnaeceBvH75TnAtLHw3slcg4WbPXsdrOKxZDEjpHE0XIKu8HN5jQ3U/3Fn3eh7fN0dIlW0mirBvY8HievGcVYJITzxKa4/qeoxkG3gx7Z4iGV1IFC+1uxWSDa04f5l/iZF0lnDF+nXgg8ki6yo39KC3EXd8tUpRu2Hms8StK7iK+HHyV/7df9gCgyYQiPjfjrD25IDrlnm/rx4woAAVItPr7uqx+7k0nvbggUnF9jtfrUEGE045FXveXP6yqJSW7V6blWwzO8BiiCpGRpbU1R5ALokTK4pDxy2BqOsRDduNATa/2Nh14SKs4+j8Tftoy4y+HJ5Pnx+DSdO3N1GtaTpTbMyEzA+ShdMYR8cTtOFpUe4iIMTqS2TUyUOFk+E0BgkSWpXMZJOrWGeimVFGMKvzyVQMPNMdKr1u/lFYMTBZgZR3WSYbH0lqUpcGJdL9GMVOmNXXFASyK0ahRmGNjSeGMhZDGItlf4KFO3E15Wb74k3FYiHEsavHGUy96hbRcAXgUWIx6+wYdd2HYgjkUpbEoHGxZN6e2cxxFYyDPWTbl1WzKll7+G5O6lfUs1Pu/RBLQ/m+j8pJoGsO6hLDlS8l2x3CNhwG7/mDSpHe24SRxbvzaThAvvSBEeS0VpgeOvEAIMuTT5vloPBlnS1eeyQexJPePqxZkkpk+eSnRKyNWWmJaq+HreONrqOcEWhibZtJTOf14H4aI/zkGK423AB9X22JhlG8nbjEhWVLimuLNyPKFI3zz4FYW23m3Iu2hus4jiPFFvvAQuInuC/8q7J+GbbdMXcV5YtKNEiUIUv63EPZak3N1cZ4X/WRe9FQ6tHfE7v4bv613OWHIDyEtgjhSIA/uw5Vco5DD75JMXP7iYA/0vBTgr8nQ5aE7ZRFIveuaMD3nexwciZNE3ytIC6UqCjFWqz3JH0/vxMeOZn+EhJnwhAXIKdzkRoSnjlHHlGIvwWS8lTf83/vFRtVVTRoZtB3GZN1NzjB+rMZDGJHT746Ac5CkzTzn9hmys1O2n6mU/JsnXTmNb7nVH7Asbq1lB8Xw102itjPS4byZxSl2Vy5Z9VEyntgrBM/p1GbDWDcDXeIOhCNqJ5nQ0+5npNf7+tPYNouYarj99hkUrl7llhrkcB4zj2V+5SurhaUmGcs9H7YWF01kOIGp7icCnQUhL+Of2VGhGq2ItqXMbIVpovW6vKoUVmoiny8Sfu48wJXs1TWtt/KmWkUg8HItzlNVWcPUBt5c0nIwHo85GgGWpJIWUlXPs0vEHqlp9KSaD5OsErT9I4aJjNnyhMQ10qHZeR3j6h5fnSUoLUXqp7LW5qB7HsvswC+Av4SCMGyDx3oC0I6Mz/jpazX8LFsEEqOW1sHEYfi34Y7qqwKBFdcNl+YLbbP8JDl27GXYcZdbF9XuSz3PIFuqCGhuwH3APxkKVcWrqtwfG21wbsw8NCkWJjkwJ7AV/Uw7JUovEwZaUS5csL7zS0NseVGuY+pSSySW0Iy7zPjXijqbaJ0X6t6WGIwekscWovjOogFnQX/oWmKAyRVkd9+WherZRQsMpoNcDuS3TyWax8/hBLVUP0hOWdWlvc4+8gDO+LYfk2TkVAt485Mx7nmpRznUqU3joRF39tVvXJOPTwmb9D+EeR/64By1f/D8Sctc7kUr6XJKUxJmR3ULP9jQO9Bnn2tc0oRsP7pHcunIMhA57YPW8CFpKwQkGX9+44bMpg4vU2atO5fGB55WFK2zdvp19m0uEsCaZB/dPj/NXRSbqF5LJkgxq36ZyOxpOcaAeWFaawMrEKqkPLDDOuXSuSXRvccEHyKT+EDmM8dlZ1hMAwgj45j2pGpQXQ7XQ447TmprGjqmeuVB8W6Bke4sYVDo5kPO2c5G4hnxfd65t+f6+C7MtnXT633LCwyvMG8jlGOojzrRglw4Oda09zEfmZ3OQnDscwz0myEd8zeLfYpd+kqtYqcPerDPkk7Y/5zLDrUKWofZqfDgfu8UqLkNU0WfqkAJHLsXyimYZmSbFSUC6igvYYRLH2BRI9MYV0K1NSgKAykH8FxIWj13AEu7abZRoB4wCdTSUiYiHlFKUKEsDaAtOTk5K/////0r/////SwB0lGJNcAKFlGgVdJRSlIwDcG9zlE1wAnWMCWhhc19nYXVzc5RLAIwFZ2F1c3OURwAAAAAAAAAAdWJ1Yi4=",
|
| 56 |
"dtype": "float32",
|
| 57 |
"_shape": [
|
| 58 |
2
|
|
|
|
| 67 |
"num_timesteps": 1001472,
|
| 68 |
"_total_timesteps": 1000000,
|
| 69 |
"_num_timesteps_at_start": 0,
|
| 70 |
+
"seed": 4,
|
| 71 |
"action_noise": null,
|
| 72 |
+
"start_time": 1670518608185674844,
|
| 73 |
"learning_rate": {
|
| 74 |
":type:": "<class 'function'>",
|
| 75 |
+
":serialized:": "gAWVhQIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwOX21ha2VfZnVuY3Rpb26Uk5QoaACMDV9idWlsdGluX3R5cGWUk5SMCENvZGVUeXBllIWUUpQoSwFLAEsASwFLAUsTQwSIAFMAlE6FlCmMAV+UhZSMUy9ob21lL21heGltaWxpYW4vdmVudi9saWIvcHl0aG9uMy44L3NpdGUtcGFja2FnZXMvc3RhYmxlX2Jhc2VsaW5lczMvY29tbW9uL3V0aWxzLnB5lIwEZnVuY5RLgEMCAAGUjAN2YWyUhZQpdJRSlH2UKIwLX19wYWNrYWdlX1+UjBhzdGFibGVfYmFzZWxpbmVzMy5jb21tb26UjAhfX25hbWVfX5SMHnN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi51dGlsc5SMCF9fZmlsZV9flGgMdU5OaACMEF9tYWtlX2VtcHR5X2NlbGyUk5QpUpSFlHSUUpSMHGNsb3VkcGlja2xlLmNsb3VkcGlja2xlX2Zhc3SUjBJfZnVuY3Rpb25fc2V0c3RhdGWUk5RoHn2UfZQoaBZoDYwMX19xdWFsbmFtZV9flIwZY29uc3RhbnRfZm4uPGxvY2Fscz4uZnVuY5SMD19fYW5ub3RhdGlvbnNfX5R9lIwOX19rd2RlZmF1bHRzX1+UTowMX19kZWZhdWx0c19flE6MCl9fbW9kdWxlX1+UaBeMB19fZG9jX1+UTowLX19jbG9zdXJlX1+UaACMCl9tYWtlX2NlbGyUk5RHPx/V8usTiESFlFKUhZSMF19jbG91ZHBpY2tsZV9zdWJtb2R1bGVzlF2UjAtfX2dsb2JhbHNfX5R9lHWGlIZSMC4="
|
| 76 |
},
|
| 77 |
+
"tensorboard_log": "runs/seals/Swimmer-v0__ppo__6__1670518602/seals-Swimmer-v0",
|
| 78 |
"lr_schedule": {
|
| 79 |
":type:": "<class 'function'>",
|
| 80 |
+
":serialized:": "gAWVhQIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwOX21ha2VfZnVuY3Rpb26Uk5QoaACMDV9idWlsdGluX3R5cGWUk5SMCENvZGVUeXBllIWUUpQoSwFLAEsASwFLAUsTQwSIAFMAlE6FlCmMAV+UhZSMUy9ob21lL21heGltaWxpYW4vdmVudi9saWIvcHl0aG9uMy44L3NpdGUtcGFja2FnZXMvc3RhYmxlX2Jhc2VsaW5lczMvY29tbW9uL3V0aWxzLnB5lIwEZnVuY5RLgEMCAAGUjAN2YWyUhZQpdJRSlH2UKIwLX19wYWNrYWdlX1+UjBhzdGFibGVfYmFzZWxpbmVzMy5jb21tb26UjAhfX25hbWVfX5SMHnN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi51dGlsc5SMCF9fZmlsZV9flGgMdU5OaACMEF9tYWtlX2VtcHR5X2NlbGyUk5QpUpSFlHSUUpSMHGNsb3VkcGlja2xlLmNsb3VkcGlja2xlX2Zhc3SUjBJfZnVuY3Rpb25fc2V0c3RhdGWUk5RoHn2UfZQoaBZoDYwMX19xdWFsbmFtZV9flIwZY29uc3RhbnRfZm4uPGxvY2Fscz4uZnVuY5SMD19fYW5ub3RhdGlvbnNfX5R9lIwOX19rd2RlZmF1bHRzX1+UTowMX19kZWZhdWx0c19flE6MCl9fbW9kdWxlX1+UaBeMB19fZG9jX1+UTowLX19jbG9zdXJlX1+UaACMCl9tYWtlX2NlbGyUk5RHPx/V8usTiESFlFKUhZSMF19jbG91ZHBpY2tsZV9zdWJtb2R1bGVzlF2UjAtfX2dsb2JhbHNfX5R9lHWGlIZSMC4="
|
| 81 |
},
|
| 82 |
"_last_obs": null,
|
| 83 |
"_last_episode_starts": {
|
|
|
|
| 86 |
},
|
| 87 |
"_last_original_obs": {
|
| 88 |
":type:": "<class 'numpy.ndarray'>",
|
| 89 |
+
":serialized:": "gAWVxQAAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJZQAAAAAAAAADSLDgK8o5+/Ls3qY8nbuD/YK6kN6u2DP91sbYiQz6O/SGkGxpnWtL+wwT0ewOB/v10XlRZey6i/yIQS66WJsr+GJMnOSeGyP6R4WnIMx6k/lIwFbnVtcHmUjAVkdHlwZZSTlIwCZjiUiYiHlFKUKEsDjAE8lE5OTkr/////Sv////9LAHSUYksBSwqGlIwBQ5R0lFKULg=="
|
| 90 |
},
|
| 91 |
"_episode_num": 0,
|
| 92 |
"use_sde": false,
|
|
|
|
| 94 |
"_current_progress_remaining": -0.0014719999999999178,
|
| 95 |
"ep_info_buffer": {
|
| 96 |
":type:": "<class 'collections.deque'>",
|
| 97 |
+
":serialized:": "gAWVgRAAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKUKH2UKIwBcpSMFW51bXB5LmNvcmUubXVsdGlhcnJheZSMBnNjYWxhcpSTlIwFbnVtcHmUjAVkdHlwZZSTlIwCZjiUiYiHlFKUKEsDjAE8lE5OTkr/////Sv////9LAHSUYkMIs+veigQLdUCUhpRSlIwBbJRN6AOMAXSUR0DEvMXg3tKJdX2UKGgGaAloD0MIDjLJyNkxdUCUhpRSlGgVTegDaBZHQMS9QbiIcip1fZQoaAZoCWgPQwgWhsjp6xB1QJSGlFKUaBVN6ANoFkdAxMeXCBPKuHV9lChoBmgJaA9DCAJFLGKYxHRAlIaUUpRoFU3oA2gWR0DEyBGNkvsadX2UKGgGaAloD0MI1EM0uoOkdECUhpRSlGgVTegDaBZHQMTSO7fYSQJ1fZQoaAZoCWgPQwhMpgpGZdN0QJSGlFKUaBVN6ANoFkdAxNK2AMDwIHV9lChoBmgJaA9DCFzn3y6753RAlIaUUpRoFU3oA2gWR0DE3O1SZSeidX2UKGgGaAloD0MIXYqryn5zdUCUhpRSlGgVTegDaBZHQMTdaDEWIoF1fZQoaAZoCWgPQwgT86yk1V51QJSGlFKUaBVN6ANoFkdAxOe5+1Bt13V9lChoBmgJaA9DCHUg66kVpHRAlIaUUpRoFU3oA2gWR0DE6DanFYMfdX2UKGgGaAloD0MIJ9h/nRsndUCUhpRSlGgVTegDaBZHQMTyiNHpbEB1fZQoaAZoCWgPQwizQpHuJ5N1QJSGlFKUaBVN6ANoFkdAxPL/QJokA3V9lChoBmgJaA9DCEeQSrEjWXVAlIaUUpRoFU3oA2gWR0DE/TBT4tYkdX2UKGgGaAloD0MIt0WZDfLQdECUhpRSlGgVTegDaBZHQMT9q127nPp1fZQoaAZoCWgPQwgNiuYB7Mt0QJSGlFKUaBVN6ANoFkdAxQfZKwpvxnV9lChoBmgJaA9DCJnVO9yOAHVAlIaUUpRoFU3oA2gWR0DFCFTkIX0odX2UKGgGaAloD0MI7zuGx76RdUCUhpRSlGgVTegDaBZHQMUTkFg+hXd1fZQoaAZoCWgPQwj3WztRUr51QJSGlFKUaBVN6ANoFkdAxRQkQCjk/HV9lChoBmgJaA9DCMYwJ2jTtnRAlIaUUpRoFU3oA2gWR0DFHsQ0CRwIdX2UKGgGaAloD0MIwRpn01H5dECUhpRSlGgVTegDaBZHQMUfQAhStNl1fZQoaAZoCWgPQwhAahMnt8x1QJSGlFKUaBVN6ANoFkdAxSlxydWhiHV9lChoBmgJaA9DCCC29Ghql3VAlIaUUpRoFU3oA2gWR0DFKek/r0J4dX2UKGgGaAloD0MI0ETY8HRSdUCUhpRSlGgVTegDaBZHQMU0JgTZg5R1fZQoaAZoCWgPQwhNvtnmxnx0QJSGlFKUaBVN6ANoFkdAxTSnEk0JnnV9lChoBmgJaA9DCCXnxB6aSXVAlIaUUpRoFU3oA2gWR0DFQccvysjndX2UKGgGaAloD0MI6xotB3padUCUhpRSlGgVTegDaBZHQMVCQt4iX6Z1fZQoaAZoCWgPQwj/lgD8k+h0QJSGlFKUaBVN6ANoFkdAxUxqjJMg2nV9lChoBmgJaA9DCL3hPnIreHVAlIaUUpRoFU3oA2gWR0DFTOaYNRWMdX2UKGgGaAloD0MIaThlbn4GdUCUhpRSlGgVTegDaBZHQMVXHgGKQ7t1fZQoaAZoCWgPQwjRdeEHpzh1QJSGlFKUaBVN6ANoFkdAxVeW4p+c6XV9lChoBmgJaA9DCD+toj90xHRAlIaUUpRoFU3oA2gWR0DFYcb+HaexdX2UKGgGaAloD0MIdsHgmrs3dUCUhpRSlGgVTegDaBZHQMViSJdSl311fZQoaAZoCWgPQwj5u3fUmIh0QJSGlFKUaBVN6ANoFkdAxWySIznA7HV9lChoBmgJaA9DCE3aVN0jh3RAlIaUUpRoFU3oA2gWR0DFbQuvW6K+dX2UKGgGaAloD0MIN8XjohrTdUCUhpRSlGgVTegDaBZHQMV3M6Fdszl1fZQoaAZoCWgPQwgDJ9vAXZ50QJSGlFKUaBVN6ANoFkdAxXesgW8AaXV9lChoBmgJaA9DCITVWMLa+XRAlIaUUpRoFU3oA2gWR0DFgdQqqfe2dX2UKGgGaAloD0MIQRGLGHbHdECUhpRSlGgVTegDaBZHQMWCTxffGdZ1fZQoaAZoCWgPQwgkQiPY+LR0QJSGlFKUaBVN6ANoFkdAxYLIwdKdx3V9lChoBmgJaA9DCDrJVpcTH3VAlIaUUpRoFU3oA2gWR0DFjYjYbsF/dX2UKGgGaAloD0MIZJC7CBNHdUCUhpRSlGgVTegDaBZHQMWOA6vq1PZ1fZQoaAZoCWgPQwhWgO82L2R0QJSGlFKUaBVN6ANoFkdAxZgpazu4PXV9lChoBmgJaA9DCFMFo5J6t3RAlIaUUpRoFU3oA2gWR0DFmKQBo24vdX2UKGgGaAloD0MIVACMZ9CidUCUhpRSlGgVTegDaBZHQMWi2RSpBHF1fZQoaAZoCWgPQwhoWfePxS91QJSGlFKUaBVN6ANoFkdAxaNQRQJokHV9lChoBmgJaA9DCLKd76dGw3VAlIaUUpRoFU3oA2gWR0DFrYEunMt9dX2UKGgGaAloD0MIC5krg+rAdUCUhpRSlGgVTegDaBZHQMWt+0jC53F1fZQoaAZoCWgPQwgcCTTYVFV1QJSGlFKUaBVN6ANoFkdAxbgkgRsdk3V9lChoBmgJaA9DCJ1n7Et2iHVAlIaUUpRoFU3oA2gWR0DFuKG01IiDdX2UKGgGaAloD0MI0/VE10VMdUCUhpRSlGgVTegDaBZHQMXEpZpztC11fZQoaAZoCWgPQwgjaw2lNjB1QJSGlFKUaBVN6ANoFkdAxcUgnUDuB3V9lChoBmgJaA9DCG2QSUbOX3VAlIaUUpRoFU3oA2gWR0DFz2FE/jbSdX2UKGgGaAloD0MI64uEtpwVdUCUhpRSlGgVTegDaBZHQMXP2jin5zp1fZQoaAZoCWgPQwgqAMYzKFd1QJSGlFKUaBVN6ANoFkdAxdoIy/sVtXV9lChoBmgJaA9DCEDbataZg3VAlIaUUpRoFU3oA2gWR0DF2oVO/L1VdX2UKGgGaAloD0MIEvjDz//BdUCUhpRSlGgVTegDaBZHQMXkwgYYR/V1fZQoaAZoCWgPQwg0oN6M2pB1QJSGlFKUaBVN6ANoFkdAxeU+ed07sHV9lChoBmgJaA9DCGOa6V7nA3ZAlIaUUpRoFU3oA2gWR0DF72/5tWMkdX2UKGgGaAloD0MIwTkjSvuqdUCUhpRSlGgVTegDaBZHQMXv7N03fhx1fZQoaAZoCWgPQwj3ViQmaD52QJSGlFKUaBVN6ANoFkdAxfoxYT0xunV9lChoBmgJaA9DCIXrUbgeQnVAlIaUUpRoFU3oA2gWR0DF+qzgdfb9dX2UKGgGaAloD0MI2eicn+IYdUCUhpRSlGgVTegDaBZHQMYE6jn/1g91fZQoaAZoCWgPQwgMA5ZchWp1QJSGlFKUaBVN6ANoFkdAxgVoX/HYH3V9lChoBmgJaA9DCL1SliGO8nRAlIaUUpRoFU3oA2gWR0DGD5HFefI0dX2UKGgGaAloD0MIqU4Hsl53dUCUhpRSlGgVTegDaBZHQMYQDIi1Rch1fZQoaAZoCWgPQwggelImNcl1QJSGlFKUaBVN6ANoFkdAxhpWC04R3HV9lChoBmgJaA9DCPqzHyniD3ZAlIaUUpRoFU3oA2gWR0DGGtIzch1UdX2UKGgGaAloD0MIOLpKd1dOdUCUhpRSlGgVTegDaBZHQMYlE/Q8fV91fZQoaAZoCWgPQwhc5nRZTFV1QJSGlFKUaBVN6ANoFkdAxiWLlA/s3XV9lChoBmgJaA9DCHfYRGZuinVAlIaUUpRoFU3oA2gWR0DGL8Aht+CsdX2UKGgGaAloD0MInIu/7QkddUCUhpRSlGgVTegDaBZHQMYwPKGL1mJ1fZQoaAZoCWgPQwiJQPUP4m11QJSGlFKUaBVN6ANoFkdAxjp06/7BPHV9lChoBmgJaA9DCDSD+MAOTnZAlIaUUpRoFU3oA2gWR0DGOu4I2OyWdX2UKGgGaAloD0MIKEaWzDFgdUCUhpRSlGgVTegDaBZHQMZFJVEd/8V1fZQoaAZoCWgPQwjkE7LzNiZ2QJSGlFKUaBVN6ANoFkdAxkdl8fFJhHV9lChoBmgJaA9DCHJw6ZizTHVAlIaUUpRoFU3oA2gWR0DGUanrY5DJdX2UKGgGaAloD0MIkPRpFf1HdUCUhpRSlGgVTegDaBZHQMZSIrv1DjR1fZQoaAZoCWgPQwibAS7IVgF2QJSGlFKUaBVN6ANoFkdAxlxiHIp6QnV9lChoBmgJaA9DCJp4B3hS0XVAlIaUUpRoFU3oA2gWR0DGXN/L/0dzdX2UKGgGaAloD0MIPRBZpIlHdUCUhpRSlGgVTegDaBZHQMZnB+oDPnl1fZQoaAZoCWgPQwgrieyDrHZ1QJSGlFKUaBVN6ANoFkdAxmeCadc0L3V9lChoBmgJaA9DCJlmutfJ+XVAlIaUUpRoFU3oA2gWR0DGZ/Zof0VadX2UKGgGaAloD0MIkxgEVs40dkCUhpRSlGgVTegDaBZHQMZyIxoysS11fZQoaAZoCWgPQwjedqG5Tj52QJSGlFKUaBVN6ANoFkdAxnKeHhS9/XV9lChoBmgJaA9DCOrnTUXqmnVAlIaUUpRoFU3oA2gWR0DGfMqjFhoedX2UKGgGaAloD0MIRj8aThnldUCUhpRSlGgVTegDaBZHQMZ9RHaews51fZQoaAZoCWgPQwg26iEaXYp1QJSGlFKUaBVN6ANoFkdAxodinqFAV3V9lChoBmgJaA9DCGjon+DiMnVAlIaUUpRoFU3oA2gWR0DGh92M4tHydX2UKGgGaAloD0MIqvHSTSIRdkCUhpRSlGgVTegDaBZHQMaSC6ZH/cZ1fZQoaAZoCWgPQwhZp8r3zIF1QJSGlFKUaBVN6ANoFkdAxpKFg2IfsHV9lChoBmgJaA9DCNtv7USJC3VAlIaUUpRoFU3oA2gWR0DGnLojyFwldX2UKGgGaAloD0MIp7OTwZEvdkCUhpRSlGgVTegDaBZHQMadNAf2bod1fZQoaAZoCWgPQwg1fuGVZOB1QJSGlFKUaBVN6ANoFkdAxqdWhTOxB3V9lChoBmgJaA9DCFPr/UZ7HXVAlIaUUpRoFU3oA2gWR0DGp9EU9IPLdX2UKGgGaAloD0MIIjfDDbgddkCUhpRSlGgVTegDaBZHQMax9NLUTct1fZQoaAZoCWgPQwg4LuOmBgd2QJSGlFKUaBVN6ANoFkdAxrJtrY5DJHV9lChoBmgJaA9DCFMj9DM1e3VAlIaUUpRoFU3oA2gWR0DGvJ5RXOnmdX2UKGgGaAloD0MIRGywcBKKdUCUhpRSlGgVTegDaBZHQMa9GSUs4DN1fZQoaAZoCWgPQwgNjSeCOOZ1QJSGlFKUaBVN6ANoFkdAxscs7o0Q9XV9lChoBmgJaA9DCE62gTvQsXVAlIaUUpRoFU3oA2gWR0DGyWFotcv/dWUu"
|
| 98 |
},
|
| 99 |
"ep_success_buffer": {
|
| 100 |
":type:": "<class 'collections.deque'>",
|
|
|
|
| 111 |
"n_epochs": 20,
|
| 112 |
"clip_range": {
|
| 113 |
":type:": "<class 'function'>",
|
| 114 |
+
":serialized:": "gAWVhQIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwOX21ha2VfZnVuY3Rpb26Uk5QoaACMDV9idWlsdGluX3R5cGWUk5SMCENvZGVUeXBllIWUUpQoSwFLAEsASwFLAUsTQwSIAFMAlE6FlCmMAV+UhZSMUy9ob21lL21heGltaWxpYW4vdmVudi9saWIvcHl0aG9uMy44L3NpdGUtcGFja2FnZXMvc3RhYmxlX2Jhc2VsaW5lczMvY29tbW9uL3V0aWxzLnB5lIwEZnVuY5RLgEMCAAGUjAN2YWyUhZQpdJRSlH2UKIwLX19wYWNrYWdlX1+UjBhzdGFibGVfYmFzZWxpbmVzMy5jb21tb26UjAhfX25hbWVfX5SMHnN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi51dGlsc5SMCF9fZmlsZV9flGgMdU5OaACMEF9tYWtlX2VtcHR5X2NlbGyUk5QpUpSFlHSUUpSMHGNsb3VkcGlja2xlLmNsb3VkcGlja2xlX2Zhc3SUjBJfZnVuY3Rpb25fc2V0c3RhdGWUk5RoHn2UfZQoaBZoDYwMX19xdWFsbmFtZV9flIwZY29uc3RhbnRfZm4uPGxvY2Fscz4uZnVuY5SMD19fYW5ub3RhdGlvbnNfX5R9lIwOX19rd2RlZmF1bHRzX1+UTowMX19kZWZhdWx0c19flE6MCl9fbW9kdWxlX1+UaBeMB19fZG9jX1+UTowLX19jbG9zdXJlX1+UaACMCl9tYWtlX2NlbGyUk5RHP7mZmZmZmZqFlFKUhZSMF19jbG91ZHBpY2tsZV9zdWJtb2R1bGVzlF2UjAtfX2dsb2JhbHNfX5R9lHWGlIZSMC4="
|
| 115 |
},
|
| 116 |
"clip_range_vf": null,
|
| 117 |
"normalize_advantage": true,
|
ppo-seals-Swimmer-v0/policy.optimizer.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:27db56d314e8454595d15e351d63481f2c0a895a3b3385b7f93542a010957f3f
|
| 3 |
+
size 89328
|
ppo-seals-Swimmer-v0/policy.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:102dea9605c32c2f519e9eebc913e8081b3f1ffa72695ca4560a82ef3c76ac1a
|
| 3 |
+
size 44917
|
ppo-seals-Swimmer-v0/system_info.txt
CHANGED
|
@@ -1,6 +1,6 @@
|
|
| 1 |
-
OS: Linux-5.4.0-
|
| 2 |
Python: 3.8.10
|
| 3 |
-
Stable-Baselines3: 1.6.
|
| 4 |
PyTorch: 1.11.0+cu102
|
| 5 |
GPU Enabled: False
|
| 6 |
Numpy: 1.22.3
|
|
|
|
| 1 |
+
OS: Linux-5.4.0-125-generic-x86_64-with-glibc2.29 #141-Ubuntu SMP Wed Aug 10 13:42:03 UTC 2022
|
| 2 |
Python: 3.8.10
|
| 3 |
+
Stable-Baselines3: 1.6.2
|
| 4 |
PyTorch: 1.11.0+cu102
|
| 5 |
GPU Enabled: False
|
| 6 |
Numpy: 1.22.3
|
replay.mp4
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9ca6c99625ec8c3009852c2318d9e2b942d02db7a2a4aac6cbc4c1fa687a054a
|
| 3 |
+
size 1427851
|
results.json
CHANGED
|
@@ -1 +1 @@
|
|
| 1 |
-
{"mean_reward":
|
|
|
|
| 1 |
+
{"mean_reward": 347.7021517, "std_reward": 5.879248281492488, "is_deterministic": true, "n_eval_episodes": 10, "eval_datetime": "2022-12-31T18:32:13.524726"}
|
train_eval_metrics.zip
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:aa47e56f834c1e15ab109e044748b33a1cde7a4d9997e8df7c41901019ef2cc7
|
| 3 |
+
size 33383
|
vec_normalize.pkl
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:bfbcacb608931d53fe0942c153afade94aa893b39f29840c9ae9f5d85fca04f0
|
| 3 |
+
size 4125
|