Initial commit
Browse files- README.md +12 -4
- args.yml +11 -5
- config.yml +5 -1
- replay.mp4 +2 -2
- results.json +1 -1
- sac-seals-Walker2d-v0.zip +2 -2
- sac-seals-Walker2d-v0/_stable_baselines3_version +1 -1
- sac-seals-Walker2d-v0/actor.optimizer.pth +1 -1
- sac-seals-Walker2d-v0/critic.optimizer.pth +1 -1
- sac-seals-Walker2d-v0/data +27 -25
- sac-seals-Walker2d-v0/ent_coef_optimizer.pth +2 -2
- sac-seals-Walker2d-v0/policy.pth +1 -1
- sac-seals-Walker2d-v0/pytorch_variables.pth +1 -1
- sac-seals-Walker2d-v0/system_info.txt +2 -2
- train_eval_metrics.zip +2 -2
README.md
CHANGED
|
@@ -10,7 +10,7 @@ model-index:
|
|
| 10 |
results:
|
| 11 |
- metrics:
|
| 12 |
- type: mean_reward
|
| 13 |
-
value:
|
| 14 |
name: mean_reward
|
| 15 |
task:
|
| 16 |
type: reinforcement-learning
|
|
@@ -37,15 +37,21 @@ SB3 Contrib: https://github.com/Stable-Baselines-Team/stable-baselines3-contrib
|
|
| 37 |
|
| 38 |
```
|
| 39 |
# Download model and save it into the logs/ folder
|
| 40 |
-
python -m
|
| 41 |
python enjoy.py --algo sac --env seals/Walker2d-v0 -f logs/
|
| 42 |
```
|
| 43 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 44 |
## Training (with the RL Zoo)
|
| 45 |
```
|
| 46 |
python train.py --algo sac --env seals/Walker2d-v0 -f logs/
|
| 47 |
# Upload the model and generate video (when possible)
|
| 48 |
-
python -m
|
| 49 |
```
|
| 50 |
|
| 51 |
## Hyperparameters
|
|
@@ -58,7 +64,9 @@ OrderedDict([('batch_size', 128),
|
|
| 58 |
('n_timesteps', 1000000.0),
|
| 59 |
('policy', 'MlpPolicy'),
|
| 60 |
('policy_kwargs',
|
| 61 |
-
'
|
|
|
|
|
|
|
| 62 |
('tau', 0.02),
|
| 63 |
('train_freq', 1),
|
| 64 |
('normalize', False)])
|
|
|
|
| 10 |
results:
|
| 11 |
- metrics:
|
| 12 |
- type: mean_reward
|
| 13 |
+
value: 2492.52 +/- 1181.09
|
| 14 |
name: mean_reward
|
| 15 |
task:
|
| 16 |
type: reinforcement-learning
|
|
|
|
| 37 |
|
| 38 |
```
|
| 39 |
# Download model and save it into the logs/ folder
|
| 40 |
+
python -m rl_zoo3.load_from_hub --algo sac --env seals/Walker2d-v0 -orga HumanCompatibleAI -f logs/
|
| 41 |
python enjoy.py --algo sac --env seals/Walker2d-v0 -f logs/
|
| 42 |
```
|
| 43 |
|
| 44 |
+
If you installed the RL Zoo3 via pip (`pip install rl_zoo3`), from anywhere you can do:
|
| 45 |
+
```
|
| 46 |
+
python -m rl_zoo3.load_from_hub --algo sac --env seals/Walker2d-v0 -orga HumanCompatibleAI -f logs/
|
| 47 |
+
rl_zoo3 enjoy --algo sac --env seals/Walker2d-v0 -f logs/
|
| 48 |
+
```
|
| 49 |
+
|
| 50 |
## Training (with the RL Zoo)
|
| 51 |
```
|
| 52 |
python train.py --algo sac --env seals/Walker2d-v0 -f logs/
|
| 53 |
# Upload the model and generate video (when possible)
|
| 54 |
+
python -m rl_zoo3.push_to_hub --algo sac --env seals/Walker2d-v0 -f logs/ -orga HumanCompatibleAI
|
| 55 |
```
|
| 56 |
|
| 57 |
## Hyperparameters
|
|
|
|
| 64 |
('n_timesteps', 1000000.0),
|
| 65 |
('policy', 'MlpPolicy'),
|
| 66 |
('policy_kwargs',
|
| 67 |
+
{'log_std_init': 0.1955317469998743,
|
| 68 |
+
'net_arch': [400, 300],
|
| 69 |
+
'use_sde': False}),
|
| 70 |
('tau', 0.02),
|
| 71 |
('train_freq', 1),
|
| 72 |
('normalize', False)])
|
args.yml
CHANGED
|
@@ -1,6 +1,8 @@
|
|
| 1 |
!!python/object/apply:collections.OrderedDict
|
| 2 |
- - - algo
|
| 3 |
- sac
|
|
|
|
|
|
|
| 4 |
- - device
|
| 5 |
- cpu
|
| 6 |
- - env
|
|
@@ -16,7 +18,7 @@
|
|
| 16 |
- - hyperparams
|
| 17 |
- null
|
| 18 |
- - log_folder
|
| 19 |
-
-
|
| 20 |
- - log_interval
|
| 21 |
- -1
|
| 22 |
- - max_total_trials
|
|
@@ -41,6 +43,8 @@
|
|
| 41 |
- null
|
| 42 |
- - optimize_hyperparameters
|
| 43 |
- false
|
|
|
|
|
|
|
| 44 |
- - pruner
|
| 45 |
- median
|
| 46 |
- - sampler
|
|
@@ -50,13 +54,13 @@
|
|
| 50 |
- - save_replay_buffer
|
| 51 |
- false
|
| 52 |
- - seed
|
| 53 |
-
-
|
| 54 |
- - storage
|
| 55 |
- null
|
| 56 |
- - study_name
|
| 57 |
- null
|
| 58 |
- - tensorboard_log
|
| 59 |
-
- runs/seals/Walker2d-
|
| 60 |
- - track
|
| 61 |
- true
|
| 62 |
- - trained_agent
|
|
@@ -70,6 +74,8 @@
|
|
| 70 |
- - verbose
|
| 71 |
- 1
|
| 72 |
- - wandb_entity
|
| 73 |
-
-
|
| 74 |
- - wandb_project_name
|
| 75 |
-
- seals-experts-
|
|
|
|
|
|
|
|
|
| 1 |
!!python/object/apply:collections.OrderedDict
|
| 2 |
- - - algo
|
| 3 |
- sac
|
| 4 |
+
- - conf_file
|
| 5 |
+
- hyperparams/python/sac.py
|
| 6 |
- - device
|
| 7 |
- cpu
|
| 8 |
- - env
|
|
|
|
| 18 |
- - hyperparams
|
| 19 |
- null
|
| 20 |
- - log_folder
|
| 21 |
+
- logs
|
| 22 |
- - log_interval
|
| 23 |
- -1
|
| 24 |
- - max_total_trials
|
|
|
|
| 43 |
- null
|
| 44 |
- - optimize_hyperparameters
|
| 45 |
- false
|
| 46 |
+
- - progress
|
| 47 |
+
- false
|
| 48 |
- - pruner
|
| 49 |
- median
|
| 50 |
- - sampler
|
|
|
|
| 54 |
- - save_replay_buffer
|
| 55 |
- false
|
| 56 |
- - seed
|
| 57 |
+
- 5
|
| 58 |
- - storage
|
| 59 |
- null
|
| 60 |
- - study_name
|
| 61 |
- null
|
| 62 |
- - tensorboard_log
|
| 63 |
+
- runs/seals/Walker2d-v0__sac__5__1672507712
|
| 64 |
- - track
|
| 65 |
- true
|
| 66 |
- - trained_agent
|
|
|
|
| 74 |
- - verbose
|
| 75 |
- 1
|
| 76 |
- - wandb_entity
|
| 77 |
+
- ernestum
|
| 78 |
- - wandb_project_name
|
| 79 |
+
- seals-experts-normalized
|
| 80 |
+
- - yaml_file
|
| 81 |
+
- null
|
config.yml
CHANGED
|
@@ -14,7 +14,11 @@
|
|
| 14 |
- - policy
|
| 15 |
- MlpPolicy
|
| 16 |
- - policy_kwargs
|
| 17 |
-
-
|
|
|
|
|
|
|
|
|
|
|
|
|
| 18 |
- - tau
|
| 19 |
- 0.02
|
| 20 |
- - train_freq
|
|
|
|
| 14 |
- - policy
|
| 15 |
- MlpPolicy
|
| 16 |
- - policy_kwargs
|
| 17 |
+
- log_std_init: 0.1955317469998743
|
| 18 |
+
net_arch:
|
| 19 |
+
- 400
|
| 20 |
+
- 300
|
| 21 |
+
use_sde: false
|
| 22 |
- - tau
|
| 23 |
- 0.02
|
| 24 |
- - train_freq
|
replay.mp4
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f1d6d32b7ac2c7937e297249eecdbba14be9b3dc1b809e1415eacc0583980fb6
|
| 3 |
+
size 1373262
|
results.json
CHANGED
|
@@ -1 +1 @@
|
|
| 1 |
-
{"mean_reward":
|
|
|
|
| 1 |
+
{"mean_reward": 2492.5241772, "std_reward": 1181.0920613431326, "is_deterministic": true, "n_eval_episodes": 10, "eval_datetime": "2023-01-02T11:00:28.709939"}
|
sac-seals-Walker2d-v0.zip
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:42b9465decbc006ee3c68af4bd6e6166d0ac22a72cd30dd7bb7dab2e2bd80d95
|
| 3 |
+
size 5806513
|
sac-seals-Walker2d-v0/_stable_baselines3_version
CHANGED
|
@@ -1 +1 @@
|
|
| 1 |
-
1.6.
|
|
|
|
| 1 |
+
1.6.2
|
sac-seals-Walker2d-v0/actor.optimizer.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1058141
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1bcdfff6139b595523b30377aef80b9384a8ebb756b53fd10a3ba8542228d123
|
| 3 |
size 1058141
|
sac-seals-Walker2d-v0/critic.optimizer.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 2098489
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1a430cf9de43e68c4ba2a25448e0c7e7a89ff1736d38d71ff2492cddacc00b1d
|
| 3 |
size 2098489
|
sac-seals-Walker2d-v0/data
CHANGED
|
@@ -4,17 +4,17 @@
|
|
| 4 |
":serialized:": "gAWVMAAAAAAAAACMHnN0YWJsZV9iYXNlbGluZXMzLnNhYy5wb2xpY2llc5SMCVNBQ1BvbGljeZSTlC4=",
|
| 5 |
"__module__": "stable_baselines3.sac.policies",
|
| 6 |
"__doc__": "\n Policy class (with both actor and critic) for SAC.\n\n :param observation_space: Observation space\n :param action_space: Action space\n :param lr_schedule: Learning rate schedule (could be constant)\n :param net_arch: The specification of the policy and value networks.\n :param activation_fn: Activation function\n :param use_sde: Whether to use State Dependent Exploration or not\n :param log_std_init: Initial value for the log standard deviation\n :param sde_net_arch: Network architecture for extracting features\n when using gSDE. If None, the latent features from the policy will be used.\n Pass an empty list to use the states as features.\n :param use_expln: Use ``expln()`` function instead of ``exp()`` when using gSDE to ensure\n a positive standard deviation (cf paper). It allows to keep variance\n above zero and prevent it from growing too fast. In practice, ``exp()`` is usually enough.\n :param clip_mean: Clip the mean output when using gSDE to avoid numerical instability.\n :param features_extractor_class: Features extractor to use.\n :param features_extractor_kwargs: Keyword arguments\n to pass to the features extractor.\n :param normalize_images: Whether to normalize images or not,\n dividing by 255.0 (True by default)\n :param optimizer_class: The optimizer to use,\n ``th.optim.Adam`` by default\n :param optimizer_kwargs: Additional keyword arguments,\n excluding the learning rate, to pass to the optimizer\n :param n_critics: Number of critic networks to create.\n :param share_features_extractor: Whether to share or not the features extractor\n between the actor and the critic (this saves computation time)\n ",
|
| 7 |
-
"__init__": "<function SACPolicy.__init__ at
|
| 8 |
-
"_build": "<function SACPolicy._build at
|
| 9 |
-
"_get_constructor_parameters": "<function SACPolicy._get_constructor_parameters at
|
| 10 |
-
"reset_noise": "<function SACPolicy.reset_noise at
|
| 11 |
-
"make_actor": "<function SACPolicy.make_actor at
|
| 12 |
-
"make_critic": "<function SACPolicy.make_critic at
|
| 13 |
-
"forward": "<function SACPolicy.forward at
|
| 14 |
-
"_predict": "<function SACPolicy._predict at
|
| 15 |
-
"set_training_mode": "<function SACPolicy.set_training_mode at
|
| 16 |
"__abstractmethods__": "frozenset()",
|
| 17 |
-
"_abc_impl": "<_abc_data object at
|
| 18 |
},
|
| 19 |
"verbose": 1,
|
| 20 |
"policy_kwargs": {
|
|
@@ -40,7 +40,7 @@
|
|
| 40 |
},
|
| 41 |
"action_space": {
|
| 42 |
":type:": "<class 'gym.spaces.box.Box'>",
|
| 43 |
-
":serialized:": "gAWVEwwAAAAAAACMDmd5bS5zcGFjZXMuYm94lIwDQm94lJOUKYGUfZQojAVkdHlwZZSMBW51bXB5lGgFk5SMAmY0lImIh5RSlChLA4wBPJROTk5K/////0r/////SwB0lGKMBl9zaGFwZZRLBoWUjANsb3eUjBJudW1weS5jb3JlLm51bWVyaWOUjAtfZnJvbWJ1ZmZlcpSTlCiWGAAAAAAAAAAAAIC/AACAvwAAgL8AAIC/AACAvwAAgL+UaApLBoWUjAFDlHSUUpSMBGhpZ2iUaBIolhgAAAAAAAAAAACAPwAAgD8AAIA/AACAPwAAgD8AAIA/lGgKSwaFlGgVdJRSlIwNYm91bmRlZF9iZWxvd5RoEiiWBgAAAAAAAAABAQEBAQGUaAeMAmIxlImIh5RSlChLA4wBfJROTk5K/////0r/////
|
| 44 |
"dtype": "float32",
|
| 45 |
"_shape": [
|
| 46 |
6
|
|
@@ -55,17 +55,17 @@
|
|
| 55 |
"num_timesteps": 1000000,
|
| 56 |
"_total_timesteps": 1000000,
|
| 57 |
"_num_timesteps_at_start": 0,
|
| 58 |
-
"seed":
|
| 59 |
"action_noise": null,
|
| 60 |
-
"start_time":
|
| 61 |
"learning_rate": {
|
| 62 |
":type:": "<class 'function'>",
|
| 63 |
-
":serialized:": "
|
| 64 |
},
|
| 65 |
-
"tensorboard_log": "runs/seals/Walker2d-
|
| 66 |
"lr_schedule": {
|
| 67 |
":type:": "<class 'function'>",
|
| 68 |
-
":serialized:": "
|
| 69 |
},
|
| 70 |
"_last_obs": null,
|
| 71 |
"_last_episode_starts": {
|
|
@@ -74,7 +74,7 @@
|
|
| 74 |
},
|
| 75 |
"_last_original_obs": {
|
| 76 |
":type:": "<class 'numpy.ndarray'>",
|
| 77 |
-
":serialized:": "
|
| 78 |
},
|
| 79 |
"_episode_num": 1000,
|
| 80 |
"use_sde": false,
|
|
@@ -82,7 +82,7 @@
|
|
| 82 |
"_current_progress_remaining": 0.0,
|
| 83 |
"ep_info_buffer": {
|
| 84 |
":type:": "<class 'collections.deque'>",
|
| 85 |
-
":serialized:": "gAWVgRAAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKUKH2UKIwBcpSMFW51bXB5LmNvcmUubXVsdGlhcnJheZSMBnNjYWxhcpSTlIwFbnVtcHmUjAVkdHlwZZSTlIwCZjiUiYiHlFKUKEsDjAE8lE5OTkr/////Sv////
|
| 86 |
},
|
| 87 |
"ep_success_buffer": {
|
| 88 |
":type:": "<class 'collections.deque'>",
|
|
@@ -100,13 +100,13 @@
|
|
| 100 |
":type:": "<class 'abc.ABCMeta'>",
|
| 101 |
":serialized:": "gAWVNQAAAAAAAACMIHN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi5idWZmZXJzlIwMUmVwbGF5QnVmZmVylJOULg==",
|
| 102 |
"__module__": "stable_baselines3.common.buffers",
|
| 103 |
-
"__doc__": "\n Replay buffer used in off-policy algorithms like SAC/TD3.\n\n :param buffer_size: Max number of element in the buffer\n :param observation_space: Observation space\n :param action_space: Action space\n :param device
|
| 104 |
-
"__init__": "<function ReplayBuffer.__init__ at
|
| 105 |
-
"add": "<function ReplayBuffer.add at
|
| 106 |
-
"sample": "<function ReplayBuffer.sample at
|
| 107 |
-
"_get_samples": "<function ReplayBuffer._get_samples at
|
| 108 |
"__abstractmethods__": "frozenset()",
|
| 109 |
-
"_abc_impl": "<_abc_data object at
|
| 110 |
},
|
| 111 |
"replay_buffer_kwargs": {},
|
| 112 |
"train_freq": {
|
|
@@ -116,5 +116,7 @@
|
|
| 116 |
"use_sde_at_warmup": false,
|
| 117 |
"target_entropy": -6.0,
|
| 118 |
"ent_coef": "auto",
|
| 119 |
-
"target_update_interval": 1
|
|
|
|
|
|
|
| 120 |
}
|
|
|
|
| 4 |
":serialized:": "gAWVMAAAAAAAAACMHnN0YWJsZV9iYXNlbGluZXMzLnNhYy5wb2xpY2llc5SMCVNBQ1BvbGljeZSTlC4=",
|
| 5 |
"__module__": "stable_baselines3.sac.policies",
|
| 6 |
"__doc__": "\n Policy class (with both actor and critic) for SAC.\n\n :param observation_space: Observation space\n :param action_space: Action space\n :param lr_schedule: Learning rate schedule (could be constant)\n :param net_arch: The specification of the policy and value networks.\n :param activation_fn: Activation function\n :param use_sde: Whether to use State Dependent Exploration or not\n :param log_std_init: Initial value for the log standard deviation\n :param sde_net_arch: Network architecture for extracting features\n when using gSDE. If None, the latent features from the policy will be used.\n Pass an empty list to use the states as features.\n :param use_expln: Use ``expln()`` function instead of ``exp()`` when using gSDE to ensure\n a positive standard deviation (cf paper). It allows to keep variance\n above zero and prevent it from growing too fast. In practice, ``exp()`` is usually enough.\n :param clip_mean: Clip the mean output when using gSDE to avoid numerical instability.\n :param features_extractor_class: Features extractor to use.\n :param features_extractor_kwargs: Keyword arguments\n to pass to the features extractor.\n :param normalize_images: Whether to normalize images or not,\n dividing by 255.0 (True by default)\n :param optimizer_class: The optimizer to use,\n ``th.optim.Adam`` by default\n :param optimizer_kwargs: Additional keyword arguments,\n excluding the learning rate, to pass to the optimizer\n :param n_critics: Number of critic networks to create.\n :param share_features_extractor: Whether to share or not the features extractor\n between the actor and the critic (this saves computation time)\n ",
|
| 7 |
+
"__init__": "<function SACPolicy.__init__ at 0x7ff1e3c87ee0>",
|
| 8 |
+
"_build": "<function SACPolicy._build at 0x7ff1e3c87f70>",
|
| 9 |
+
"_get_constructor_parameters": "<function SACPolicy._get_constructor_parameters at 0x7ff1e3c10040>",
|
| 10 |
+
"reset_noise": "<function SACPolicy.reset_noise at 0x7ff1e3c100d0>",
|
| 11 |
+
"make_actor": "<function SACPolicy.make_actor at 0x7ff1e3c10160>",
|
| 12 |
+
"make_critic": "<function SACPolicy.make_critic at 0x7ff1e3c101f0>",
|
| 13 |
+
"forward": "<function SACPolicy.forward at 0x7ff1e3c10280>",
|
| 14 |
+
"_predict": "<function SACPolicy._predict at 0x7ff1e3c10310>",
|
| 15 |
+
"set_training_mode": "<function SACPolicy.set_training_mode at 0x7ff1e3c103a0>",
|
| 16 |
"__abstractmethods__": "frozenset()",
|
| 17 |
+
"_abc_impl": "<_abc_data object at 0x7ff1e3c86cc0>"
|
| 18 |
},
|
| 19 |
"verbose": 1,
|
| 20 |
"policy_kwargs": {
|
|
|
|
| 40 |
},
|
| 41 |
"action_space": {
|
| 42 |
":type:": "<class 'gym.spaces.box.Box'>",
|
| 43 |
+
":serialized:": "gAWVEwwAAAAAAACMDmd5bS5zcGFjZXMuYm94lIwDQm94lJOUKYGUfZQojAVkdHlwZZSMBW51bXB5lGgFk5SMAmY0lImIh5RSlChLA4wBPJROTk5K/////0r/////SwB0lGKMBl9zaGFwZZRLBoWUjANsb3eUjBJudW1weS5jb3JlLm51bWVyaWOUjAtfZnJvbWJ1ZmZlcpSTlCiWGAAAAAAAAAAAAIC/AACAvwAAgL8AAIC/AACAvwAAgL+UaApLBoWUjAFDlHSUUpSMBGhpZ2iUaBIolhgAAAAAAAAAAACAPwAAgD8AAIA/AACAPwAAgD8AAIA/lGgKSwaFlGgVdJRSlIwNYm91bmRlZF9iZWxvd5RoEiiWBgAAAAAAAAABAQEBAQGUaAeMAmIxlImIh5RSlChLA4wBfJROTk5K/////0r/////SwB0lGJLBoWUaBV0lFKUjA1ib3VuZGVkX2Fib3ZllGgSKJYGAAAAAAAAAAEBAQEBAZRoIUsGhZRoFXSUUpSMCl9ucF9yYW5kb22UjBRudW1weS5yYW5kb20uX3BpY2tsZZSMEl9fcmFuZG9tc3RhdGVfY3RvcpSTlIwHTVQxOTkzN5SFlFKUfZQojA1iaXRfZ2VuZXJhdG9ylGgwjAVzdGF0ZZR9lCiMA2tleZRoEiiWwAkAAAAAAAAAAACA6tN/xWH5wD0oAa1E8OERvEjDYHkfreNmqh4QKt4HW7JZN7H00QJoY2idx8vw52ILeV09wf7H9uSOspEY6SeEKg4LS00Ddxk7dv8QQctbl0qMB+rcHr+TWXcQcbAq35aY5P7OxpdMQRlbjJbbbfnmF1Hv43cL/nHbV1u4nUUjWFla8ouYUvfnLoE63QquxZMPaU4t5xfsVlVZEJitFjDnZAN8CmzaBN8BvxmdUYf/qFE1vlkhpFk4XZTv9UHpfbhwllQFJtkqXTULW1VsByAmZ1OFo5fpJCy0Mv5/U4jfrLFCNLFHmgtwHd+lkxeGKSScibI5vWrl0U161lQhpLgLZZEVQpVKoguvbBXMaTPYJlfu0BFSgoN/2yWossX+9AS1tclZqSU8SC47dMOBZUX3Q8zUe9EET2+7+DPHL2zzsqoKPTri33+Xrh2qtYokX18RehQ9kkLHQ8j2IZV8nshrU82myYzGkmBOUk/N7PSTjxXdhoAm6NKcvKcCjYF+LM1/zf1pwLtF1rZQH/S9J5btwynyaP+DfnBmdI/OKzBgzv5yFq9Rjc+S/g7AAllaFPZYKcLyxdGv6x+4tcVXsnQ4rmf2/DXa/9HtNjmUQ/mmCad+ooEatARFaJK6Vdkjr4xNhvV+HlUlDTD9B4c69tbif8dT6Xp3QYNzsHVgrNcKFtXJM3gg/4ymO97ZtAY908wcKiQn9fcXX+jz53zqaSYdrJwKjNqOMYgEm8vBaMIG8hrYKhY/6TF+6z8UAkv+4o9/SsPMmoemXYAefM8jf/QNc9GfWnFOhJ8j5/0Vs77smpx7ELrcj8Y2ERAUu3FZ9b3YtQFQGhwzgcwCH0yY6Ig04ersEOmGJp00Ejl+V5kNeG7rc/nMk9QBTFVqqBQ7TQjgUTjZyqVNPhEkiaLyhdw1bpw5wl24C9vUK6xIxXMDLA2PEu/XWZ/mrklBUGohRNOHmXjPJFxaVT1h0X7LTkP0bEZHOeOx334toT6fR0rx8oP28gFglY7LohEaM2ERFHYGsA9BVF2/Yc58iou71Vh6MOOkaj+aRx9roaYYmjASvXleLXdYaDNmUUato4bLxkS8oC89wu9SBBViVrowjYAIMAhLs6brp7IjIBqiyNiIPvT2roNPimRpFGt1q+9zwWmWcB2k83KGDYoAybWNVtx9gSD+qOtM5YefZU9hnC81EUXymxeOU3PqNew1uN/Xjrl9jOrbQ7rU/A5MHHEIsCfY88LaRr2Z7GZZioCgemcv66XEBQRxyRqGkJicgCt8pQkOBZICaNOBvLNMSRylajx9R/EMpho+5Csx1c05XIOEV661JWxrPAcgtAzriNDWA+Vp+bk1heFvFhO30GtFqhDpxbGsBbDbp9mx5PksBDYUeYLRzER8V4mIHqjO5ziQLmYeQa4Qg0D4GcEF62dTBrD1n420bx9C+/OwmIPa1eEOwc2EFQh5JX+cVA5IgNeLcyX36mrWhuUEMYFZHz/zzRr27dmFCUZcsr0bP8D1pvLNJsZ09G2rYmFaq6UXqiCHa8Q6/MaqHrQwo18aluCLh/6avB9bPlN7F1v90xwCLHkFFla3Paakp+NwZfTz5zLDLCMrKX9Mu6Ke3QUR8ZlckvxKV7Hw+z2OntpCfVfDh0160gQAAHBS5qhEBsuyVQ3P8TVuBzFIrgsEFg/k2DoL3LEhz/nfhybKHB9BNJwvnlE/AtmkoEdBkFKSfsFjWzdINaN/HlNI5ZdKRhYrMRqj9WLJ36WJj26RF8dTTkVzlYJcRMPqiHz02VD2v/YIclDfISll5JtNDQHAcde4vGF4tkVdm/6SSmd1Ca58AzO/vUTRm0ewnwgAmBe528CpHpQyinIehsj6ab2uRk5bn6yCtoRzMYURXxLhyr6aYgr40LZ9srkuU9aTxdd0pzmXkvY/DVoxsmBM8cZjgG3HNhKlrYOVbOXtJsHGmPdnBdJTGfMXfjwRRPO04FVEDkFVmhlj5eZLZiG2Hc6GTDp0RD2rtrrMKZUfCViSoNvlEBNhmKL+wkaVWJyc2L/traE6bMd13hB007gsQ100pwpVkAzveKf8vEWiC1KRCRFpsouPP6w2/FdW5aKAPxUQQWuBm6uLojNxPcMLigD+mxJ2UtR4dJQ8u6Mg0H++Pazi2Dad/+2hB+x4IrohYhcC8Iyp3c1zt2Gddp5U+xR1tSZH1q1XDd7nuaC6eSsvtdCeShv13GLmR/AaeHlTWN+gzholHKKOYR4MdvCFBPOpIO/vWoqAlX9qf3K3k+Z5PPzE1E5IZbWNb9OEF5owiMBdIU9X12IrESH47szKCX5U7i/3SyDAkW11hFZtw5Ez7aKdzveU8K24QUa5GlAFnuq/abt790qkVfTmk5xD5XmHrjeV1fXK2ydiCcHcfHHSlYYWCpyNKVAcT0XVrx5l4/21HWs1KuboY+07aLtGO8LWFhTe90oKRh4dUAMUqmVlahO++B0/m/wkTFZYMTxE7/FxL3CICtMzxq45kzD56h4/F7jVFZEEotD5oh4m2nZWQAVvYgRJ5ZygY/yKGqa1/WFjRSuX75cnQcSUNgMJAOtYb7Xk/9x25mv+NMk3MsE/nfEW5++J60cHLBrpR/4DLoaznQNyeBxpnhbZ+2DKiLhqW1ennLder6LxlRZIRZbZtpknrP5D3emYhfUMIfuA3MJowL73IQOYxodNGBH4bQZsqZWsIVPCJVIKb3ydqK3ktqjdi2w6y9o4lMFFu/0yQH3L0wBC1C3Dp8u9AGnHx3Obq/WHKtrFqxu0+JAKD4ulBujifw99/VSZCadmz8qOpqcyYeys/z4t7aowfJ8E+diddt1UEW+mY3PSyanIXtptsiEja3umEClp3Ev+lYsthI/PfbZpkHHlwWQiuzZ0fH6vZbbC8ImoNQTraXNbppwag86MJKFttkOFerkl2iSyVScEYKmC+pAt988ddz0TbCIkJUXNhHjZPNXhw4xvXZ/jjLO61U2/WeV6vj75Jdf7yPl5/GOQJAja6p68DLmmVJn1BURa05iwhL3jc+9WeW/0ZJs6pQRoDtuwZMeGogotgxsusha14t4WXbYJupB7fxIKSKO2dS+TQMnKDgangIRiYZTAPeTMifmChFb4E0aG+wuJzbMhHzLAFvfW4218KXyFkiqCB9w+kj9I/CS941KoazHt/9FE+MG1tTJtXiyLwPEcIbQeaAnTG4CHMwNTUJS0ILtX2siBj/oI/WZffZz9IOEpt0E/55ZCB82x8SrLzu19Q6dL7Wp0FfVASHqw/FMBOgcbZ/0KpMIIlaQ2z6Ca0o8Bd6eiw5nRJ/KSpXhE0m2MFDJ9RzSUaAeMAnU0lImIh5RSlChLA2gLTk5OSv////9K/////0sAdJRiTXAChZRoFXSUUpSMA3Bvc5RNcAJ1jAloYXNfZ2F1c3OUSwCMBWdhdXNzlEcAAAAAAAAAAHVidWIu",
|
| 44 |
"dtype": "float32",
|
| 45 |
"_shape": [
|
| 46 |
6
|
|
|
|
| 55 |
"num_timesteps": 1000000,
|
| 56 |
"_total_timesteps": 1000000,
|
| 57 |
"_num_timesteps_at_start": 0,
|
| 58 |
+
"seed": 9,
|
| 59 |
"action_noise": null,
|
| 60 |
+
"start_time": 1672507717010268280,
|
| 61 |
"learning_rate": {
|
| 62 |
":type:": "<class 'function'>",
|
| 63 |
+
":serialized:": "gAWVhQIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwOX21ha2VfZnVuY3Rpb26Uk5QoaACMDV9idWlsdGluX3R5cGWUk5SMCENvZGVUeXBllIWUUpQoSwFLAEsASwFLAUsTQwSIAFMAlE6FlCmMAV+UhZSMUy9ob21lL21heGltaWxpYW4vdmVudi9saWIvcHl0aG9uMy44L3NpdGUtcGFja2FnZXMvc3RhYmxlX2Jhc2VsaW5lczMvY29tbW9uL3V0aWxzLnB5lIwEZnVuY5RLgEMCAAGUjAN2YWyUhZQpdJRSlH2UKIwLX19wYWNrYWdlX1+UjBhzdGFibGVfYmFzZWxpbmVzMy5jb21tb26UjAhfX25hbWVfX5SMHnN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi51dGlsc5SMCF9fZmlsZV9flGgMdU5OaACMEF9tYWtlX2VtcHR5X2NlbGyUk5QpUpSFlHSUUpSMHGNsb3VkcGlja2xlLmNsb3VkcGlja2xlX2Zhc3SUjBJfZnVuY3Rpb25fc2V0c3RhdGWUk5RoHn2UfZQoaBZoDYwMX19xdWFsbmFtZV9flIwZY29uc3RhbnRfZm4uPGxvY2Fscz4uZnVuY5SMD19fYW5ub3RhdGlvbnNfX5R9lIwOX19rd2RlZmF1bHRzX1+UTowMX19kZWZhdWx0c19flE6MCl9fbW9kdWxlX1+UaBeMB19fZG9jX1+UTowLX19jbG9zdXJlX1+UaACMCl9tYWtlX2NlbGyUk5RHP0Mn2ZsNUUKFlFKUhZSMF19jbG91ZHBpY2tsZV9zdWJtb2R1bGVzlF2UjAtfX2dsb2JhbHNfX5R9lHWGlIZSMC4="
|
| 64 |
},
|
| 65 |
+
"tensorboard_log": "runs/seals/Walker2d-v0__sac__5__1672507712/seals-Walker2d-v0",
|
| 66 |
"lr_schedule": {
|
| 67 |
":type:": "<class 'function'>",
|
| 68 |
+
":serialized:": "gAWVhQIAAAAAAACMF2Nsb3VkcGlja2xlLmNsb3VkcGlja2xllIwOX21ha2VfZnVuY3Rpb26Uk5QoaACMDV9idWlsdGluX3R5cGWUk5SMCENvZGVUeXBllIWUUpQoSwFLAEsASwFLAUsTQwSIAFMAlE6FlCmMAV+UhZSMUy9ob21lL21heGltaWxpYW4vdmVudi9saWIvcHl0aG9uMy44L3NpdGUtcGFja2FnZXMvc3RhYmxlX2Jhc2VsaW5lczMvY29tbW9uL3V0aWxzLnB5lIwEZnVuY5RLgEMCAAGUjAN2YWyUhZQpdJRSlH2UKIwLX19wYWNrYWdlX1+UjBhzdGFibGVfYmFzZWxpbmVzMy5jb21tb26UjAhfX25hbWVfX5SMHnN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi51dGlsc5SMCF9fZmlsZV9flGgMdU5OaACMEF9tYWtlX2VtcHR5X2NlbGyUk5QpUpSFlHSUUpSMHGNsb3VkcGlja2xlLmNsb3VkcGlja2xlX2Zhc3SUjBJfZnVuY3Rpb25fc2V0c3RhdGWUk5RoHn2UfZQoaBZoDYwMX19xdWFsbmFtZV9flIwZY29uc3RhbnRfZm4uPGxvY2Fscz4uZnVuY5SMD19fYW5ub3RhdGlvbnNfX5R9lIwOX19rd2RlZmF1bHRzX1+UTowMX19kZWZhdWx0c19flE6MCl9fbW9kdWxlX1+UaBeMB19fZG9jX1+UTowLX19jbG9zdXJlX1+UaACMCl9tYWtlX2NlbGyUk5RHP0Mn2ZsNUUKFlFKUhZSMF19jbG91ZHBpY2tsZV9zdWJtb2R1bGVzlF2UjAtfX2dsb2JhbHNfX5R9lHWGlIZSMC4="
|
| 69 |
},
|
| 70 |
"_last_obs": null,
|
| 71 |
"_last_episode_starts": {
|
|
|
|
| 74 |
},
|
| 75 |
"_last_original_obs": {
|
| 76 |
":type:": "<class 'numpy.ndarray'>",
|
| 77 |
+
":serialized:": "gAWVBQEAAAAAAACMEm51bXB5LmNvcmUubnVtZXJpY5SMC19mcm9tYnVmZmVylJOUKJaQAAAAAAAAAIvFGgl8eDpAO61oaHjV2z9YjVEyzPDtP1UOOyBwgwTAqnNbCbj+BcAMb9X5g37sP/o0SdjQDvG/jFXJcos2vD9eq2djgyzxvx5u28VWnQBA+R/GdA/SDEAAAAAAAAAkwJkzz8CDWBTAAAAAAAAAJMAAAAAAAAAkwAAAAAAAACTAQHYuWltXgj9SMKGnPyAhQJSMBW51bXB5lIwFZHR5cGWUk5SMAmY4lImIh5RSlChLA4wBPJROTk5K/////0r/////SwB0lGJLAUsShpSMAUOUdJRSlC4="
|
| 78 |
},
|
| 79 |
"_episode_num": 1000,
|
| 80 |
"use_sde": false,
|
|
|
|
| 82 |
"_current_progress_remaining": 0.0,
|
| 83 |
"ep_info_buffer": {
|
| 84 |
":type:": "<class 'collections.deque'>",
|
| 85 |
+
":serialized:": "gAWVgRAAAAAAAACMC2NvbGxlY3Rpb25zlIwFZGVxdWWUk5QpS2SGlFKUKH2UKIwBcpSMFW51bXB5LmNvcmUubXVsdGlhcnJheZSMBnNjYWxhcpSTlIwFbnVtcHmUjAVkdHlwZZSTlIwCZjiUiYiHlFKUKEsDjAE8lE5OTkr/////Sv////9LAHSUYkMInz4Cf3hprECUhpRSlIwBbJRN6AOMAXSUR0DgKOt/MnqndX2UKGgGaAloD0MIdAzIXt9vsUCUhpRSlGgVTegDaBZHQOAs2HEsJ6Z1fZQoaAZoCWgPQwjXFp6XZlCxQJSGlFKUaBVN6ANoFkdA4DCvfe+EiHV9lChoBmgJaA9DCHhGW5WcbrFAlIaUUpRoFU3oA2gWR0DgNIG/r0J4dX2UKGgGaAloD0MIYvVHGL4Lp0CUhpRSlGgVTegDaBZHQOA4N/cYZVJ1fZQoaAZoCWgPQwiEfqZeGyWxQJSGlFKUaBVN6ANoFkdA4Dv1H+6y0XV9lChoBmgJaA9DCCPZI9QcJqdAlIaUUpRoFU3oA2gWR0DgP7KhRIjGdX2UKGgGaAloD0MIF4BG6d6IsUCUhpRSlGgVTegDaBZHQOBDZQ+GGmF1fZQoaAZoCWgPQwgWURN9duGwQJSGlFKUaBVN6ANoFkdA4Echwzk6tHV9lChoBmgJaA9DCKT/5VpkkZxAlIaUUpRoFU3oA2gWR0DgStch1TzedX2UKGgGaAloD0MIhCo1e6BzSMCUhpRSlGgVTegDaBZHQOBOnFZPl+51fZQoaAZoCWgPQwjDZoALwm6xQJSGlFKUaBVN6ANoFkdA4FJbyj59E3V9lChoBmgJaA9DCBZu+UjClrFAlIaUUpRoFU3oA2gWR0DgViPVKf4AdX2UKGgGaAloD0MI7QvohQeLsUCUhpRSlGgVTegDaBZHQOBZ6WxnnMd1fZQoaAZoCWgPQwgfEVMicZCxQJSGlFKUaBVN6ANoFkdA4F2e/foA4nV9lChoBmgJaA9DCJd1/1iIL2vAlIaUUpRoFU3oA2gWR0DgYV8QT238dX2UKGgGaAloD0MIMEeP3w9usUCUhpRSlGgVTegDaBZHQOBlFglUp/h1fZQoaAZoCWgPQwg6IAn7Wm6xQJSGlFKUaBVN6ANoFkdA4GjTF7dBSnV9lChoBmgJaA9DCLX/AdaqKqhAlIaUUpRoFU3oA2gWR0DgbI7FxXGPdX2UKGgGaAloD0MIqn8QyRDon0CUhpRSlGgVTegDaBZHQOBwVdjy4F11fZQoaAZoCWgPQwghkbbxJ3SjQJSGlFKUaBVN6ANoFkdA4HQItorWiHV9lChoBmgJaA9DCPG3PUF+pbFAlIaUUpRoFU3oA2gWR0Dgd8GC8OCodX2UKGgGaAloD0MIthMlIenDrECUhpRSlGgVTegDaBZHQOB7ih8lXzV1fZQoaAZoCWgPQwg+XHLcUWKxQJSGlFKUaBVN6ANoFkdA4H876bF0gnV9lChoBmgJaA9DCHFZhc0AFaZAlIaUUpRoFU3oA2gWR0Dggutnh86WdX2UKGgGaAloD0MIs2Dij6IcpkCUhpRSlGgVTegDaBZHQOCHRDeXRgJ1fZQoaAZoCWgPQwi86CtIy2SnQJSGlFKUaBVN6ANoFkdA4Ir33CsOonV9lChoBmgJaA9DCB75g4GHiKhAlIaUUpRoFU3oA2gWR0Dgjr9V94NadX2UKGgGaAloD0MIK4arA1Bml0CUhpRSlGgVTegDaBZHQOCSfGy7f511fZQoaAZoCWgPQwjdIjDWF0ePQJSGlFKUaBVN6ANoFkdA4JY4Iq9XcXV9lChoBmgJaA9DCOQxA5VxEHFAlIaUUpRoFU3oA2gWR0Dgmf6paRp2dX2UKGgGaAloD0MID3uhgKUbsUCUhpRSlGgVTegDaBZHQOCd16yyD7J1fZQoaAZoCWgPQwiPxTapuGuxQJSGlFKUaBVN6ANoFkdA4KHB1anrIHV9lChoBmgJaA9DCOYHrvJUkK5AlIaUUpRoFU3oA2gWR0DgpYzx4IKMdX2UKGgGaAloD0MIoMa9+W3+kECUhpRSlGgVTegDaBZHQOCpS4Z0jkd1fZQoaAZoCWgPQwgV/aGZZ2yvQJSGlFKUaBVN6ANoFkdA4K0cRIBikXV9lChoBmgJaA9DCDC6vDmU6rBAlIaUUpRoFU3oA2gWR0DgsORtfG+9dX2UKGgGaAloD0MI1lJA2texr0CUhpRSlGgVTegDaBZHQOC0mGi1y/91fZQoaAZoCWgPQwg491ePQ0GhQJSGlFKUaBVN6ANoFkdA4Lhcb0Fr23V9lChoBmgJaA9DCBR15h4ut7BAlIaUUpRoFU3oA2gWR0DgvBaNNJvpdX2UKGgGaAloD0MI0TsVcFMBsUCUhpRSlGgVTegDaBZHQOC/3l07r9l1fZQoaAZoCWgPQwg6WWq91w2tQJSGlFKUaBVN6ANoFkdA4MOj0AT7EnV9lChoBmgJaA9DCEtbXOMbYqVAlIaUUpRoFU3oA2gWR0Dgx3azzErHdX2UKGgGaAloD0MIBHEeTqCyj0CUhpRSlGgVTegDaBZHQODLSo2CNCJ1fZQoaAZoCWgPQwh3gv3X0eKkQJSGlFKUaBVN6ANoFkdA4M8Jw4sEq3V9lChoBmgJaA9DCCFAho4dX6FAlIaUUpRoFU3oA2gWR0Dg0twh5gPVdX2UKGgGaAloD0MIxLKZQ153sECUhpRSlGgVTegDaBZHQODWmABmwq11fZQoaAZoCWgPQwg9mBQfxxmhQJSGlFKUaBVN6ANoFkdA4Npm+zUqhHV9lChoBmgJaA9DCFn5ZTDmSLBAlIaUUpRoFU3oA2gWR0Dg3kVzJ6ppdX2UKGgGaAloD0MIwa27efL6okCUhpRSlGgVTegDaBZHQODiCzY5DJF1fZQoaAZoCWgPQwiI83ACA8KwQJSGlFKUaBVN6ANoFkdA4OZxh3qzJXV9lChoBmgJaA9DCJi+1xAUI7BAlIaUUpRoFU3oA2gWR0Dg6jEla8pTdX2UKGgGaAloD0MINPRPcFUHsUCUhpRSlGgVTegDaBZHQODt+G3azu51fZQoaAZoCWgPQwiqQ26GC42xQJSGlFKUaBVN6ANoFkdA4PGrj+rEL3V9lChoBmgJaA9DCMLexJCkmrFAlIaUUpRoFU3oA2gWR0Dg9Vp5qM3qdX2UKGgGaAloD0MIyERKs000sUCUhpRSlGgVTegDaBZHQOD5B249X911fZQoaAZoCWgPQwjyBpj5LjWLQJSGlFKUaBVN6ANoFkdA4PzISpJf6XV9lChoBmgJaA9DCN9rCI7Lx1dAlIaUUpRoFU3oA2gWR0DhAHPsMRYjdX2UKGgGaAloD0MISQ9DqxM2jkCUhpRSlGgVTegDaBZHQOEEKXeP7vZ1fZQoaAZoCWgPQwjFPCtpTUWnQJSGlFKUaBVN6ANoFkdA4QfWxi5NGnV9lChoBmgJaA9DCH1cGyrGc5pAlIaUUpRoFU3oA2gWR0DhC5uoOQQudX2UKGgGaAloD0MITS8xlrnwsECUhpRSlGgVTegDaBZHQOEPVRIWgvl1fZQoaAZoCWgPQwieCyO9aHOuQJSGlFKUaBVN6ANoFkdA4RMLCm2srHV9lChoBmgJaA9DCI/FNqn0L7BAlIaUUpRoFU3oA2gWR0DhFruV1Oj7dX2UKGgGaAloD0MILuQR3FT8sECUhpRSlGgVTegDaBZHQOEaa4BHTZx1fZQoaAZoCWgPQwglQE0tu16IQJSGlFKUaBVN6ANoFkdA4R4T+hGpdnV9lChoBmgJaA9DCN4DdF8KmLBAlIaUUpRoFU3oA2gWR0DhIcaXnhbXdX2UKGgGaAloD0MI7MA5I0YXsUCUhpRSlGgVTegDaBZHQOElnbYoRZl1fZQoaAZoCWgPQwiGOUGbhMqwQJSGlFKUaBVN6ANoFkdA4SlAhSUC73V9lChoBmgJaA9DCCR8728Y47FAlIaUUpRoFU3oA2gWR0DhLO/OLzf8dX2UKGgGaAloD0MI1xh0QuhYTkCUhpRSlGgVTegDaBZHQOEwoczO5ax1fZQoaAZoCWgPQwjcRgN4y2ucQJSGlFKUaBVN6ANoFkdA4TRVW38XN3V9lChoBmgJaA9DCNO9TuqL73xAlIaUUpRoFU3oA2gWR0DhOAMD7IkrdX2UKGgGaAloD0MIhIJStFLrpkCUhpRSlGgVTegDaBZHQOE7rq1Z1V51fZQoaAZoCWgPQwj1nzU/xk+nQJSGlFKUaBVN6ANoFkdA4T9aBhH9WXV9lChoBmgJaA9DCLITXoIzHJ1AlIaUUpRoFU3oA2gWR0DhQ5kvxH5KdX2UKGgGaAloD0MIGcqJdgVqlECUhpRSlGgVTegDaBZHQOFHQgJgLJF1fZQoaAZoCWgPQwj9EYYBYzyqQJSGlFKUaBVN6ANoFkdA4UrIX+dbxHV9lChoBmgJaA9DCMQI4dG6SrFAlIaUUpRoFU3oA2gWR0DhTkUyULUkdX2UKGgGaAloD0MIiBOYThc7sUCUhpRSlGgVTegDaBZHQOFRzHo1UER1fZQoaAZoCWgPQwhM32sI+tmwQJSGlFKUaBVN6ANoFkdA4VV0tR3u/nV9lChoBmgJaA9DCHiAJy3QFLFAlIaUUpRoFU3oA2gWR0DhWPZZeRgadX2UKGgGaAloD0MIBi0kYOxBsUCUhpRSlGgVTegDaBZHQOFcdyjafz11fZQoaAZoCWgPQwgjvD0IHfmwQJSGlFKUaBVN6ANoFkdA4V/koy9EkXV9lChoBmgJaA9DCAySPq36vLBAlIaUUpRoFU3oA2gWR0DhY1IB6KLsdX2UKGgGaAloD0MI6IU7F/brsECUhpRSlGgVTegDaBZHQOFmyAIhQnB1fZQoaAZoCWgPQwjfwyXHqYuwQJSGlFKUaBVN6ANoFkdA4Wo0e9allHV9lChoBmgJaA9DCH8Xtmbz1LBAlIaUUpRoFU3oA2gWR0DhbaGHXVbzdX2UKGgGaAloD0MIrmcIx8z3sECUhpRSlGgVTegDaBZHQOFxDd1ZDAt1fZQoaAZoCWgPQwjGpL+Xkg6xQJSGlFKUaBVN6ANoFkdA4XR+5B1LanV9lChoBmgJaA9DCIKsp1b/p5hAlIaUUpRoFU3oA2gWR0Dhd+eI2wV1dX2UKGgGaAloD0MIKH6MuQvFmECUhpRSlGgVTegDaBZHQOF7XYKD0191fZQoaAZoCWgPQwjFOlW+Z+ZbQJSGlFKUaBVN6ANoFkdA4X78ByKekHV9lChoBmgJaA9DCKkWEcUEJJBAlIaUUpRoFU3oA2gWR0Dhgny0O3DvdX2UKGgGaAloD0MIbJIf8csLokCUhpRSlGgVTegDaBZHQOGF3swvg3t1fZQoaAZoCWgPQwhDA7Fshg6hQJSGlFKUaBVN6ANoFkdA4Yk/LxiG4HV9lChoBmgJaA9DCCCWzRxydqJAlIaUUpRoFU3oA2gWR0DhjOTSpBHDdX2UKGgGaAloD0MIhH8RNC73sUCUhpRSlGgVTegDaBZHQOGQXfKISDh1fZQoaAZoCWgPQwidY0D2GjaoQJSGlFKUaBVN6ANoFkdA4ZO2pRO1v3V9lChoBmgJaA9DCCuKV1kXXLBAlIaUUpRoFU3oA2gWR0DhlxZzSThYdWUu"
|
| 86 |
},
|
| 87 |
"ep_success_buffer": {
|
| 88 |
":type:": "<class 'collections.deque'>",
|
|
|
|
| 100 |
":type:": "<class 'abc.ABCMeta'>",
|
| 101 |
":serialized:": "gAWVNQAAAAAAAACMIHN0YWJsZV9iYXNlbGluZXMzLmNvbW1vbi5idWZmZXJzlIwMUmVwbGF5QnVmZmVylJOULg==",
|
| 102 |
"__module__": "stable_baselines3.common.buffers",
|
| 103 |
+
"__doc__": "\n Replay buffer used in off-policy algorithms like SAC/TD3.\n\n :param buffer_size: Max number of element in the buffer\n :param observation_space: Observation space\n :param action_space: Action space\n :param device: PyTorch device\n :param n_envs: Number of parallel environments\n :param optimize_memory_usage: Enable a memory efficient variant\n of the replay buffer which reduces by almost a factor two the memory used,\n at a cost of more complexity.\n See https://github.com/DLR-RM/stable-baselines3/issues/37#issuecomment-637501195\n and https://github.com/DLR-RM/stable-baselines3/pull/28#issuecomment-637559274\n Cannot be used in combination with handle_timeout_termination.\n :param handle_timeout_termination: Handle timeout termination (due to timelimit)\n separately and treat the task as infinite horizon task.\n https://github.com/DLR-RM/stable-baselines3/issues/284\n ",
|
| 104 |
+
"__init__": "<function ReplayBuffer.__init__ at 0x7ff1e3c5c0d0>",
|
| 105 |
+
"add": "<function ReplayBuffer.add at 0x7ff1e3c5c160>",
|
| 106 |
+
"sample": "<function ReplayBuffer.sample at 0x7ff1e3c5c1f0>",
|
| 107 |
+
"_get_samples": "<function ReplayBuffer._get_samples at 0x7ff1e3c5c280>",
|
| 108 |
"__abstractmethods__": "frozenset()",
|
| 109 |
+
"_abc_impl": "<_abc_data object at 0x7ff1e3cdd300>"
|
| 110 |
},
|
| 111 |
"replay_buffer_kwargs": {},
|
| 112 |
"train_freq": {
|
|
|
|
| 116 |
"use_sde_at_warmup": false,
|
| 117 |
"target_entropy": -6.0,
|
| 118 |
"ent_coef": "auto",
|
| 119 |
+
"target_update_interval": 1,
|
| 120 |
+
"batch_norm_stats": [],
|
| 121 |
+
"batch_norm_stats_target": []
|
| 122 |
}
|
sac-seals-Walker2d-v0/ent_coef_optimizer.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b514180cdb79442072ef81cdef6d2e1f5ec3a856d35b7f0df418e70f75ff997e
|
| 3 |
+
size 1507
|
sac-seals-Walker2d-v0/policy.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 2625861
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:235465ba14b0776771df7bf881136a26d6906703aa8829e262aaaa44b9674fa4
|
| 3 |
size 2625861
|
sac-seals-Walker2d-v0/pytorch_variables.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 747
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:87a3ecd2725bf728bd2ee5403c415377146315f773f5bf623e0180137fdba22b
|
| 3 |
size 747
|
sac-seals-Walker2d-v0/system_info.txt
CHANGED
|
@@ -1,6 +1,6 @@
|
|
| 1 |
-
OS: Linux-5.4.0-
|
| 2 |
Python: 3.8.10
|
| 3 |
-
Stable-Baselines3: 1.6.
|
| 4 |
PyTorch: 1.11.0+cu102
|
| 5 |
GPU Enabled: False
|
| 6 |
Numpy: 1.22.3
|
|
|
|
| 1 |
+
OS: Linux-5.4.0-125-generic-x86_64-with-glibc2.29 #141-Ubuntu SMP Wed Aug 10 13:42:03 UTC 2022
|
| 2 |
Python: 3.8.10
|
| 3 |
+
Stable-Baselines3: 1.6.2
|
| 4 |
PyTorch: 1.11.0+cu102
|
| 5 |
GPU Enabled: False
|
| 6 |
Numpy: 1.22.3
|
train_eval_metrics.zip
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:dcabe724964590e4adfb5c96889aba152036ebe86b434e0a6ca315f54aaa9079
|
| 3 |
+
size 34786
|