Commit
·
9b49d79
1
Parent(s):
5101655
Upload folder using huggingface_hub
Browse files- .summary/0/events.out.tfevents.1694521277.rhmmedcatt-ProLiant-ML350-Gen10 +3 -0
- README.md +56 -0
- checkpoint_p0/best_000000224_917504_reward_0.815.pth +3 -0
- checkpoint_p0/checkpoint_000000246_1007616.pth +3 -0
- config.json +142 -0
- git.diff +0 -0
- replay.mp4 +0 -0
- sf_log.txt +181 -0
.summary/0/events.out.tfevents.1694521277.rhmmedcatt-ProLiant-ML350-Gen10
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7de6ffd23dcba7e216952ab00f02a8d679000dec55e57c28732695c82413d334
|
| 3 |
+
size 84366
|
README.md
ADDED
|
@@ -0,0 +1,56 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
library_name: sample-factory
|
| 3 |
+
tags:
|
| 4 |
+
- deep-reinforcement-learning
|
| 5 |
+
- reinforcement-learning
|
| 6 |
+
- sample-factory
|
| 7 |
+
model-index:
|
| 8 |
+
- name: APPO
|
| 9 |
+
results:
|
| 10 |
+
- task:
|
| 11 |
+
type: reinforcement-learning
|
| 12 |
+
name: reinforcement-learning
|
| 13 |
+
dataset:
|
| 14 |
+
name: doom_basic
|
| 15 |
+
type: doom_basic
|
| 16 |
+
metrics:
|
| 17 |
+
- type: mean_reward
|
| 18 |
+
value: 0.77 +/- 0.13
|
| 19 |
+
name: mean_reward
|
| 20 |
+
verified: false
|
| 21 |
+
---
|
| 22 |
+
|
| 23 |
+
A(n) **APPO** model trained on the **doom_basic** environment.
|
| 24 |
+
|
| 25 |
+
This model was trained using Sample-Factory 2.0: https://github.com/alex-petrenko/sample-factory.
|
| 26 |
+
Documentation for how to use Sample-Factory can be found at https://www.samplefactory.dev/
|
| 27 |
+
|
| 28 |
+
|
| 29 |
+
## Downloading the model
|
| 30 |
+
|
| 31 |
+
After installing Sample-Factory, download the model with:
|
| 32 |
+
```
|
| 33 |
+
python -m sample_factory.huggingface.load_from_hub -r MattStammers/vizdoom_basic
|
| 34 |
+
```
|
| 35 |
+
|
| 36 |
+
|
| 37 |
+
## Using the model
|
| 38 |
+
|
| 39 |
+
To run the model after download, use the `enjoy` script corresponding to this environment:
|
| 40 |
+
```
|
| 41 |
+
python -m <path.to.enjoy.module> --algo=APPO --env=doom_basic --train_dir=./train_dir --experiment=vizdoom_basic
|
| 42 |
+
```
|
| 43 |
+
|
| 44 |
+
|
| 45 |
+
You can also upload models to the Hugging Face Hub using the same script with the `--push_to_hub` flag.
|
| 46 |
+
See https://www.samplefactory.dev/10-huggingface/huggingface/ for more details
|
| 47 |
+
|
| 48 |
+
## Training with this model
|
| 49 |
+
|
| 50 |
+
To continue training with this model, use the `train` script corresponding to this environment:
|
| 51 |
+
```
|
| 52 |
+
python -m <path.to.train.module> --algo=APPO --env=doom_basic --train_dir=./train_dir --experiment=vizdoom_basic --restart_behavior=resume --train_for_env_steps=10000000000
|
| 53 |
+
```
|
| 54 |
+
|
| 55 |
+
Note, you may have to adjust `--train_for_env_steps` to a suitably high number as the experiment will resume at the number of steps it concluded at.
|
| 56 |
+
|
checkpoint_p0/best_000000224_917504_reward_0.815.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2c755a883e08b46f81b73e08a539d7cf4f6181e556acceef80d73a4e936f84f9
|
| 3 |
+
size 34922470
|
checkpoint_p0/checkpoint_000000246_1007616.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8d6187e7885a9977ff2ca58d15928f94edc65c38f207bc232cd061c93bdc675b
|
| 3 |
+
size 34922884
|
config.json
ADDED
|
@@ -0,0 +1,142 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"help": false,
|
| 3 |
+
"algo": "APPO",
|
| 4 |
+
"env": "doom_basic",
|
| 5 |
+
"experiment": "default_experiment",
|
| 6 |
+
"train_dir": "/home/cogstack/Documents/optuna/environments/sample_factory/train_dir",
|
| 7 |
+
"restart_behavior": "restart",
|
| 8 |
+
"device": "gpu",
|
| 9 |
+
"seed": null,
|
| 10 |
+
"num_policies": 1,
|
| 11 |
+
"async_rl": true,
|
| 12 |
+
"serial_mode": false,
|
| 13 |
+
"batched_sampling": false,
|
| 14 |
+
"num_batches_to_accumulate": 2,
|
| 15 |
+
"worker_num_splits": 2,
|
| 16 |
+
"policy_workers_per_policy": 1,
|
| 17 |
+
"max_policy_lag": 1000,
|
| 18 |
+
"num_workers": 8,
|
| 19 |
+
"num_envs_per_worker": 4,
|
| 20 |
+
"batch_size": 1024,
|
| 21 |
+
"num_batches_per_epoch": 1,
|
| 22 |
+
"num_epochs": 1,
|
| 23 |
+
"rollout": 32,
|
| 24 |
+
"recurrence": 32,
|
| 25 |
+
"shuffle_minibatches": false,
|
| 26 |
+
"gamma": 0.99,
|
| 27 |
+
"reward_scale": 1.0,
|
| 28 |
+
"reward_clip": 1000.0,
|
| 29 |
+
"value_bootstrap": false,
|
| 30 |
+
"normalize_returns": true,
|
| 31 |
+
"exploration_loss_coeff": 0.001,
|
| 32 |
+
"value_loss_coeff": 0.5,
|
| 33 |
+
"kl_loss_coeff": 0.0,
|
| 34 |
+
"exploration_loss": "symmetric_kl",
|
| 35 |
+
"gae_lambda": 0.95,
|
| 36 |
+
"ppo_clip_ratio": 0.1,
|
| 37 |
+
"ppo_clip_value": 0.2,
|
| 38 |
+
"with_vtrace": false,
|
| 39 |
+
"vtrace_rho": 1.0,
|
| 40 |
+
"vtrace_c": 1.0,
|
| 41 |
+
"optimizer": "adam",
|
| 42 |
+
"adam_eps": 1e-06,
|
| 43 |
+
"adam_beta1": 0.9,
|
| 44 |
+
"adam_beta2": 0.999,
|
| 45 |
+
"max_grad_norm": 4.0,
|
| 46 |
+
"learning_rate": 0.0001,
|
| 47 |
+
"lr_schedule": "constant",
|
| 48 |
+
"lr_schedule_kl_threshold": 0.008,
|
| 49 |
+
"lr_adaptive_min": 1e-06,
|
| 50 |
+
"lr_adaptive_max": 0.01,
|
| 51 |
+
"obs_subtract_mean": 0.0,
|
| 52 |
+
"obs_scale": 255.0,
|
| 53 |
+
"normalize_input": true,
|
| 54 |
+
"normalize_input_keys": null,
|
| 55 |
+
"decorrelate_experience_max_seconds": 0,
|
| 56 |
+
"decorrelate_envs_on_one_worker": true,
|
| 57 |
+
"actor_worker_gpus": [],
|
| 58 |
+
"set_workers_cpu_affinity": true,
|
| 59 |
+
"force_envs_single_thread": false,
|
| 60 |
+
"default_niceness": 0,
|
| 61 |
+
"log_to_file": true,
|
| 62 |
+
"experiment_summaries_interval": 10,
|
| 63 |
+
"flush_summaries_interval": 30,
|
| 64 |
+
"stats_avg": 100,
|
| 65 |
+
"summaries_use_frameskip": true,
|
| 66 |
+
"heartbeat_interval": 20,
|
| 67 |
+
"heartbeat_reporting_interval": 600,
|
| 68 |
+
"train_for_env_steps": 1000000,
|
| 69 |
+
"train_for_seconds": 10000000000,
|
| 70 |
+
"save_every_sec": 120,
|
| 71 |
+
"keep_checkpoints": 2,
|
| 72 |
+
"load_checkpoint_kind": "latest",
|
| 73 |
+
"save_milestones_sec": -1,
|
| 74 |
+
"save_best_every_sec": 5,
|
| 75 |
+
"save_best_metric": "reward",
|
| 76 |
+
"save_best_after": 100000,
|
| 77 |
+
"benchmark": false,
|
| 78 |
+
"encoder_mlp_layers": [
|
| 79 |
+
512,
|
| 80 |
+
512
|
| 81 |
+
],
|
| 82 |
+
"encoder_conv_architecture": "convnet_simple",
|
| 83 |
+
"encoder_conv_mlp_layers": [
|
| 84 |
+
512
|
| 85 |
+
],
|
| 86 |
+
"use_rnn": true,
|
| 87 |
+
"rnn_size": 512,
|
| 88 |
+
"rnn_type": "gru",
|
| 89 |
+
"rnn_num_layers": 1,
|
| 90 |
+
"decoder_mlp_layers": [],
|
| 91 |
+
"nonlinearity": "elu",
|
| 92 |
+
"policy_initialization": "orthogonal",
|
| 93 |
+
"policy_init_gain": 1.0,
|
| 94 |
+
"actor_critic_share_weights": true,
|
| 95 |
+
"adaptive_stddev": true,
|
| 96 |
+
"continuous_tanh_scale": 0.0,
|
| 97 |
+
"initial_stddev": 1.0,
|
| 98 |
+
"use_env_info_cache": false,
|
| 99 |
+
"env_gpu_actions": false,
|
| 100 |
+
"env_gpu_observations": true,
|
| 101 |
+
"env_frameskip": 4,
|
| 102 |
+
"env_framestack": 1,
|
| 103 |
+
"pixel_format": "CHW",
|
| 104 |
+
"use_record_episode_statistics": false,
|
| 105 |
+
"with_wandb": false,
|
| 106 |
+
"wandb_user": null,
|
| 107 |
+
"wandb_project": "sample_factory",
|
| 108 |
+
"wandb_group": null,
|
| 109 |
+
"wandb_job_type": "SF",
|
| 110 |
+
"wandb_tags": [],
|
| 111 |
+
"with_pbt": false,
|
| 112 |
+
"pbt_mix_policies_in_one_env": true,
|
| 113 |
+
"pbt_period_env_steps": 5000000,
|
| 114 |
+
"pbt_start_mutation": 20000000,
|
| 115 |
+
"pbt_replace_fraction": 0.3,
|
| 116 |
+
"pbt_mutation_rate": 0.15,
|
| 117 |
+
"pbt_replace_reward_gap": 0.1,
|
| 118 |
+
"pbt_replace_reward_gap_absolute": 1e-06,
|
| 119 |
+
"pbt_optimize_gamma": false,
|
| 120 |
+
"pbt_target_objective": "true_objective",
|
| 121 |
+
"pbt_perturb_min": 1.1,
|
| 122 |
+
"pbt_perturb_max": 1.5,
|
| 123 |
+
"num_agents": -1,
|
| 124 |
+
"num_humans": 0,
|
| 125 |
+
"num_bots": -1,
|
| 126 |
+
"start_bot_difficulty": null,
|
| 127 |
+
"timelimit": null,
|
| 128 |
+
"res_w": 128,
|
| 129 |
+
"res_h": 72,
|
| 130 |
+
"wide_aspect_ratio": false,
|
| 131 |
+
"eval_env_frameskip": 1,
|
| 132 |
+
"fps": 35,
|
| 133 |
+
"command_line": "--env=doom_health_gathering_supreme --num_workers=8 --num_envs_per_worker=4 --train_for_env_steps=4000000",
|
| 134 |
+
"cli_args": {
|
| 135 |
+
"env": "doom_health_gathering_supreme",
|
| 136 |
+
"num_workers": 8,
|
| 137 |
+
"num_envs_per_worker": 4,
|
| 138 |
+
"train_for_env_steps": 4000000
|
| 139 |
+
},
|
| 140 |
+
"git_hash": "b12d96985caa7a7552d0840afdd14065f56f9f9a",
|
| 141 |
+
"git_repo_name": "https://github.com/MattStammers/optuna.git"
|
| 142 |
+
}
|
git.diff
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
replay.mp4
ADDED
|
Binary file (370 kB). View file
|
|
|
sf_log.txt
ADDED
|
@@ -0,0 +1,181 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
[2023-09-12 13:21:22,562][09743] Using GPUs [0] for process 0 (actually maps to GPUs [0])
|
| 2 |
+
[2023-09-12 13:21:22,562][09743] Set environment var CUDA_VISIBLE_DEVICES to '0' (GPU indices [0]) for learning process 0
|
| 3 |
+
[2023-09-12 13:21:22,598][09743] Num visible devices: 1
|
| 4 |
+
[2023-09-12 13:21:22,637][09743] Starting seed is not provided
|
| 5 |
+
[2023-09-12 13:21:22,638][09743] Using GPUs [0] for process 0 (actually maps to GPUs [0])
|
| 6 |
+
[2023-09-12 13:21:22,638][09743] Initializing actor-critic model on device cuda:0
|
| 7 |
+
[2023-09-12 13:21:22,638][09743] RunningMeanStd input shape: (3, 72, 128)
|
| 8 |
+
[2023-09-12 13:21:22,639][09743] RunningMeanStd input shape: (1,)
|
| 9 |
+
[2023-09-12 13:21:22,659][09743] ConvEncoder: input_channels=3
|
| 10 |
+
[2023-09-12 13:21:22,911][09743] Conv encoder output size: 512
|
| 11 |
+
[2023-09-12 13:21:22,911][09743] Policy head output size: 512
|
| 12 |
+
[2023-09-12 13:21:22,935][09743] Created Actor Critic model with architecture:
|
| 13 |
+
[2023-09-12 13:21:22,935][09743] ActorCriticSharedWeights(
|
| 14 |
+
(obs_normalizer): ObservationNormalizer(
|
| 15 |
+
(running_mean_std): RunningMeanStdDictInPlace(
|
| 16 |
+
(running_mean_std): ModuleDict(
|
| 17 |
+
(obs): RunningMeanStdInPlace()
|
| 18 |
+
)
|
| 19 |
+
)
|
| 20 |
+
)
|
| 21 |
+
(returns_normalizer): RecursiveScriptModule(original_name=RunningMeanStdInPlace)
|
| 22 |
+
(encoder): VizdoomEncoder(
|
| 23 |
+
(basic_encoder): ConvEncoder(
|
| 24 |
+
(enc): RecursiveScriptModule(
|
| 25 |
+
original_name=ConvEncoderImpl
|
| 26 |
+
(conv_head): RecursiveScriptModule(
|
| 27 |
+
original_name=Sequential
|
| 28 |
+
(0): RecursiveScriptModule(original_name=Conv2d)
|
| 29 |
+
(1): RecursiveScriptModule(original_name=ELU)
|
| 30 |
+
(2): RecursiveScriptModule(original_name=Conv2d)
|
| 31 |
+
(3): RecursiveScriptModule(original_name=ELU)
|
| 32 |
+
(4): RecursiveScriptModule(original_name=Conv2d)
|
| 33 |
+
(5): RecursiveScriptModule(original_name=ELU)
|
| 34 |
+
)
|
| 35 |
+
(mlp_layers): RecursiveScriptModule(
|
| 36 |
+
original_name=Sequential
|
| 37 |
+
(0): RecursiveScriptModule(original_name=Linear)
|
| 38 |
+
(1): RecursiveScriptModule(original_name=ELU)
|
| 39 |
+
)
|
| 40 |
+
)
|
| 41 |
+
)
|
| 42 |
+
)
|
| 43 |
+
(core): ModelCoreRNN(
|
| 44 |
+
(core): GRU(512, 512)
|
| 45 |
+
)
|
| 46 |
+
(decoder): MlpDecoder(
|
| 47 |
+
(mlp): Identity()
|
| 48 |
+
)
|
| 49 |
+
(critic_linear): Linear(in_features=512, out_features=1, bias=True)
|
| 50 |
+
(action_parameterization): ActionParameterizationDefault(
|
| 51 |
+
(distribution_linear): Linear(in_features=512, out_features=4, bias=True)
|
| 52 |
+
)
|
| 53 |
+
)
|
| 54 |
+
[2023-09-12 13:21:24,096][09743] Using optimizer <class 'torch.optim.adam.Adam'>
|
| 55 |
+
[2023-09-12 13:21:24,096][09743] No checkpoints found
|
| 56 |
+
[2023-09-12 13:21:24,097][09743] Did not load from checkpoint, starting from scratch!
|
| 57 |
+
[2023-09-12 13:21:24,097][09743] Initialized policy 0 weights for model version 0
|
| 58 |
+
[2023-09-12 13:21:24,098][09743] LearnerWorker_p0 finished initialization!
|
| 59 |
+
[2023-09-12 13:21:24,098][09743] Using GPUs [0] for process 0 (actually maps to GPUs [0])
|
| 60 |
+
[2023-09-12 13:21:24,463][09929] Worker 1 uses CPU cores [4, 5, 6, 7]
|
| 61 |
+
[2023-09-12 13:21:24,475][09931] Worker 2 uses CPU cores [8, 9, 10, 11]
|
| 62 |
+
[2023-09-12 13:21:24,499][09964] Worker 5 uses CPU cores [20, 21, 22, 23]
|
| 63 |
+
[2023-09-12 13:21:24,535][09967] Worker 4 uses CPU cores [16, 17, 18, 19]
|
| 64 |
+
[2023-09-12 13:21:24,545][09928] Using GPUs [0] for process 0 (actually maps to GPUs [0])
|
| 65 |
+
[2023-09-12 13:21:24,545][09928] Set environment var CUDA_VISIBLE_DEVICES to '0' (GPU indices [0]) for inference process 0
|
| 66 |
+
[2023-09-12 13:21:24,566][09928] Num visible devices: 1
|
| 67 |
+
[2023-09-12 13:21:24,567][09932] Worker 3 uses CPU cores [12, 13, 14, 15]
|
| 68 |
+
[2023-09-12 13:21:24,645][09965] Worker 7 uses CPU cores [28, 29, 30, 31]
|
| 69 |
+
[2023-09-12 13:21:24,665][09968] Worker 6 uses CPU cores [24, 25, 26, 27]
|
| 70 |
+
[2023-09-12 13:21:24,689][09930] Worker 0 uses CPU cores [0, 1, 2, 3]
|
| 71 |
+
[2023-09-12 13:21:25,314][09928] RunningMeanStd input shape: (3, 72, 128)
|
| 72 |
+
[2023-09-12 13:21:25,315][09928] RunningMeanStd input shape: (1,)
|
| 73 |
+
[2023-09-12 13:21:25,326][09928] ConvEncoder: input_channels=3
|
| 74 |
+
[2023-09-12 13:21:25,447][09928] Conv encoder output size: 512
|
| 75 |
+
[2023-09-12 13:21:25,448][09928] Policy head output size: 512
|
| 76 |
+
[2023-09-12 13:21:25,839][09964] Doom resolution: 160x120, resize resolution: (128, 72)
|
| 77 |
+
[2023-09-12 13:21:25,839][09968] Doom resolution: 160x120, resize resolution: (128, 72)
|
| 78 |
+
[2023-09-12 13:21:25,839][09967] Doom resolution: 160x120, resize resolution: (128, 72)
|
| 79 |
+
[2023-09-12 13:21:25,840][09965] Doom resolution: 160x120, resize resolution: (128, 72)
|
| 80 |
+
[2023-09-12 13:21:25,840][09931] Doom resolution: 160x120, resize resolution: (128, 72)
|
| 81 |
+
[2023-09-12 13:21:25,848][09932] Doom resolution: 160x120, resize resolution: (128, 72)
|
| 82 |
+
[2023-09-12 13:21:25,851][09930] Doom resolution: 160x120, resize resolution: (128, 72)
|
| 83 |
+
[2023-09-12 13:21:25,852][09929] Doom resolution: 160x120, resize resolution: (128, 72)
|
| 84 |
+
[2023-09-12 13:21:26,147][09967] Decorrelating experience for 0 frames...
|
| 85 |
+
[2023-09-12 13:21:26,147][09965] Decorrelating experience for 0 frames...
|
| 86 |
+
[2023-09-12 13:21:26,215][09964] Decorrelating experience for 0 frames...
|
| 87 |
+
[2023-09-12 13:21:26,239][09929] Decorrelating experience for 0 frames...
|
| 88 |
+
[2023-09-12 13:21:26,258][09968] Decorrelating experience for 0 frames...
|
| 89 |
+
[2023-09-12 13:21:26,273][09931] Decorrelating experience for 0 frames...
|
| 90 |
+
[2023-09-12 13:21:26,286][09930] Decorrelating experience for 0 frames...
|
| 91 |
+
[2023-09-12 13:21:26,418][09967] Decorrelating experience for 32 frames...
|
| 92 |
+
[2023-09-12 13:21:26,493][09964] Decorrelating experience for 32 frames...
|
| 93 |
+
[2023-09-12 13:21:26,522][09965] Decorrelating experience for 32 frames...
|
| 94 |
+
[2023-09-12 13:21:26,525][09929] Decorrelating experience for 32 frames...
|
| 95 |
+
[2023-09-12 13:21:26,551][09932] Decorrelating experience for 0 frames...
|
| 96 |
+
[2023-09-12 13:21:26,556][09931] Decorrelating experience for 32 frames...
|
| 97 |
+
[2023-09-12 13:21:26,568][09930] Decorrelating experience for 32 frames...
|
| 98 |
+
[2023-09-12 13:21:26,775][09967] Decorrelating experience for 64 frames...
|
| 99 |
+
[2023-09-12 13:21:26,821][09932] Decorrelating experience for 32 frames...
|
| 100 |
+
[2023-09-12 13:21:26,852][09964] Decorrelating experience for 64 frames...
|
| 101 |
+
[2023-09-12 13:21:26,919][09931] Decorrelating experience for 64 frames...
|
| 102 |
+
[2023-09-12 13:21:26,929][09930] Decorrelating experience for 64 frames...
|
| 103 |
+
[2023-09-12 13:21:27,103][09929] Decorrelating experience for 64 frames...
|
| 104 |
+
[2023-09-12 13:21:27,160][09968] Decorrelating experience for 32 frames...
|
| 105 |
+
[2023-09-12 13:21:27,164][09965] Decorrelating experience for 64 frames...
|
| 106 |
+
[2023-09-12 13:21:27,195][09932] Decorrelating experience for 64 frames...
|
| 107 |
+
[2023-09-12 13:21:27,201][09964] Decorrelating experience for 96 frames...
|
| 108 |
+
[2023-09-12 13:21:27,330][09931] Decorrelating experience for 96 frames...
|
| 109 |
+
[2023-09-12 13:21:27,451][09929] Decorrelating experience for 96 frames...
|
| 110 |
+
[2023-09-12 13:21:27,465][09967] Decorrelating experience for 96 frames...
|
| 111 |
+
[2023-09-12 13:21:27,498][09965] Decorrelating experience for 96 frames...
|
| 112 |
+
[2023-09-12 13:21:27,507][09930] Decorrelating experience for 96 frames...
|
| 113 |
+
[2023-09-12 13:21:27,588][09968] Decorrelating experience for 64 frames...
|
| 114 |
+
[2023-09-12 13:21:27,634][09932] Decorrelating experience for 96 frames...
|
| 115 |
+
[2023-09-12 13:21:27,903][09968] Decorrelating experience for 96 frames...
|
| 116 |
+
[2023-09-12 13:21:28,649][09743] Signal inference workers to stop experience collection...
|
| 117 |
+
[2023-09-12 13:21:28,653][09928] InferenceWorker_p0-w0: stopping experience collection
|
| 118 |
+
[2023-09-12 13:21:32,650][09743] Signal inference workers to resume experience collection...
|
| 119 |
+
[2023-09-12 13:21:32,651][09928] InferenceWorker_p0-w0: resuming experience collection
|
| 120 |
+
[2023-09-12 13:21:35,991][09928] Updated weights for policy 0, policy_version 10 (0.0392)
|
| 121 |
+
[2023-09-12 13:21:39,213][09928] Updated weights for policy 0, policy_version 20 (0.0009)
|
| 122 |
+
[2023-09-12 13:21:42,363][09928] Updated weights for policy 0, policy_version 30 (0.0009)
|
| 123 |
+
[2023-09-12 13:21:42,527][09743] Saving new best policy, reward=-1.655!
|
| 124 |
+
[2023-09-12 13:21:45,525][09928] Updated weights for policy 0, policy_version 40 (0.0009)
|
| 125 |
+
[2023-09-12 13:21:47,532][09743] Saving new best policy, reward=-0.936!
|
| 126 |
+
[2023-09-12 13:21:48,749][09928] Updated weights for policy 0, policy_version 50 (0.0009)
|
| 127 |
+
[2023-09-12 13:21:51,927][09928] Updated weights for policy 0, policy_version 60 (0.0008)
|
| 128 |
+
[2023-09-12 13:21:52,564][09743] Saving new best policy, reward=0.078!
|
| 129 |
+
[2023-09-12 13:21:55,197][09928] Updated weights for policy 0, policy_version 70 (0.0009)
|
| 130 |
+
[2023-09-12 13:21:57,529][09743] Saving new best policy, reward=0.521!
|
| 131 |
+
[2023-09-12 13:21:58,483][09928] Updated weights for policy 0, policy_version 80 (0.0010)
|
| 132 |
+
[2023-09-12 13:22:01,740][09928] Updated weights for policy 0, policy_version 90 (0.0008)
|
| 133 |
+
[2023-09-12 13:22:02,527][09743] Saving new best policy, reward=0.599!
|
| 134 |
+
[2023-09-12 13:22:05,213][09928] Updated weights for policy 0, policy_version 100 (0.0012)
|
| 135 |
+
[2023-09-12 13:22:07,589][09743] Saving new best policy, reward=0.680!
|
| 136 |
+
[2023-09-12 13:22:08,636][09928] Updated weights for policy 0, policy_version 110 (0.0017)
|
| 137 |
+
[2023-09-12 13:22:11,996][09928] Updated weights for policy 0, policy_version 120 (0.0009)
|
| 138 |
+
[2023-09-12 13:22:12,526][09743] Saving new best policy, reward=0.735!
|
| 139 |
+
[2023-09-12 13:22:15,370][09928] Updated weights for policy 0, policy_version 130 (0.0008)
|
| 140 |
+
[2023-09-12 13:22:17,527][09743] Saving new best policy, reward=0.755!
|
| 141 |
+
[2023-09-12 13:22:18,771][09928] Updated weights for policy 0, policy_version 140 (0.0009)
|
| 142 |
+
[2023-09-12 13:22:22,142][09928] Updated weights for policy 0, policy_version 150 (0.0009)
|
| 143 |
+
[2023-09-12 13:22:22,526][09743] Saving new best policy, reward=0.781!
|
| 144 |
+
[2023-09-12 13:22:25,580][09928] Updated weights for policy 0, policy_version 160 (0.0008)
|
| 145 |
+
[2023-09-12 13:22:27,573][09743] Saving new best policy, reward=0.791!
|
| 146 |
+
[2023-09-12 13:22:28,937][09928] Updated weights for policy 0, policy_version 170 (0.0009)
|
| 147 |
+
[2023-09-12 13:22:32,402][09928] Updated weights for policy 0, policy_version 180 (0.0008)
|
| 148 |
+
[2023-09-12 13:22:32,526][09743] Saving new best policy, reward=0.794!
|
| 149 |
+
[2023-09-12 13:22:35,725][09928] Updated weights for policy 0, policy_version 190 (0.0009)
|
| 150 |
+
[2023-09-12 13:22:37,529][09743] Saving new best policy, reward=0.806!
|
| 151 |
+
[2023-09-12 13:22:39,128][09928] Updated weights for policy 0, policy_version 200 (0.0010)
|
| 152 |
+
[2023-09-12 13:22:42,471][09928] Updated weights for policy 0, policy_version 210 (0.0009)
|
| 153 |
+
[2023-09-12 13:22:45,922][09928] Updated weights for policy 0, policy_version 220 (0.0009)
|
| 154 |
+
[2023-09-12 13:22:47,533][09743] Saving new best policy, reward=0.815!
|
| 155 |
+
[2023-09-12 13:22:49,374][09928] Updated weights for policy 0, policy_version 230 (0.0009)
|
| 156 |
+
[2023-09-12 13:22:52,742][09928] Updated weights for policy 0, policy_version 240 (0.0009)
|
| 157 |
+
[2023-09-12 13:22:54,862][09743] Stopping Batcher_0...
|
| 158 |
+
[2023-09-12 13:22:54,862][09743] Saving /home/cogstack/Documents/optuna/environments/sample_factory/train_dir/default_experiment/checkpoint_p0/checkpoint_000000246_1007616.pth...
|
| 159 |
+
[2023-09-12 13:22:54,863][09743] Loop batcher_evt_loop terminating...
|
| 160 |
+
[2023-09-12 13:22:54,876][09965] Stopping RolloutWorker_w7...
|
| 161 |
+
[2023-09-12 13:22:54,877][09965] Loop rollout_proc7_evt_loop terminating...
|
| 162 |
+
[2023-09-12 13:22:54,877][09932] Stopping RolloutWorker_w3...
|
| 163 |
+
[2023-09-12 13:22:54,877][09930] Stopping RolloutWorker_w0...
|
| 164 |
+
[2023-09-12 13:22:54,877][09968] Stopping RolloutWorker_w6...
|
| 165 |
+
[2023-09-12 13:22:54,877][09932] Loop rollout_proc3_evt_loop terminating...
|
| 166 |
+
[2023-09-12 13:22:54,877][09930] Loop rollout_proc0_evt_loop terminating...
|
| 167 |
+
[2023-09-12 13:22:54,878][09968] Loop rollout_proc6_evt_loop terminating...
|
| 168 |
+
[2023-09-12 13:22:54,880][09964] Stopping RolloutWorker_w5...
|
| 169 |
+
[2023-09-12 13:22:54,880][09931] Stopping RolloutWorker_w2...
|
| 170 |
+
[2023-09-12 13:22:54,880][09964] Loop rollout_proc5_evt_loop terminating...
|
| 171 |
+
[2023-09-12 13:22:54,880][09931] Loop rollout_proc2_evt_loop terminating...
|
| 172 |
+
[2023-09-12 13:22:54,881][09929] Stopping RolloutWorker_w1...
|
| 173 |
+
[2023-09-12 13:22:54,881][09929] Loop rollout_proc1_evt_loop terminating...
|
| 174 |
+
[2023-09-12 13:22:54,882][09967] Stopping RolloutWorker_w4...
|
| 175 |
+
[2023-09-12 13:22:54,882][09967] Loop rollout_proc4_evt_loop terminating...
|
| 176 |
+
[2023-09-12 13:22:54,885][09928] Weights refcount: 2 0
|
| 177 |
+
[2023-09-12 13:22:54,887][09928] Stopping InferenceWorker_p0-w0...
|
| 178 |
+
[2023-09-12 13:22:54,887][09928] Loop inference_proc0-0_evt_loop terminating...
|
| 179 |
+
[2023-09-12 13:22:54,931][09743] Saving /home/cogstack/Documents/optuna/environments/sample_factory/train_dir/default_experiment/checkpoint_p0/checkpoint_000000246_1007616.pth...
|
| 180 |
+
[2023-09-12 13:22:55,021][09743] Stopping LearnerWorker_p0...
|
| 181 |
+
[2023-09-12 13:22:55,022][09743] Loop learner_proc0_evt_loop terminating...
|