Upload ddpm_celebahq checkpoints and training logs
Browse files- ddpm_celebahq/ckpt/step499999/ema.pt +3 -0
- ddpm_celebahq/ckpt/step499999/meta.pt +3 -0
- ddpm_celebahq/ckpt/step499999/model.pt +3 -0
- ddpm_celebahq/ckpt/step499999/optimizer.pt +3 -0
- ddpm_celebahq/config-2023-03-09-21-50-59.yaml +64 -0
- ddpm_celebahq/config-2023-03-14-11-16-58.yaml +54 -0
- ddpm_celebahq/output-2023-03-09-21-50-59.log +893 -0
- ddpm_celebahq/output-2023-03-14-11-16-58.log +389 -0
- ddpm_celebahq/samples.zip +3 -0
- ddpm_celebahq/tensorboard/events.out.tfevents.1678369865.boot-SYS-4029GP-TRT.222806.0 +3 -0
- ddpm_celebahq/tensorboard/events.out.tfevents.1678763820.admin.cluster.local.39024.0 +3 -0
ddpm_celebahq/ckpt/step499999/ema.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b0b8cba5209cfa811be38b5c7ae40b8b42f708cf28f6872f04e02f3d7cafe59c
|
| 3 |
+
size 454829245
|
ddpm_celebahq/ckpt/step499999/meta.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5d05d95c13e6329e3917032458a208b5751a560da25c8fe8060e950de6c8d72c
|
| 3 |
+
size 425
|
ddpm_celebahq/ckpt/step499999/model.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b3ae79cc4526fea57e873133fde2c0f071d0d8abd6444cc4d8d2e422eda777d9
|
| 3 |
+
size 454852221
|
ddpm_celebahq/ckpt/step499999/optimizer.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1576c1ac48a4acc07ad4e13f7529ca2326e2166bbb5c5ad12c87a36a2b9a065e
|
| 3 |
+
size 909770019
|
ddpm_celebahq/config-2023-03-09-21-50-59.yaml
ADDED
|
@@ -0,0 +1,64 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
data:
|
| 2 |
+
dataroot: /data/fengxin/xuyifeng/data/CelebA-HQ/
|
| 3 |
+
img_channels: 3
|
| 4 |
+
img_size: 256
|
| 5 |
+
name: CelebA-HQ
|
| 6 |
+
dataloader:
|
| 7 |
+
micro_batch: 8
|
| 8 |
+
num_workers: 4
|
| 9 |
+
pin_memory: true
|
| 10 |
+
prefetch_factor: 2
|
| 11 |
+
diffusion:
|
| 12 |
+
beta_end: 0.02
|
| 13 |
+
beta_schedule: linear
|
| 14 |
+
beta_start: 0.0001
|
| 15 |
+
objective: pred_eps
|
| 16 |
+
total_steps: 1000
|
| 17 |
+
var_type: fixed_small
|
| 18 |
+
model:
|
| 19 |
+
dim: 128
|
| 20 |
+
dim_mults:
|
| 21 |
+
- 1
|
| 22 |
+
- 1
|
| 23 |
+
- 2
|
| 24 |
+
- 2
|
| 25 |
+
- 4
|
| 26 |
+
- 4
|
| 27 |
+
dropout: 0.0
|
| 28 |
+
ema_decay: 0.9999
|
| 29 |
+
ema_gradual: true
|
| 30 |
+
in_channels: 3
|
| 31 |
+
n_heads: 1
|
| 32 |
+
num_res_blocks: 2
|
| 33 |
+
out_channels: 3
|
| 34 |
+
type: unet
|
| 35 |
+
use_attn:
|
| 36 |
+
- false
|
| 37 |
+
- false
|
| 38 |
+
- false
|
| 39 |
+
- false
|
| 40 |
+
- true
|
| 41 |
+
- false
|
| 42 |
+
sample:
|
| 43 |
+
load_ema: true
|
| 44 |
+
micro_batch: 50
|
| 45 |
+
mode: sample
|
| 46 |
+
n_denoise: 20
|
| 47 |
+
n_progressive: 20
|
| 48 |
+
n_samples: null
|
| 49 |
+
save_dir: null
|
| 50 |
+
skip_steps: null
|
| 51 |
+
weights: null
|
| 52 |
+
seed: 2022
|
| 53 |
+
train:
|
| 54 |
+
batch_size: 64
|
| 55 |
+
clip_grad_norm: 1.0
|
| 56 |
+
n_samples: 36
|
| 57 |
+
n_steps: 500000
|
| 58 |
+
optim:
|
| 59 |
+
lr: 2.0e-05
|
| 60 |
+
type: Adam
|
| 61 |
+
print_freq: 400
|
| 62 |
+
resume: null
|
| 63 |
+
sample_freq: 5000
|
| 64 |
+
save_freq: 10000
|
ddpm_celebahq/config-2023-03-14-11-16-58.yaml
ADDED
|
@@ -0,0 +1,54 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
data:
|
| 2 |
+
dataroot: /amax/xyf/data/CelebA-HQ/
|
| 3 |
+
img_channels: 3
|
| 4 |
+
img_size: 256
|
| 5 |
+
name: CelebA-HQ
|
| 6 |
+
dataloader:
|
| 7 |
+
micro_batch: 8
|
| 8 |
+
num_workers: 4
|
| 9 |
+
pin_memory: true
|
| 10 |
+
prefetch_factor: 2
|
| 11 |
+
diffusion:
|
| 12 |
+
beta_end: 0.02
|
| 13 |
+
beta_schedule: linear
|
| 14 |
+
beta_start: 0.0001
|
| 15 |
+
objective: pred_eps
|
| 16 |
+
total_steps: 1000
|
| 17 |
+
var_type: fixed_small
|
| 18 |
+
model:
|
| 19 |
+
dim: 128
|
| 20 |
+
dim_mults:
|
| 21 |
+
- 1
|
| 22 |
+
- 1
|
| 23 |
+
- 2
|
| 24 |
+
- 2
|
| 25 |
+
- 4
|
| 26 |
+
- 4
|
| 27 |
+
dropout: 0.0
|
| 28 |
+
ema_decay: 0.9999
|
| 29 |
+
ema_gradual: true
|
| 30 |
+
in_channels: 3
|
| 31 |
+
n_heads: 1
|
| 32 |
+
num_res_blocks: 2
|
| 33 |
+
out_channels: 3
|
| 34 |
+
type: unet
|
| 35 |
+
use_attn:
|
| 36 |
+
- false
|
| 37 |
+
- false
|
| 38 |
+
- false
|
| 39 |
+
- false
|
| 40 |
+
- true
|
| 41 |
+
- false
|
| 42 |
+
seed: 2022
|
| 43 |
+
train:
|
| 44 |
+
batch_size: 64
|
| 45 |
+
clip_grad_norm: 1.0
|
| 46 |
+
n_samples: 36
|
| 47 |
+
n_steps: 500000
|
| 48 |
+
optim:
|
| 49 |
+
lr: 2.0e-05
|
| 50 |
+
type: Adam
|
| 51 |
+
print_freq: 400
|
| 52 |
+
resume: ./runs/ddpm_celebahq/ckpt/step349999/
|
| 53 |
+
sample_freq: 5000
|
| 54 |
+
save_freq: 10000
|
ddpm_celebahq/output-2023-03-09-21-50-59.log
ADDED
|
@@ -0,0 +1,893 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
2023-03-09 21:51:05,177 - INFO - Experiment directory: ./runs/ddpm_celebahq/
|
| 2 |
+
2023-03-09 21:51:05,179 - INFO - Number of processes: 4
|
| 3 |
+
2023-03-09 21:51:05,179 - INFO - Distributed type: MULTI_GPU
|
| 4 |
+
2023-03-09 21:51:05,179 - INFO - Mixed precision: fp16
|
| 5 |
+
2023-03-09 21:51:05,408 - INFO - Size of training set: 24183
|
| 6 |
+
2023-03-09 21:51:05,409 - INFO - Batch size per process: 16
|
| 7 |
+
2023-03-09 21:51:05,409 - INFO - Total batch size: 64
|
| 8 |
+
2023-03-09 21:51:07,635 - INFO - Start training...
|
| 9 |
+
2023-03-09 21:58:30,584 - INFO - [Train] step: 399, loss: 0.016238, lr: 0.000020
|
| 10 |
+
2023-03-09 22:05:47,133 - INFO - [Train] step: 799, loss: 0.007404, lr: 0.000020
|
| 11 |
+
2023-03-09 22:13:05,590 - INFO - [Train] step: 1199, loss: 0.006821, lr: 0.000020
|
| 12 |
+
2023-03-09 22:20:23,556 - INFO - [Train] step: 1599, loss: 0.005147, lr: 0.000020
|
| 13 |
+
2023-03-09 22:27:40,698 - INFO - [Train] step: 1999, loss: 0.011604, lr: 0.000020
|
| 14 |
+
2023-03-09 22:34:59,254 - INFO - [Train] step: 2399, loss: 0.016084, lr: 0.000020
|
| 15 |
+
2023-03-09 22:42:17,326 - INFO - [Train] step: 2799, loss: 0.006824, lr: 0.000020
|
| 16 |
+
2023-03-09 22:49:34,674 - INFO - [Train] step: 3199, loss: 0.046819, lr: 0.000020
|
| 17 |
+
2023-03-09 22:56:50,526 - INFO - [Train] step: 3599, loss: 0.033713, lr: 0.000020
|
| 18 |
+
2023-03-09 23:04:08,330 - INFO - [Train] step: 3999, loss: 0.026199, lr: 0.000020
|
| 19 |
+
2023-03-09 23:11:23,415 - INFO - [Train] step: 4399, loss: 0.026416, lr: 0.000020
|
| 20 |
+
2023-03-09 23:18:42,066 - INFO - [Train] step: 4799, loss: 0.006708, lr: 0.000020
|
| 21 |
+
2023-03-09 23:30:13,897 - INFO - [Train] step: 5199, loss: 0.010492, lr: 0.000020
|
| 22 |
+
2023-03-09 23:37:32,392 - INFO - [Train] step: 5599, loss: 0.005229, lr: 0.000020
|
| 23 |
+
2023-03-09 23:44:51,720 - INFO - [Train] step: 5999, loss: 0.006261, lr: 0.000020
|
| 24 |
+
2023-03-09 23:52:07,853 - INFO - [Train] step: 6399, loss: 0.014251, lr: 0.000020
|
| 25 |
+
2023-03-09 23:59:25,022 - INFO - [Train] step: 6799, loss: 0.012289, lr: 0.000020
|
| 26 |
+
2023-03-10 00:06:43,161 - INFO - [Train] step: 7199, loss: 0.002434, lr: 0.000020
|
| 27 |
+
2023-03-10 00:13:57,566 - INFO - [Train] step: 7599, loss: 0.018379, lr: 0.000020
|
| 28 |
+
2023-03-10 00:21:15,724 - INFO - [Train] step: 7999, loss: 0.008544, lr: 0.000020
|
| 29 |
+
2023-03-10 00:28:33,165 - INFO - [Train] step: 8399, loss: 0.010076, lr: 0.000020
|
| 30 |
+
2023-03-10 00:35:49,420 - INFO - [Train] step: 8799, loss: 0.017195, lr: 0.000020
|
| 31 |
+
2023-03-10 00:43:07,072 - INFO - [Train] step: 9199, loss: 0.008099, lr: 0.000020
|
| 32 |
+
2023-03-10 00:50:24,625 - INFO - [Train] step: 9599, loss: 0.005269, lr: 0.000020
|
| 33 |
+
2023-03-10 00:57:39,376 - INFO - [Train] step: 9999, loss: 0.005796, lr: 0.000020
|
| 34 |
+
2023-03-10 01:09:15,302 - INFO - [Train] step: 10399, loss: 0.003558, lr: 0.000020
|
| 35 |
+
2023-03-10 01:16:31,629 - INFO - [Train] step: 10799, loss: 0.010980, lr: 0.000020
|
| 36 |
+
2023-03-10 01:23:48,757 - INFO - [Train] step: 11199, loss: 0.007747, lr: 0.000020
|
| 37 |
+
2023-03-10 01:31:04,343 - INFO - [Train] step: 11599, loss: 0.004996, lr: 0.000020
|
| 38 |
+
2023-03-10 01:38:19,341 - INFO - [Train] step: 11999, loss: 0.001535, lr: 0.000020
|
| 39 |
+
2023-03-10 01:45:36,014 - INFO - [Train] step: 12399, loss: 0.005831, lr: 0.000020
|
| 40 |
+
2023-03-10 01:52:53,085 - INFO - [Train] step: 12799, loss: 0.031975, lr: 0.000020
|
| 41 |
+
2023-03-10 02:00:11,647 - INFO - [Train] step: 13199, loss: 0.013152, lr: 0.000020
|
| 42 |
+
2023-03-10 02:07:27,111 - INFO - [Train] step: 13599, loss: 0.019296, lr: 0.000020
|
| 43 |
+
2023-03-10 02:14:43,542 - INFO - [Train] step: 13999, loss: 0.005936, lr: 0.000020
|
| 44 |
+
2023-03-10 02:22:00,369 - INFO - [Train] step: 14399, loss: 0.010608, lr: 0.000020
|
| 45 |
+
2023-03-10 02:29:16,965 - INFO - [Train] step: 14799, loss: 0.007580, lr: 0.000020
|
| 46 |
+
2023-03-10 02:40:48,895 - INFO - [Train] step: 15199, loss: 0.003113, lr: 0.000020
|
| 47 |
+
2023-03-10 02:48:04,704 - INFO - [Train] step: 15599, loss: 0.009304, lr: 0.000020
|
| 48 |
+
2023-03-10 02:55:21,157 - INFO - [Train] step: 15999, loss: 0.006565, lr: 0.000020
|
| 49 |
+
2023-03-10 03:02:37,927 - INFO - [Train] step: 16399, loss: 0.033882, lr: 0.000020
|
| 50 |
+
2023-03-10 03:09:52,583 - INFO - [Train] step: 16799, loss: 0.010231, lr: 0.000020
|
| 51 |
+
2023-03-10 03:17:10,800 - INFO - [Train] step: 17199, loss: 0.017256, lr: 0.000020
|
| 52 |
+
2023-03-10 03:24:26,176 - INFO - [Train] step: 17599, loss: 0.005058, lr: 0.000020
|
| 53 |
+
2023-03-10 03:31:43,453 - INFO - [Train] step: 17999, loss: 0.014167, lr: 0.000020
|
| 54 |
+
2023-03-10 03:38:59,585 - INFO - [Train] step: 18399, loss: 0.003806, lr: 0.000020
|
| 55 |
+
2023-03-10 03:46:14,442 - INFO - [Train] step: 18799, loss: 0.006275, lr: 0.000020
|
| 56 |
+
2023-03-10 03:53:30,966 - INFO - [Train] step: 19199, loss: 0.002486, lr: 0.000020
|
| 57 |
+
2023-03-10 04:00:48,302 - INFO - [Train] step: 19599, loss: 0.005383, lr: 0.000020
|
| 58 |
+
2023-03-10 04:08:06,157 - INFO - [Train] step: 19999, loss: 0.008164, lr: 0.000020
|
| 59 |
+
2023-03-10 04:19:42,860 - INFO - [Train] step: 20399, loss: 0.006277, lr: 0.000020
|
| 60 |
+
2023-03-10 04:26:59,838 - INFO - [Train] step: 20799, loss: 0.012384, lr: 0.000020
|
| 61 |
+
2023-03-10 04:34:15,988 - INFO - [Train] step: 21199, loss: 0.020082, lr: 0.000020
|
| 62 |
+
2023-03-10 04:41:34,137 - INFO - [Train] step: 21599, loss: 0.034662, lr: 0.000020
|
| 63 |
+
2023-03-10 04:48:49,688 - INFO - [Train] step: 21999, loss: 0.003974, lr: 0.000020
|
| 64 |
+
2023-03-10 04:56:04,860 - INFO - [Train] step: 22399, loss: 0.004995, lr: 0.000020
|
| 65 |
+
2023-03-10 05:03:20,548 - INFO - [Train] step: 22799, loss: 0.013045, lr: 0.000020
|
| 66 |
+
2023-03-10 05:10:37,177 - INFO - [Train] step: 23199, loss: 0.014377, lr: 0.000020
|
| 67 |
+
2023-03-10 05:17:52,965 - INFO - [Train] step: 23599, loss: 0.002784, lr: 0.000020
|
| 68 |
+
2023-03-10 05:25:09,824 - INFO - [Train] step: 23999, loss: 0.003347, lr: 0.000020
|
| 69 |
+
2023-03-10 05:32:24,999 - INFO - [Train] step: 24399, loss: 0.021391, lr: 0.000020
|
| 70 |
+
2023-03-10 05:39:42,817 - INFO - [Train] step: 24799, loss: 0.018484, lr: 0.000020
|
| 71 |
+
2023-03-10 05:51:14,978 - INFO - [Train] step: 25199, loss: 0.006395, lr: 0.000020
|
| 72 |
+
2023-03-10 05:58:30,956 - INFO - [Train] step: 25599, loss: 0.015259, lr: 0.000020
|
| 73 |
+
2023-03-10 06:05:45,975 - INFO - [Train] step: 25999, loss: 0.002671, lr: 0.000020
|
| 74 |
+
2023-03-10 06:13:03,950 - INFO - [Train] step: 26399, loss: 0.008560, lr: 0.000020
|
| 75 |
+
2023-03-10 06:20:20,709 - INFO - [Train] step: 26799, loss: 0.002947, lr: 0.000020
|
| 76 |
+
2023-03-10 06:27:38,418 - INFO - [Train] step: 27199, loss: 0.018766, lr: 0.000020
|
| 77 |
+
2023-03-10 06:34:55,265 - INFO - [Train] step: 27599, loss: 0.009439, lr: 0.000020
|
| 78 |
+
2023-03-10 06:42:14,738 - INFO - [Train] step: 27999, loss: 0.002472, lr: 0.000020
|
| 79 |
+
2023-03-10 06:49:31,222 - INFO - [Train] step: 28399, loss: 0.003887, lr: 0.000020
|
| 80 |
+
2023-03-10 06:56:48,833 - INFO - [Train] step: 28799, loss: 0.033714, lr: 0.000020
|
| 81 |
+
2023-03-10 07:04:06,271 - INFO - [Train] step: 29199, loss: 0.017518, lr: 0.000020
|
| 82 |
+
2023-03-10 07:11:21,329 - INFO - [Train] step: 29599, loss: 0.006144, lr: 0.000020
|
| 83 |
+
2023-03-10 07:18:39,172 - INFO - [Train] step: 29999, loss: 0.005183, lr: 0.000020
|
| 84 |
+
2023-03-10 07:30:13,465 - INFO - [Train] step: 30399, loss: 0.003152, lr: 0.000020
|
| 85 |
+
2023-03-10 07:37:28,762 - INFO - [Train] step: 30799, loss: 0.002465, lr: 0.000020
|
| 86 |
+
2023-03-10 07:44:45,788 - INFO - [Train] step: 31199, loss: 0.003165, lr: 0.000020
|
| 87 |
+
2023-03-10 07:52:00,984 - INFO - [Train] step: 31599, loss: 0.005176, lr: 0.000020
|
| 88 |
+
2023-03-10 07:59:17,046 - INFO - [Train] step: 31999, loss: 0.003373, lr: 0.000020
|
| 89 |
+
2023-03-10 08:06:33,476 - INFO - [Train] step: 32399, loss: 0.007051, lr: 0.000020
|
| 90 |
+
2023-03-10 08:13:50,938 - INFO - [Train] step: 32799, loss: 0.003261, lr: 0.000020
|
| 91 |
+
2023-03-10 08:21:08,959 - INFO - [Train] step: 33199, loss: 0.016323, lr: 0.000020
|
| 92 |
+
2023-03-10 08:28:26,800 - INFO - [Train] step: 33599, loss: 0.004968, lr: 0.000020
|
| 93 |
+
2023-03-10 08:35:44,074 - INFO - [Train] step: 33999, loss: 0.005054, lr: 0.000020
|
| 94 |
+
2023-03-10 08:43:01,315 - INFO - [Train] step: 34399, loss: 0.007740, lr: 0.000020
|
| 95 |
+
2023-03-10 08:50:17,678 - INFO - [Train] step: 34799, loss: 0.009216, lr: 0.000020
|
| 96 |
+
2023-03-10 09:01:52,143 - INFO - [Train] step: 35199, loss: 0.004876, lr: 0.000020
|
| 97 |
+
2023-03-10 09:09:08,134 - INFO - [Train] step: 35599, loss: 0.008269, lr: 0.000020
|
| 98 |
+
2023-03-10 09:16:24,668 - INFO - [Train] step: 35999, loss: 0.002973, lr: 0.000020
|
| 99 |
+
2023-03-10 09:23:41,458 - INFO - [Train] step: 36399, loss: 0.002490, lr: 0.000020
|
| 100 |
+
2023-03-10 09:30:56,751 - INFO - [Train] step: 36799, loss: 0.006066, lr: 0.000020
|
| 101 |
+
2023-03-10 09:38:13,062 - INFO - [Train] step: 37199, loss: 0.004478, lr: 0.000020
|
| 102 |
+
2023-03-10 09:45:31,368 - INFO - [Train] step: 37599, loss: 0.007083, lr: 0.000020
|
| 103 |
+
2023-03-10 09:52:49,265 - INFO - [Train] step: 37999, loss: 0.018550, lr: 0.000020
|
| 104 |
+
2023-03-10 10:00:07,366 - INFO - [Train] step: 38399, loss: 0.006511, lr: 0.000020
|
| 105 |
+
2023-03-10 10:07:25,016 - INFO - [Train] step: 38799, loss: 0.003466, lr: 0.000020
|
| 106 |
+
2023-03-10 10:14:42,752 - INFO - [Train] step: 39199, loss: 0.009857, lr: 0.000020
|
| 107 |
+
2023-03-10 10:21:59,510 - INFO - [Train] step: 39599, loss: 0.012272, lr: 0.000020
|
| 108 |
+
2023-03-10 10:29:18,309 - INFO - [Train] step: 39999, loss: 0.007467, lr: 0.000020
|
| 109 |
+
2023-03-10 10:40:54,399 - INFO - [Train] step: 40399, loss: 0.013849, lr: 0.000020
|
| 110 |
+
2023-03-10 10:48:11,760 - INFO - [Train] step: 40799, loss: 0.010391, lr: 0.000020
|
| 111 |
+
2023-03-10 10:55:29,146 - INFO - [Train] step: 41199, loss: 0.009051, lr: 0.000020
|
| 112 |
+
2023-03-10 11:02:47,250 - INFO - [Train] step: 41599, loss: 0.005582, lr: 0.000020
|
| 113 |
+
2023-03-10 11:10:04,085 - INFO - [Train] step: 41999, loss: 0.005462, lr: 0.000020
|
| 114 |
+
2023-03-10 11:17:22,185 - INFO - [Train] step: 42399, loss: 0.022137, lr: 0.000020
|
| 115 |
+
2023-03-10 11:24:40,610 - INFO - [Train] step: 42799, loss: 0.027267, lr: 0.000020
|
| 116 |
+
2023-03-10 11:31:59,002 - INFO - [Train] step: 43199, loss: 0.005770, lr: 0.000020
|
| 117 |
+
2023-03-10 11:39:16,781 - INFO - [Train] step: 43599, loss: 0.025620, lr: 0.000020
|
| 118 |
+
2023-03-10 11:46:34,961 - INFO - [Train] step: 43999, loss: 0.002744, lr: 0.000020
|
| 119 |
+
2023-03-10 11:53:51,848 - INFO - [Train] step: 44399, loss: 0.015335, lr: 0.000020
|
| 120 |
+
2023-03-10 12:01:09,549 - INFO - [Train] step: 44799, loss: 0.019616, lr: 0.000020
|
| 121 |
+
2023-03-10 12:12:44,341 - INFO - [Train] step: 45199, loss: 0.009860, lr: 0.000020
|
| 122 |
+
2023-03-10 12:20:03,303 - INFO - [Train] step: 45599, loss: 0.006583, lr: 0.000020
|
| 123 |
+
2023-03-10 12:27:24,168 - INFO - [Train] step: 45999, loss: 0.008201, lr: 0.000020
|
| 124 |
+
2023-03-10 12:34:41,659 - INFO - [Train] step: 46399, loss: 0.002160, lr: 0.000020
|
| 125 |
+
2023-03-10 12:41:59,609 - INFO - [Train] step: 46799, loss: 0.012910, lr: 0.000020
|
| 126 |
+
2023-03-10 12:49:18,104 - INFO - [Train] step: 47199, loss: 0.009714, lr: 0.000020
|
| 127 |
+
2023-03-10 12:56:35,756 - INFO - [Train] step: 47599, loss: 0.004828, lr: 0.000020
|
| 128 |
+
2023-03-10 13:03:54,534 - INFO - [Train] step: 47999, loss: 0.003431, lr: 0.000020
|
| 129 |
+
2023-03-10 13:11:11,995 - INFO - [Train] step: 48399, loss: 0.004929, lr: 0.000020
|
| 130 |
+
2023-03-10 13:18:28,549 - INFO - [Train] step: 48799, loss: 0.020942, lr: 0.000020
|
| 131 |
+
2023-03-10 13:25:45,285 - INFO - [Train] step: 49199, loss: 0.004077, lr: 0.000020
|
| 132 |
+
2023-03-10 13:33:03,392 - INFO - [Train] step: 49599, loss: 0.016495, lr: 0.000020
|
| 133 |
+
2023-03-10 13:40:20,901 - INFO - [Train] step: 49999, loss: 0.004044, lr: 0.000020
|
| 134 |
+
2023-03-10 13:51:56,475 - INFO - [Train] step: 50399, loss: 0.024780, lr: 0.000020
|
| 135 |
+
2023-03-10 13:59:12,853 - INFO - [Train] step: 50799, loss: 0.005866, lr: 0.000020
|
| 136 |
+
2023-03-10 14:06:30,615 - INFO - [Train] step: 51199, loss: 0.018788, lr: 0.000020
|
| 137 |
+
2023-03-10 14:13:50,480 - INFO - [Train] step: 51599, loss: 0.018487, lr: 0.000020
|
| 138 |
+
2023-03-10 14:21:06,523 - INFO - [Train] step: 51999, loss: 0.008031, lr: 0.000020
|
| 139 |
+
2023-03-10 14:28:23,838 - INFO - [Train] step: 52399, loss: 0.003580, lr: 0.000020
|
| 140 |
+
2023-03-10 14:35:42,469 - INFO - [Train] step: 52799, loss: 0.004332, lr: 0.000020
|
| 141 |
+
2023-03-10 14:42:58,578 - INFO - [Train] step: 53199, loss: 0.016786, lr: 0.000020
|
| 142 |
+
2023-03-10 14:50:16,003 - INFO - [Train] step: 53599, loss: 0.002716, lr: 0.000020
|
| 143 |
+
2023-03-10 14:57:34,888 - INFO - [Train] step: 53999, loss: 0.010843, lr: 0.000020
|
| 144 |
+
2023-03-10 15:04:51,602 - INFO - [Train] step: 54399, loss: 0.007810, lr: 0.000020
|
| 145 |
+
2023-03-10 15:12:13,052 - INFO - [Train] step: 54799, loss: 0.006223, lr: 0.000020
|
| 146 |
+
2023-03-10 15:23:51,498 - INFO - [Train] step: 55199, loss: 0.011482, lr: 0.000020
|
| 147 |
+
2023-03-10 15:31:11,962 - INFO - [Train] step: 55599, loss: 0.008382, lr: 0.000020
|
| 148 |
+
2023-03-10 15:38:32,487 - INFO - [Train] step: 55999, loss: 0.023479, lr: 0.000020
|
| 149 |
+
2023-03-10 15:45:51,197 - INFO - [Train] step: 56399, loss: 0.015853, lr: 0.000020
|
| 150 |
+
2023-03-10 15:53:06,868 - INFO - [Train] step: 56799, loss: 0.012826, lr: 0.000020
|
| 151 |
+
2023-03-10 16:00:23,570 - INFO - [Train] step: 57199, loss: 0.003291, lr: 0.000020
|
| 152 |
+
2023-03-10 16:07:42,788 - INFO - [Train] step: 57599, loss: 0.003508, lr: 0.000020
|
| 153 |
+
2023-03-10 16:14:59,776 - INFO - [Train] step: 57999, loss: 0.035959, lr: 0.000020
|
| 154 |
+
2023-03-10 16:22:15,987 - INFO - [Train] step: 58399, loss: 0.005783, lr: 0.000020
|
| 155 |
+
2023-03-10 16:29:34,201 - INFO - [Train] step: 58799, loss: 0.014543, lr: 0.000020
|
| 156 |
+
2023-03-10 16:36:52,447 - INFO - [Train] step: 59199, loss: 0.008074, lr: 0.000020
|
| 157 |
+
2023-03-10 16:44:09,412 - INFO - [Train] step: 59599, loss: 0.016773, lr: 0.000020
|
| 158 |
+
2023-03-10 16:51:26,636 - INFO - [Train] step: 59999, loss: 0.007519, lr: 0.000020
|
| 159 |
+
2023-03-10 17:03:01,677 - INFO - [Train] step: 60399, loss: 0.003334, lr: 0.000020
|
| 160 |
+
2023-03-10 17:10:16,785 - INFO - [Train] step: 60799, loss: 0.005840, lr: 0.000020
|
| 161 |
+
2023-03-10 17:17:35,382 - INFO - [Train] step: 61199, loss: 0.008660, lr: 0.000020
|
| 162 |
+
2023-03-10 17:24:52,153 - INFO - [Train] step: 61599, loss: 0.008871, lr: 0.000020
|
| 163 |
+
2023-03-10 17:32:10,681 - INFO - [Train] step: 61999, loss: 0.006048, lr: 0.000020
|
| 164 |
+
2023-03-10 17:39:27,797 - INFO - [Train] step: 62399, loss: 0.020593, lr: 0.000020
|
| 165 |
+
2023-03-10 17:46:44,571 - INFO - [Train] step: 62799, loss: 0.009767, lr: 0.000020
|
| 166 |
+
2023-03-10 17:54:01,274 - INFO - [Train] step: 63199, loss: 0.003158, lr: 0.000020
|
| 167 |
+
2023-03-10 18:01:18,416 - INFO - [Train] step: 63599, loss: 0.002259, lr: 0.000020
|
| 168 |
+
2023-03-10 18:08:34,963 - INFO - [Train] step: 63999, loss: 0.002775, lr: 0.000020
|
| 169 |
+
2023-03-10 18:15:51,740 - INFO - [Train] step: 64399, loss: 0.004474, lr: 0.000020
|
| 170 |
+
2023-03-10 18:23:06,750 - INFO - [Train] step: 64799, loss: 0.007809, lr: 0.000020
|
| 171 |
+
2023-03-10 18:34:40,797 - INFO - [Train] step: 65199, loss: 0.004959, lr: 0.000020
|
| 172 |
+
2023-03-10 18:41:59,736 - INFO - [Train] step: 65599, loss: 0.005450, lr: 0.000020
|
| 173 |
+
2023-03-10 18:49:15,689 - INFO - [Train] step: 65999, loss: 0.008792, lr: 0.000020
|
| 174 |
+
2023-03-10 18:56:32,995 - INFO - [Train] step: 66399, loss: 0.005820, lr: 0.000020
|
| 175 |
+
2023-03-10 19:03:50,754 - INFO - [Train] step: 66799, loss: 0.002413, lr: 0.000020
|
| 176 |
+
2023-03-10 19:11:05,628 - INFO - [Train] step: 67199, loss: 0.007062, lr: 0.000020
|
| 177 |
+
2023-03-10 19:18:22,840 - INFO - [Train] step: 67599, loss: 0.005656, lr: 0.000020
|
| 178 |
+
2023-03-10 19:25:39,721 - INFO - [Train] step: 67999, loss: 0.012289, lr: 0.000020
|
| 179 |
+
2023-03-10 19:32:56,472 - INFO - [Train] step: 68399, loss: 0.010628, lr: 0.000020
|
| 180 |
+
2023-03-10 19:40:13,012 - INFO - [Train] step: 68799, loss: 0.003341, lr: 0.000020
|
| 181 |
+
2023-03-10 19:47:28,366 - INFO - [Train] step: 69199, loss: 0.002595, lr: 0.000020
|
| 182 |
+
2023-03-10 19:54:44,694 - INFO - [Train] step: 69599, loss: 0.019788, lr: 0.000020
|
| 183 |
+
2023-03-10 20:02:00,111 - INFO - [Train] step: 69999, loss: 0.003614, lr: 0.000020
|
| 184 |
+
2023-03-10 20:13:34,462 - INFO - [Train] step: 70399, loss: 0.011459, lr: 0.000020
|
| 185 |
+
2023-03-10 20:20:50,961 - INFO - [Train] step: 70799, loss: 0.013585, lr: 0.000020
|
| 186 |
+
2023-03-10 20:28:06,482 - INFO - [Train] step: 71199, loss: 0.004927, lr: 0.000020
|
| 187 |
+
2023-03-10 20:35:22,905 - INFO - [Train] step: 71599, loss: 0.002009, lr: 0.000020
|
| 188 |
+
2023-03-10 20:42:39,794 - INFO - [Train] step: 71999, loss: 0.004501, lr: 0.000020
|
| 189 |
+
2023-03-10 20:49:57,265 - INFO - [Train] step: 72399, loss: 0.002743, lr: 0.000020
|
| 190 |
+
2023-03-10 20:57:13,105 - INFO - [Train] step: 72799, loss: 0.004917, lr: 0.000020
|
| 191 |
+
2023-03-10 21:04:30,825 - INFO - [Train] step: 73199, loss: 0.001824, lr: 0.000020
|
| 192 |
+
2023-03-10 21:11:50,625 - INFO - [Train] step: 73599, loss: 0.012039, lr: 0.000020
|
| 193 |
+
2023-03-10 21:19:09,600 - INFO - [Train] step: 73999, loss: 0.009084, lr: 0.000020
|
| 194 |
+
2023-03-10 21:26:28,085 - INFO - [Train] step: 74399, loss: 0.005894, lr: 0.000020
|
| 195 |
+
2023-03-10 21:33:43,708 - INFO - [Train] step: 74799, loss: 0.010927, lr: 0.000020
|
| 196 |
+
2023-03-10 21:45:18,475 - INFO - [Train] step: 75199, loss: 0.014266, lr: 0.000020
|
| 197 |
+
2023-03-10 21:52:35,904 - INFO - [Train] step: 75599, loss: 0.005993, lr: 0.000020
|
| 198 |
+
2023-03-10 21:59:52,987 - INFO - [Train] step: 75999, loss: 0.003691, lr: 0.000020
|
| 199 |
+
2023-03-10 22:07:08,955 - INFO - [Train] step: 76399, loss: 0.003811, lr: 0.000020
|
| 200 |
+
2023-03-10 22:14:25,320 - INFO - [Train] step: 76799, loss: 0.028426, lr: 0.000020
|
| 201 |
+
2023-03-10 22:21:40,189 - INFO - [Train] step: 77199, loss: 0.046514, lr: 0.000020
|
| 202 |
+
2023-03-10 22:28:56,695 - INFO - [Train] step: 77599, loss: 0.007021, lr: 0.000020
|
| 203 |
+
2023-03-10 22:36:12,055 - INFO - [Train] step: 77999, loss: 0.008777, lr: 0.000020
|
| 204 |
+
2023-03-10 22:43:28,535 - INFO - [Train] step: 78399, loss: 0.006468, lr: 0.000020
|
| 205 |
+
2023-03-10 22:50:46,519 - INFO - [Train] step: 78799, loss: 0.004038, lr: 0.000020
|
| 206 |
+
2023-03-10 22:58:02,566 - INFO - [Train] step: 79199, loss: 0.005986, lr: 0.000020
|
| 207 |
+
2023-03-10 23:05:19,487 - INFO - [Train] step: 79599, loss: 0.013869, lr: 0.000020
|
| 208 |
+
2023-03-10 23:12:35,331 - INFO - [Train] step: 79999, loss: 0.004121, lr: 0.000020
|
| 209 |
+
2023-03-10 23:24:10,548 - INFO - [Train] step: 80399, loss: 0.003134, lr: 0.000020
|
| 210 |
+
2023-03-10 23:31:25,471 - INFO - [Train] step: 80799, loss: 0.008664, lr: 0.000020
|
| 211 |
+
2023-03-10 23:38:43,938 - INFO - [Train] step: 81199, loss: 0.007764, lr: 0.000020
|
| 212 |
+
2023-03-10 23:46:03,983 - INFO - [Train] step: 81599, loss: 0.009887, lr: 0.000020
|
| 213 |
+
2023-03-10 23:53:22,077 - INFO - [Train] step: 81999, loss: 0.009565, lr: 0.000020
|
| 214 |
+
2023-03-11 00:00:41,551 - INFO - [Train] step: 82399, loss: 0.007369, lr: 0.000020
|
| 215 |
+
2023-03-11 00:07:59,802 - INFO - [Train] step: 82799, loss: 0.005934, lr: 0.000020
|
| 216 |
+
2023-03-11 00:15:15,535 - INFO - [Train] step: 83199, loss: 0.002533, lr: 0.000020
|
| 217 |
+
2023-03-11 00:22:32,606 - INFO - [Train] step: 83599, loss: 0.011916, lr: 0.000020
|
| 218 |
+
2023-03-11 00:29:49,651 - INFO - [Train] step: 83999, loss: 0.007953, lr: 0.000020
|
| 219 |
+
2023-03-11 00:37:06,805 - INFO - [Train] step: 84399, loss: 0.008120, lr: 0.000020
|
| 220 |
+
2023-03-11 00:44:24,594 - INFO - [Train] step: 84799, loss: 0.024131, lr: 0.000020
|
| 221 |
+
2023-03-11 00:55:56,127 - INFO - [Train] step: 85199, loss: 0.002788, lr: 0.000020
|
| 222 |
+
2023-03-11 01:03:14,558 - INFO - [Train] step: 85599, loss: 0.003005, lr: 0.000020
|
| 223 |
+
2023-03-11 01:10:31,040 - INFO - [Train] step: 85999, loss: 0.016811, lr: 0.000020
|
| 224 |
+
2023-03-11 01:17:47,138 - INFO - [Train] step: 86399, loss: 0.006534, lr: 0.000020
|
| 225 |
+
2023-03-11 01:25:05,004 - INFO - [Train] step: 86799, loss: 0.010130, lr: 0.000020
|
| 226 |
+
2023-03-11 01:32:21,719 - INFO - [Train] step: 87199, loss: 0.005580, lr: 0.000020
|
| 227 |
+
2023-03-11 01:39:37,883 - INFO - [Train] step: 87599, loss: 0.005725, lr: 0.000020
|
| 228 |
+
2023-03-11 01:46:53,426 - INFO - [Train] step: 87999, loss: 0.004628, lr: 0.000020
|
| 229 |
+
2023-03-11 01:54:09,996 - INFO - [Train] step: 88399, loss: 0.010830, lr: 0.000020
|
| 230 |
+
2023-03-11 02:01:26,041 - INFO - [Train] step: 88799, loss: 0.005298, lr: 0.000020
|
| 231 |
+
2023-03-11 02:08:44,382 - INFO - [Train] step: 89199, loss: 0.005458, lr: 0.000020
|
| 232 |
+
2023-03-11 02:16:02,606 - INFO - [Train] step: 89599, loss: 0.005595, lr: 0.000020
|
| 233 |
+
2023-03-11 02:23:19,772 - INFO - [Train] step: 89999, loss: 0.015659, lr: 0.000020
|
| 234 |
+
2023-03-11 02:34:55,630 - INFO - [Train] step: 90399, loss: 0.002870, lr: 0.000020
|
| 235 |
+
2023-03-11 02:42:11,754 - INFO - [Train] step: 90799, loss: 0.009628, lr: 0.000020
|
| 236 |
+
2023-03-11 02:49:29,614 - INFO - [Train] step: 91199, loss: 0.002470, lr: 0.000020
|
| 237 |
+
2023-03-11 02:56:48,020 - INFO - [Train] step: 91599, loss: 0.003990, lr: 0.000020
|
| 238 |
+
2023-03-11 03:04:05,590 - INFO - [Train] step: 91999, loss: 0.006200, lr: 0.000020
|
| 239 |
+
2023-03-11 03:11:22,919 - INFO - [Train] step: 92399, loss: 0.003616, lr: 0.000020
|
| 240 |
+
2023-03-11 03:18:41,486 - INFO - [Train] step: 92799, loss: 0.007222, lr: 0.000020
|
| 241 |
+
2023-03-11 03:25:57,566 - INFO - [Train] step: 93199, loss: 0.019666, lr: 0.000020
|
| 242 |
+
2023-03-11 03:33:14,429 - INFO - [Train] step: 93599, loss: 0.002036, lr: 0.000020
|
| 243 |
+
2023-03-11 03:40:31,456 - INFO - [Train] step: 93999, loss: 0.006730, lr: 0.000020
|
| 244 |
+
2023-03-11 03:47:49,603 - INFO - [Train] step: 94399, loss: 0.004959, lr: 0.000020
|
| 245 |
+
2023-03-11 03:55:06,474 - INFO - [Train] step: 94799, loss: 0.002124, lr: 0.000020
|
| 246 |
+
2023-03-11 04:06:40,560 - INFO - [Train] step: 95199, loss: 0.012709, lr: 0.000020
|
| 247 |
+
2023-03-11 04:13:59,718 - INFO - [Train] step: 95599, loss: 0.003398, lr: 0.000020
|
| 248 |
+
2023-03-11 04:21:16,701 - INFO - [Train] step: 95999, loss: 0.020992, lr: 0.000020
|
| 249 |
+
2023-03-11 04:28:34,718 - INFO - [Train] step: 96399, loss: 0.005532, lr: 0.000020
|
| 250 |
+
2023-03-11 04:35:51,488 - INFO - [Train] step: 96799, loss: 0.003470, lr: 0.000020
|
| 251 |
+
2023-03-11 04:43:08,930 - INFO - [Train] step: 97199, loss: 0.018315, lr: 0.000020
|
| 252 |
+
2023-03-11 04:50:25,607 - INFO - [Train] step: 97599, loss: 0.011032, lr: 0.000020
|
| 253 |
+
2023-03-11 04:57:41,660 - INFO - [Train] step: 97999, loss: 0.003257, lr: 0.000020
|
| 254 |
+
2023-03-11 05:05:00,375 - INFO - [Train] step: 98399, loss: 0.005701, lr: 0.000020
|
| 255 |
+
2023-03-11 05:12:18,043 - INFO - [Train] step: 98799, loss: 0.003280, lr: 0.000020
|
| 256 |
+
2023-03-11 05:19:37,076 - INFO - [Train] step: 99199, loss: 0.006920, lr: 0.000020
|
| 257 |
+
2023-03-11 05:26:55,924 - INFO - [Train] step: 99599, loss: 0.009239, lr: 0.000020
|
| 258 |
+
2023-03-11 05:34:14,742 - INFO - [Train] step: 99999, loss: 0.011367, lr: 0.000020
|
| 259 |
+
2023-03-11 05:45:51,778 - INFO - [Train] step: 100399, loss: 0.003896, lr: 0.000020
|
| 260 |
+
2023-03-11 05:53:08,370 - INFO - [Train] step: 100799, loss: 0.010033, lr: 0.000020
|
| 261 |
+
2023-03-11 06:00:24,925 - INFO - [Train] step: 101199, loss: 0.006064, lr: 0.000020
|
| 262 |
+
2023-03-11 06:07:42,317 - INFO - [Train] step: 101599, loss: 0.017514, lr: 0.000020
|
| 263 |
+
2023-03-11 06:14:59,414 - INFO - [Train] step: 101999, loss: 0.007615, lr: 0.000020
|
| 264 |
+
2023-03-11 06:22:16,385 - INFO - [Train] step: 102399, loss: 0.004339, lr: 0.000020
|
| 265 |
+
2023-03-11 06:29:34,419 - INFO - [Train] step: 102799, loss: 0.007039, lr: 0.000020
|
| 266 |
+
2023-03-11 06:36:51,732 - INFO - [Train] step: 103199, loss: 0.003875, lr: 0.000020
|
| 267 |
+
2023-03-11 06:44:08,356 - INFO - [Train] step: 103599, loss: 0.003501, lr: 0.000020
|
| 268 |
+
2023-03-11 06:51:25,593 - INFO - [Train] step: 103999, loss: 0.010158, lr: 0.000020
|
| 269 |
+
2023-03-11 06:58:43,336 - INFO - [Train] step: 104399, loss: 0.010638, lr: 0.000020
|
| 270 |
+
2023-03-11 07:06:00,136 - INFO - [Train] step: 104799, loss: 0.004833, lr: 0.000020
|
| 271 |
+
2023-03-11 07:17:37,891 - INFO - [Train] step: 105199, loss: 0.002464, lr: 0.000020
|
| 272 |
+
2023-03-11 07:24:56,134 - INFO - [Train] step: 105599, loss: 0.005881, lr: 0.000020
|
| 273 |
+
2023-03-11 07:32:13,267 - INFO - [Train] step: 105999, loss: 0.011488, lr: 0.000020
|
| 274 |
+
2023-03-11 07:39:31,029 - INFO - [Train] step: 106399, loss: 0.002424, lr: 0.000020
|
| 275 |
+
2023-03-11 07:46:48,551 - INFO - [Train] step: 106799, loss: 0.006791, lr: 0.000020
|
| 276 |
+
2023-03-11 07:54:05,686 - INFO - [Train] step: 107199, loss: 0.002028, lr: 0.000020
|
| 277 |
+
2023-03-11 08:01:23,950 - INFO - [Train] step: 107599, loss: 0.031971, lr: 0.000020
|
| 278 |
+
2023-03-11 08:08:42,180 - INFO - [Train] step: 107999, loss: 0.009882, lr: 0.000020
|
| 279 |
+
2023-03-11 08:15:59,661 - INFO - [Train] step: 108399, loss: 0.014999, lr: 0.000020
|
| 280 |
+
2023-03-11 08:23:16,675 - INFO - [Train] step: 108799, loss: 0.004669, lr: 0.000020
|
| 281 |
+
2023-03-11 08:30:34,930 - INFO - [Train] step: 109199, loss: 0.005429, lr: 0.000020
|
| 282 |
+
2023-03-11 08:37:52,734 - INFO - [Train] step: 109599, loss: 0.008722, lr: 0.000020
|
| 283 |
+
2023-03-11 08:45:10,151 - INFO - [Train] step: 109999, loss: 0.005470, lr: 0.000020
|
| 284 |
+
2023-03-11 08:56:46,492 - INFO - [Train] step: 110399, loss: 0.007998, lr: 0.000020
|
| 285 |
+
2023-03-11 09:04:03,906 - INFO - [Train] step: 110799, loss: 0.007606, lr: 0.000020
|
| 286 |
+
2023-03-11 09:11:23,703 - INFO - [Train] step: 111199, loss: 0.009633, lr: 0.000020
|
| 287 |
+
2023-03-11 09:18:43,321 - INFO - [Train] step: 111599, loss: 0.010018, lr: 0.000020
|
| 288 |
+
2023-03-11 09:26:01,501 - INFO - [Train] step: 111999, loss: 0.007148, lr: 0.000020
|
| 289 |
+
2023-03-11 09:33:20,131 - INFO - [Train] step: 112399, loss: 0.016403, lr: 0.000020
|
| 290 |
+
2023-03-11 09:40:36,599 - INFO - [Train] step: 112799, loss: 0.005345, lr: 0.000020
|
| 291 |
+
2023-03-11 09:47:55,098 - INFO - [Train] step: 113199, loss: 0.004402, lr: 0.000020
|
| 292 |
+
2023-03-11 09:55:12,045 - INFO - [Train] step: 113599, loss: 0.006984, lr: 0.000020
|
| 293 |
+
2023-03-11 10:02:28,713 - INFO - [Train] step: 113999, loss: 0.017797, lr: 0.000020
|
| 294 |
+
2023-03-11 10:09:45,621 - INFO - [Train] step: 114399, loss: 0.016687, lr: 0.000020
|
| 295 |
+
2023-03-11 10:17:02,954 - INFO - [Train] step: 114799, loss: 0.014141, lr: 0.000020
|
| 296 |
+
2023-03-11 10:28:35,432 - INFO - [Train] step: 115199, loss: 0.015732, lr: 0.000020
|
| 297 |
+
2023-03-11 10:35:53,726 - INFO - [Train] step: 115599, loss: 0.009807, lr: 0.000020
|
| 298 |
+
2023-03-11 10:43:09,793 - INFO - [Train] step: 115999, loss: 0.002469, lr: 0.000020
|
| 299 |
+
2023-03-11 10:50:27,173 - INFO - [Train] step: 116399, loss: 0.003318, lr: 0.000020
|
| 300 |
+
2023-03-11 10:57:44,603 - INFO - [Train] step: 116799, loss: 0.008228, lr: 0.000020
|
| 301 |
+
2023-03-11 11:05:00,927 - INFO - [Train] step: 117199, loss: 0.005147, lr: 0.000020
|
| 302 |
+
2023-03-11 11:12:18,829 - INFO - [Train] step: 117599, loss: 0.004911, lr: 0.000020
|
| 303 |
+
2023-03-11 11:19:38,968 - INFO - [Train] step: 117999, loss: 0.004523, lr: 0.000020
|
| 304 |
+
2023-03-11 11:26:57,541 - INFO - [Train] step: 118399, loss: 0.007089, lr: 0.000020
|
| 305 |
+
2023-03-11 11:34:14,583 - INFO - [Train] step: 118799, loss: 0.016500, lr: 0.000020
|
| 306 |
+
2023-03-11 11:41:32,397 - INFO - [Train] step: 119199, loss: 0.020201, lr: 0.000020
|
| 307 |
+
2023-03-11 11:48:50,548 - INFO - [Train] step: 119599, loss: 0.004024, lr: 0.000020
|
| 308 |
+
2023-03-11 11:56:06,960 - INFO - [Train] step: 119999, loss: 0.002289, lr: 0.000020
|
| 309 |
+
2023-03-11 12:07:42,217 - INFO - [Train] step: 120399, loss: 0.006459, lr: 0.000020
|
| 310 |
+
2023-03-11 12:14:58,765 - INFO - [Train] step: 120799, loss: 0.006613, lr: 0.000020
|
| 311 |
+
2023-03-11 12:22:15,098 - INFO - [Train] step: 121199, loss: 0.005149, lr: 0.000020
|
| 312 |
+
2023-03-11 12:29:32,400 - INFO - [Train] step: 121599, loss: 0.016409, lr: 0.000020
|
| 313 |
+
2023-03-11 12:36:49,313 - INFO - [Train] step: 121999, loss: 0.025191, lr: 0.000020
|
| 314 |
+
2023-03-11 12:44:07,550 - INFO - [Train] step: 122399, loss: 0.001851, lr: 0.000020
|
| 315 |
+
2023-03-11 12:51:21,901 - INFO - [Train] step: 122799, loss: 0.006535, lr: 0.000020
|
| 316 |
+
2023-03-11 12:58:28,452 - INFO - [Train] step: 123199, loss: 0.014650, lr: 0.000020
|
| 317 |
+
2023-03-11 13:05:38,457 - INFO - [Train] step: 123599, loss: 0.005154, lr: 0.000020
|
| 318 |
+
2023-03-11 13:12:44,535 - INFO - [Train] step: 123999, loss: 0.004007, lr: 0.000020
|
| 319 |
+
2023-03-11 13:19:54,611 - INFO - [Train] step: 124399, loss: 0.005758, lr: 0.000020
|
| 320 |
+
2023-03-11 13:27:01,543 - INFO - [Train] step: 124799, loss: 0.014380, lr: 0.000020
|
| 321 |
+
2023-03-11 13:38:24,310 - INFO - [Train] step: 125199, loss: 0.009532, lr: 0.000020
|
| 322 |
+
2023-03-11 13:45:32,553 - INFO - [Train] step: 125599, loss: 0.000561, lr: 0.000020
|
| 323 |
+
2023-03-11 13:52:38,917 - INFO - [Train] step: 125999, loss: 0.005083, lr: 0.000020
|
| 324 |
+
2023-03-11 13:59:45,355 - INFO - [Train] step: 126399, loss: 0.005843, lr: 0.000020
|
| 325 |
+
2023-03-11 14:06:52,535 - INFO - [Train] step: 126799, loss: 0.006214, lr: 0.000020
|
| 326 |
+
2023-03-11 14:14:05,261 - INFO - [Train] step: 127199, loss: 0.003786, lr: 0.000020
|
| 327 |
+
2023-03-11 14:21:13,306 - INFO - [Train] step: 127599, loss: 0.009029, lr: 0.000020
|
| 328 |
+
2023-03-11 14:28:19,694 - INFO - [Train] step: 127999, loss: 0.013584, lr: 0.000020
|
| 329 |
+
2023-03-11 14:35:24,696 - INFO - [Train] step: 128399, loss: 0.012154, lr: 0.000020
|
| 330 |
+
2023-03-11 14:42:30,960 - INFO - [Train] step: 128799, loss: 0.005755, lr: 0.000020
|
| 331 |
+
2023-03-11 14:49:36,999 - INFO - [Train] step: 129199, loss: 0.002529, lr: 0.000020
|
| 332 |
+
2023-03-11 14:56:42,846 - INFO - [Train] step: 129599, loss: 0.004997, lr: 0.000020
|
| 333 |
+
2023-03-11 15:03:48,534 - INFO - [Train] step: 129999, loss: 0.003060, lr: 0.000020
|
| 334 |
+
2023-03-11 15:15:14,170 - INFO - [Train] step: 130399, loss: 0.009632, lr: 0.000020
|
| 335 |
+
2023-03-11 15:22:20,781 - INFO - [Train] step: 130799, loss: 0.012451, lr: 0.000020
|
| 336 |
+
2023-03-11 15:29:27,337 - INFO - [Train] step: 131199, loss: 0.002347, lr: 0.000020
|
| 337 |
+
2023-03-11 15:36:31,427 - INFO - [Train] step: 131599, loss: 0.002190, lr: 0.000020
|
| 338 |
+
2023-03-11 15:43:36,322 - INFO - [Train] step: 131999, loss: 0.003123, lr: 0.000020
|
| 339 |
+
2023-03-11 15:50:42,555 - INFO - [Train] step: 132399, loss: 0.010757, lr: 0.000020
|
| 340 |
+
2023-03-11 15:57:49,124 - INFO - [Train] step: 132799, loss: 0.004436, lr: 0.000020
|
| 341 |
+
2023-03-11 16:04:56,682 - INFO - [Train] step: 133199, loss: 0.006512, lr: 0.000020
|
| 342 |
+
2023-03-11 16:12:01,622 - INFO - [Train] step: 133599, loss: 0.018851, lr: 0.000020
|
| 343 |
+
2023-03-11 16:19:08,628 - INFO - [Train] step: 133999, loss: 0.014592, lr: 0.000020
|
| 344 |
+
2023-03-11 16:26:18,786 - INFO - [Train] step: 134399, loss: 0.003378, lr: 0.000020
|
| 345 |
+
2023-03-11 16:33:23,814 - INFO - [Train] step: 134799, loss: 0.015630, lr: 0.000020
|
| 346 |
+
2023-03-11 16:44:45,689 - INFO - [Train] step: 135199, loss: 0.006343, lr: 0.000020
|
| 347 |
+
2023-03-11 16:51:51,060 - INFO - [Train] step: 135599, loss: 0.025293, lr: 0.000020
|
| 348 |
+
2023-03-11 16:58:57,722 - INFO - [Train] step: 135999, loss: 0.004040, lr: 0.000020
|
| 349 |
+
2023-03-11 17:06:05,073 - INFO - [Train] step: 136399, loss: 0.002500, lr: 0.000020
|
| 350 |
+
2023-03-11 17:13:11,321 - INFO - [Train] step: 136799, loss: 0.006101, lr: 0.000020
|
| 351 |
+
2023-03-11 17:20:15,679 - INFO - [Train] step: 137199, loss: 0.008743, lr: 0.000020
|
| 352 |
+
2023-03-11 17:27:20,741 - INFO - [Train] step: 137599, loss: 0.003077, lr: 0.000020
|
| 353 |
+
2023-03-11 17:34:26,912 - INFO - [Train] step: 137999, loss: 0.005924, lr: 0.000020
|
| 354 |
+
2023-03-11 17:41:32,289 - INFO - [Train] step: 138399, loss: 0.007569, lr: 0.000020
|
| 355 |
+
2023-03-11 17:48:38,075 - INFO - [Train] step: 138799, loss: 0.019845, lr: 0.000020
|
| 356 |
+
2023-03-11 17:55:45,131 - INFO - [Train] step: 139199, loss: 0.004091, lr: 0.000020
|
| 357 |
+
2023-03-11 18:02:50,125 - INFO - [Train] step: 139599, loss: 0.004471, lr: 0.000020
|
| 358 |
+
2023-03-11 18:09:55,961 - INFO - [Train] step: 139999, loss: 0.003974, lr: 0.000020
|
| 359 |
+
2023-03-11 18:21:16,933 - INFO - [Train] step: 140399, loss: 0.009260, lr: 0.000020
|
| 360 |
+
2023-03-11 18:28:21,349 - INFO - [Train] step: 140799, loss: 0.006411, lr: 0.000020
|
| 361 |
+
2023-03-11 18:35:26,279 - INFO - [Train] step: 141199, loss: 0.002795, lr: 0.000020
|
| 362 |
+
2023-03-11 18:42:31,160 - INFO - [Train] step: 141599, loss: 0.004380, lr: 0.000020
|
| 363 |
+
2023-03-11 18:49:36,407 - INFO - [Train] step: 141999, loss: 0.012983, lr: 0.000020
|
| 364 |
+
2023-03-11 18:56:46,293 - INFO - [Train] step: 142399, loss: 0.004742, lr: 0.000020
|
| 365 |
+
2023-03-11 19:03:51,733 - INFO - [Train] step: 142799, loss: 0.003615, lr: 0.000020
|
| 366 |
+
2023-03-11 19:10:56,925 - INFO - [Train] step: 143199, loss: 0.010707, lr: 0.000020
|
| 367 |
+
2023-03-11 19:18:02,053 - INFO - [Train] step: 143599, loss: 0.006146, lr: 0.000020
|
| 368 |
+
2023-03-11 19:25:07,419 - INFO - [Train] step: 143999, loss: 0.005632, lr: 0.000020
|
| 369 |
+
2023-03-11 19:32:13,475 - INFO - [Train] step: 144399, loss: 0.004376, lr: 0.000020
|
| 370 |
+
2023-03-11 19:39:19,108 - INFO - [Train] step: 144799, loss: 0.012538, lr: 0.000020
|
| 371 |
+
2023-03-11 19:50:38,311 - INFO - [Train] step: 145199, loss: 0.003627, lr: 0.000020
|
| 372 |
+
2023-03-11 19:57:43,024 - INFO - [Train] step: 145599, loss: 0.007376, lr: 0.000020
|
| 373 |
+
2023-03-11 20:04:47,077 - INFO - [Train] step: 145999, loss: 0.005196, lr: 0.000020
|
| 374 |
+
2023-03-11 20:11:52,541 - INFO - [Train] step: 146399, loss: 0.012478, lr: 0.000020
|
| 375 |
+
2023-03-11 20:18:58,194 - INFO - [Train] step: 146799, loss: 0.011883, lr: 0.000020
|
| 376 |
+
2023-03-11 20:26:06,031 - INFO - [Train] step: 147199, loss: 0.024361, lr: 0.000020
|
| 377 |
+
2023-03-11 20:33:15,963 - INFO - [Train] step: 147599, loss: 0.002802, lr: 0.000020
|
| 378 |
+
2023-03-11 20:40:23,229 - INFO - [Train] step: 147999, loss: 0.005191, lr: 0.000020
|
| 379 |
+
2023-03-11 20:47:27,186 - INFO - [Train] step: 148399, loss: 0.022846, lr: 0.000020
|
| 380 |
+
2023-03-11 20:54:34,896 - INFO - [Train] step: 148799, loss: 0.019694, lr: 0.000020
|
| 381 |
+
2023-03-11 21:01:40,546 - INFO - [Train] step: 149199, loss: 0.013865, lr: 0.000020
|
| 382 |
+
2023-03-11 21:08:44,898 - INFO - [Train] step: 149599, loss: 0.004263, lr: 0.000020
|
| 383 |
+
2023-03-11 21:15:49,971 - INFO - [Train] step: 149999, loss: 0.004382, lr: 0.000020
|
| 384 |
+
2023-03-11 21:27:12,031 - INFO - [Train] step: 150399, loss: 0.008572, lr: 0.000020
|
| 385 |
+
2023-03-11 21:34:18,685 - INFO - [Train] step: 150799, loss: 0.006602, lr: 0.000020
|
| 386 |
+
2023-03-11 21:41:23,922 - INFO - [Train] step: 151199, loss: 0.002497, lr: 0.000020
|
| 387 |
+
2023-03-11 21:48:28,836 - INFO - [Train] step: 151599, loss: 0.020228, lr: 0.000020
|
| 388 |
+
2023-03-11 21:55:34,197 - INFO - [Train] step: 151999, loss: 0.004011, lr: 0.000020
|
| 389 |
+
2023-03-11 22:02:39,194 - INFO - [Train] step: 152399, loss: 0.003067, lr: 0.000020
|
| 390 |
+
2023-03-11 22:09:42,802 - INFO - [Train] step: 152799, loss: 0.005858, lr: 0.000020
|
| 391 |
+
2023-03-11 22:16:47,221 - INFO - [Train] step: 153199, loss: 0.003243, lr: 0.000020
|
| 392 |
+
2023-03-11 22:23:51,698 - INFO - [Train] step: 153599, loss: 0.004074, lr: 0.000020
|
| 393 |
+
2023-03-11 22:30:49,332 - INFO - [Train] step: 153999, loss: 0.017670, lr: 0.000020
|
| 394 |
+
2023-03-11 22:37:40,426 - INFO - [Train] step: 154399, loss: 0.004066, lr: 0.000020
|
| 395 |
+
2023-03-11 22:44:32,353 - INFO - [Train] step: 154799, loss: 0.002151, lr: 0.000020
|
| 396 |
+
2023-03-11 22:55:38,194 - INFO - [Train] step: 155199, loss: 0.002201, lr: 0.000020
|
| 397 |
+
2023-03-11 23:02:30,155 - INFO - [Train] step: 155599, loss: 0.005718, lr: 0.000020
|
| 398 |
+
2023-03-11 23:09:22,007 - INFO - [Train] step: 155999, loss: 0.007180, lr: 0.000020
|
| 399 |
+
2023-03-11 23:16:13,891 - INFO - [Train] step: 156399, loss: 0.002873, lr: 0.000020
|
| 400 |
+
2023-03-11 23:23:05,463 - INFO - [Train] step: 156799, loss: 0.021473, lr: 0.000020
|
| 401 |
+
2023-03-11 23:29:57,233 - INFO - [Train] step: 157199, loss: 0.020485, lr: 0.000020
|
| 402 |
+
2023-03-11 23:36:57,840 - INFO - [Train] step: 157599, loss: 0.006915, lr: 0.000020
|
| 403 |
+
2023-03-11 23:44:01,097 - INFO - [Train] step: 157999, loss: 0.010403, lr: 0.000020
|
| 404 |
+
2023-03-11 23:50:53,077 - INFO - [Train] step: 158399, loss: 0.014333, lr: 0.000020
|
| 405 |
+
2023-03-11 23:57:44,885 - INFO - [Train] step: 158799, loss: 0.002939, lr: 0.000020
|
| 406 |
+
2023-03-12 00:04:36,563 - INFO - [Train] step: 159199, loss: 0.004551, lr: 0.000020
|
| 407 |
+
2023-03-12 00:11:28,525 - INFO - [Train] step: 159599, loss: 0.008768, lr: 0.000020
|
| 408 |
+
2023-03-12 00:18:23,190 - INFO - [Train] step: 159999, loss: 0.020475, lr: 0.000020
|
| 409 |
+
2023-03-12 00:29:36,243 - INFO - [Train] step: 160399, loss: 0.001530, lr: 0.000020
|
| 410 |
+
2023-03-12 00:36:35,224 - INFO - [Train] step: 160799, loss: 0.004393, lr: 0.000020
|
| 411 |
+
2023-03-12 00:43:30,277 - INFO - [Train] step: 161199, loss: 0.012868, lr: 0.000020
|
| 412 |
+
2023-03-12 00:50:21,957 - INFO - [Train] step: 161599, loss: 0.002418, lr: 0.000020
|
| 413 |
+
2023-03-12 00:57:14,195 - INFO - [Train] step: 161999, loss: 0.015283, lr: 0.000020
|
| 414 |
+
2023-03-12 01:04:15,672 - INFO - [Train] step: 162399, loss: 0.006445, lr: 0.000020
|
| 415 |
+
2023-03-12 01:11:10,489 - INFO - [Train] step: 162799, loss: 0.007668, lr: 0.000020
|
| 416 |
+
2023-03-12 01:18:02,531 - INFO - [Train] step: 163199, loss: 0.005140, lr: 0.000020
|
| 417 |
+
2023-03-12 01:24:54,398 - INFO - [Train] step: 163599, loss: 0.014129, lr: 0.000020
|
| 418 |
+
2023-03-12 01:31:47,247 - INFO - [Train] step: 163999, loss: 0.010362, lr: 0.000020
|
| 419 |
+
2023-03-12 01:38:39,146 - INFO - [Train] step: 164399, loss: 0.016608, lr: 0.000020
|
| 420 |
+
2023-03-12 01:45:30,907 - INFO - [Train] step: 164799, loss: 0.004401, lr: 0.000020
|
| 421 |
+
2023-03-12 01:56:35,753 - INFO - [Train] step: 165199, loss: 0.007763, lr: 0.000020
|
| 422 |
+
2023-03-12 02:03:30,472 - INFO - [Train] step: 165599, loss: 0.009738, lr: 0.000020
|
| 423 |
+
2023-03-12 02:10:28,471 - INFO - [Train] step: 165999, loss: 0.013765, lr: 0.000020
|
| 424 |
+
2023-03-12 02:17:20,691 - INFO - [Train] step: 166399, loss: 0.003277, lr: 0.000020
|
| 425 |
+
2023-03-12 02:24:12,609 - INFO - [Train] step: 166799, loss: 0.003486, lr: 0.000020
|
| 426 |
+
2023-03-12 02:31:04,304 - INFO - [Train] step: 167199, loss: 0.002242, lr: 0.000020
|
| 427 |
+
2023-03-12 02:37:59,269 - INFO - [Train] step: 167599, loss: 0.008290, lr: 0.000020
|
| 428 |
+
2023-03-12 02:44:59,522 - INFO - [Train] step: 167999, loss: 0.015730, lr: 0.000020
|
| 429 |
+
2023-03-12 02:51:53,049 - INFO - [Train] step: 168399, loss: 0.010916, lr: 0.000020
|
| 430 |
+
2023-03-12 02:58:44,097 - INFO - [Train] step: 168799, loss: 0.004053, lr: 0.000020
|
| 431 |
+
2023-03-12 03:05:34,897 - INFO - [Train] step: 169199, loss: 0.014103, lr: 0.000020
|
| 432 |
+
2023-03-12 03:12:30,312 - INFO - [Train] step: 169599, loss: 0.006347, lr: 0.000020
|
| 433 |
+
2023-03-12 03:19:21,540 - INFO - [Train] step: 169999, loss: 0.004455, lr: 0.000020
|
| 434 |
+
2023-03-12 03:30:30,352 - INFO - [Train] step: 170399, loss: 0.012792, lr: 0.000020
|
| 435 |
+
2023-03-12 03:37:21,073 - INFO - [Train] step: 170799, loss: 0.006669, lr: 0.000020
|
| 436 |
+
2023-03-12 03:44:11,903 - INFO - [Train] step: 171199, loss: 0.002961, lr: 0.000020
|
| 437 |
+
2023-03-12 03:51:02,992 - INFO - [Train] step: 171599, loss: 0.005145, lr: 0.000020
|
| 438 |
+
2023-03-12 03:57:54,072 - INFO - [Train] step: 171999, loss: 0.011066, lr: 0.000020
|
| 439 |
+
2023-03-12 04:04:48,543 - INFO - [Train] step: 172399, loss: 0.005882, lr: 0.000020
|
| 440 |
+
2023-03-12 04:11:43,037 - INFO - [Train] step: 172799, loss: 0.003114, lr: 0.000020
|
| 441 |
+
2023-03-12 04:18:43,583 - INFO - [Train] step: 173199, loss: 0.005510, lr: 0.000020
|
| 442 |
+
2023-03-12 04:25:36,207 - INFO - [Train] step: 173599, loss: 0.006343, lr: 0.000020
|
| 443 |
+
2023-03-12 04:32:27,196 - INFO - [Train] step: 173999, loss: 0.002221, lr: 0.000020
|
| 444 |
+
2023-03-12 04:39:25,276 - INFO - [Train] step: 174399, loss: 0.003699, lr: 0.000020
|
| 445 |
+
2023-03-12 04:46:20,207 - INFO - [Train] step: 174799, loss: 0.008737, lr: 0.000020
|
| 446 |
+
2023-03-12 04:57:22,073 - INFO - [Train] step: 175199, loss: 0.008197, lr: 0.000020
|
| 447 |
+
2023-03-12 05:04:12,877 - INFO - [Train] step: 175599, loss: 0.009444, lr: 0.000020
|
| 448 |
+
2023-03-12 05:11:03,882 - INFO - [Train] step: 175999, loss: 0.024660, lr: 0.000020
|
| 449 |
+
2023-03-12 05:17:56,742 - INFO - [Train] step: 176399, loss: 0.003557, lr: 0.000020
|
| 450 |
+
2023-03-12 05:24:47,841 - INFO - [Train] step: 176799, loss: 0.003185, lr: 0.000020
|
| 451 |
+
2023-03-12 05:31:39,456 - INFO - [Train] step: 177199, loss: 0.014776, lr: 0.000020
|
| 452 |
+
2023-03-12 05:38:32,916 - INFO - [Train] step: 177599, loss: 0.002506, lr: 0.000020
|
| 453 |
+
2023-03-12 05:45:24,219 - INFO - [Train] step: 177999, loss: 0.008730, lr: 0.000020
|
| 454 |
+
2023-03-12 05:52:18,227 - INFO - [Train] step: 178399, loss: 0.013050, lr: 0.000020
|
| 455 |
+
2023-03-12 05:59:12,226 - INFO - [Train] step: 178799, loss: 0.004717, lr: 0.000020
|
| 456 |
+
2023-03-12 06:06:05,526 - INFO - [Train] step: 179199, loss: 0.009336, lr: 0.000020
|
| 457 |
+
2023-03-12 06:12:57,096 - INFO - [Train] step: 179599, loss: 0.002970, lr: 0.000020
|
| 458 |
+
2023-03-12 06:19:55,450 - INFO - [Train] step: 179999, loss: 0.006409, lr: 0.000020
|
| 459 |
+
2023-03-12 06:31:10,953 - INFO - [Train] step: 180399, loss: 0.011131, lr: 0.000020
|
| 460 |
+
2023-03-12 06:38:14,042 - INFO - [Train] step: 180799, loss: 0.015210, lr: 0.000020
|
| 461 |
+
2023-03-12 06:45:10,319 - INFO - [Train] step: 181199, loss: 0.006342, lr: 0.000020
|
| 462 |
+
2023-03-12 06:52:02,810 - INFO - [Train] step: 181599, loss: 0.001723, lr: 0.000020
|
| 463 |
+
2023-03-12 06:58:53,700 - INFO - [Train] step: 181999, loss: 0.016422, lr: 0.000020
|
| 464 |
+
2023-03-12 07:05:44,340 - INFO - [Train] step: 182399, loss: 0.006542, lr: 0.000020
|
| 465 |
+
2023-03-12 07:12:34,741 - INFO - [Train] step: 182799, loss: 0.004603, lr: 0.000020
|
| 466 |
+
2023-03-12 07:19:25,717 - INFO - [Train] step: 183199, loss: 0.005993, lr: 0.000020
|
| 467 |
+
2023-03-12 07:26:17,192 - INFO - [Train] step: 183599, loss: 0.001729, lr: 0.000020
|
| 468 |
+
2023-03-12 07:33:07,612 - INFO - [Train] step: 183999, loss: 0.003530, lr: 0.000020
|
| 469 |
+
2023-03-12 07:40:02,086 - INFO - [Train] step: 184399, loss: 0.005695, lr: 0.000020
|
| 470 |
+
2023-03-12 07:46:53,392 - INFO - [Train] step: 184799, loss: 0.001636, lr: 0.000020
|
| 471 |
+
2023-03-12 07:57:54,792 - INFO - [Train] step: 185199, loss: 0.001984, lr: 0.000020
|
| 472 |
+
2023-03-12 08:04:49,249 - INFO - [Train] step: 185599, loss: 0.007725, lr: 0.000020
|
| 473 |
+
2023-03-12 08:11:49,221 - INFO - [Train] step: 185999, loss: 0.006085, lr: 0.000020
|
| 474 |
+
2023-03-12 08:18:52,197 - INFO - [Train] step: 186399, loss: 0.013102, lr: 0.000020
|
| 475 |
+
2023-03-12 08:25:46,627 - INFO - [Train] step: 186799, loss: 0.006902, lr: 0.000020
|
| 476 |
+
2023-03-12 08:32:40,250 - INFO - [Train] step: 187199, loss: 0.009928, lr: 0.000020
|
| 477 |
+
2023-03-12 08:39:40,107 - INFO - [Train] step: 187599, loss: 0.003979, lr: 0.000020
|
| 478 |
+
2023-03-12 08:46:35,140 - INFO - [Train] step: 187999, loss: 0.009272, lr: 0.000020
|
| 479 |
+
2023-03-12 08:53:27,465 - INFO - [Train] step: 188399, loss: 0.005454, lr: 0.000020
|
| 480 |
+
2023-03-12 09:00:19,237 - INFO - [Train] step: 188799, loss: 0.003551, lr: 0.000020
|
| 481 |
+
2023-03-12 09:07:10,419 - INFO - [Train] step: 189199, loss: 0.004296, lr: 0.000020
|
| 482 |
+
2023-03-12 09:14:03,058 - INFO - [Train] step: 189599, loss: 0.002813, lr: 0.000020
|
| 483 |
+
2023-03-12 09:20:54,122 - INFO - [Train] step: 189999, loss: 0.005217, lr: 0.000020
|
| 484 |
+
2023-03-12 09:31:59,481 - INFO - [Train] step: 190399, loss: 0.002141, lr: 0.000020
|
| 485 |
+
2023-03-12 09:38:50,428 - INFO - [Train] step: 190799, loss: 0.013571, lr: 0.000020
|
| 486 |
+
2023-03-12 09:45:41,102 - INFO - [Train] step: 191199, loss: 0.007098, lr: 0.000020
|
| 487 |
+
2023-03-12 09:52:31,794 - INFO - [Train] step: 191599, loss: 0.010613, lr: 0.000020
|
| 488 |
+
2023-03-12 09:59:22,935 - INFO - [Train] step: 191999, loss: 0.018198, lr: 0.000020
|
| 489 |
+
2023-03-12 10:06:14,030 - INFO - [Train] step: 192399, loss: 0.007232, lr: 0.000020
|
| 490 |
+
2023-03-12 10:13:05,087 - INFO - [Train] step: 192799, loss: 0.011173, lr: 0.000020
|
| 491 |
+
2023-03-12 10:19:56,146 - INFO - [Train] step: 193199, loss: 0.009082, lr: 0.000020
|
| 492 |
+
2023-03-12 10:26:46,874 - INFO - [Train] step: 193599, loss: 0.006499, lr: 0.000020
|
| 493 |
+
2023-03-12 10:33:37,823 - INFO - [Train] step: 193999, loss: 0.003081, lr: 0.000020
|
| 494 |
+
2023-03-12 10:40:28,912 - INFO - [Train] step: 194399, loss: 0.018146, lr: 0.000020
|
| 495 |
+
2023-03-12 10:47:19,923 - INFO - [Train] step: 194799, loss: 0.004834, lr: 0.000020
|
| 496 |
+
2023-03-12 10:58:21,853 - INFO - [Train] step: 195199, loss: 0.004834, lr: 0.000020
|
| 497 |
+
2023-03-12 11:05:15,505 - INFO - [Train] step: 195599, loss: 0.008350, lr: 0.000020
|
| 498 |
+
2023-03-12 11:12:13,927 - INFO - [Train] step: 195999, loss: 0.016337, lr: 0.000020
|
| 499 |
+
2023-03-12 11:19:06,011 - INFO - [Train] step: 196399, loss: 0.005988, lr: 0.000020
|
| 500 |
+
2023-03-12 11:25:57,859 - INFO - [Train] step: 196799, loss: 0.008619, lr: 0.000020
|
| 501 |
+
2023-03-12 11:32:48,935 - INFO - [Train] step: 197199, loss: 0.013696, lr: 0.000020
|
| 502 |
+
2023-03-12 11:39:39,997 - INFO - [Train] step: 197599, loss: 0.008003, lr: 0.000020
|
| 503 |
+
2023-03-12 11:46:41,512 - INFO - [Train] step: 197999, loss: 0.001779, lr: 0.000020
|
| 504 |
+
2023-03-12 11:53:41,396 - INFO - [Train] step: 198399, loss: 0.006573, lr: 0.000020
|
| 505 |
+
2023-03-12 12:00:31,433 - INFO - [Train] step: 198799, loss: 0.013192, lr: 0.000020
|
| 506 |
+
2023-03-12 12:07:21,575 - INFO - [Train] step: 199199, loss: 0.014054, lr: 0.000020
|
| 507 |
+
2023-03-12 12:14:14,793 - INFO - [Train] step: 199599, loss: 0.013202, lr: 0.000020
|
| 508 |
+
2023-03-12 12:21:11,527 - INFO - [Train] step: 199999, loss: 0.018106, lr: 0.000020
|
| 509 |
+
2023-03-12 12:32:14,965 - INFO - [Train] step: 200399, loss: 0.004018, lr: 0.000020
|
| 510 |
+
2023-03-12 12:39:05,484 - INFO - [Train] step: 200799, loss: 0.005736, lr: 0.000020
|
| 511 |
+
2023-03-12 12:45:56,463 - INFO - [Train] step: 201199, loss: 0.010191, lr: 0.000020
|
| 512 |
+
2023-03-12 12:52:48,232 - INFO - [Train] step: 201599, loss: 0.002486, lr: 0.000020
|
| 513 |
+
2023-03-12 12:59:48,746 - INFO - [Train] step: 201999, loss: 0.004635, lr: 0.000020
|
| 514 |
+
2023-03-12 13:06:47,484 - INFO - [Train] step: 202399, loss: 0.011719, lr: 0.000020
|
| 515 |
+
2023-03-12 13:13:46,799 - INFO - [Train] step: 202799, loss: 0.004873, lr: 0.000020
|
| 516 |
+
2023-03-12 13:20:39,268 - INFO - [Train] step: 203199, loss: 0.005700, lr: 0.000020
|
| 517 |
+
2023-03-12 13:27:31,170 - INFO - [Train] step: 203599, loss: 0.019901, lr: 0.000020
|
| 518 |
+
2023-03-12 13:34:23,418 - INFO - [Train] step: 203999, loss: 0.002666, lr: 0.000020
|
| 519 |
+
2023-03-12 13:41:19,561 - INFO - [Train] step: 204399, loss: 0.005733, lr: 0.000020
|
| 520 |
+
2023-03-12 13:48:12,653 - INFO - [Train] step: 204799, loss: 0.017386, lr: 0.000020
|
| 521 |
+
2023-03-12 13:59:15,727 - INFO - [Train] step: 205199, loss: 0.001020, lr: 0.000020
|
| 522 |
+
2023-03-12 14:06:06,603 - INFO - [Train] step: 205599, loss: 0.002335, lr: 0.000020
|
| 523 |
+
2023-03-12 14:13:01,133 - INFO - [Train] step: 205999, loss: 0.012239, lr: 0.000020
|
| 524 |
+
2023-03-12 14:19:51,828 - INFO - [Train] step: 206399, loss: 0.002556, lr: 0.000020
|
| 525 |
+
2023-03-12 14:26:42,746 - INFO - [Train] step: 206799, loss: 0.013932, lr: 0.000020
|
| 526 |
+
2023-03-12 14:33:33,620 - INFO - [Train] step: 207199, loss: 0.007203, lr: 0.000020
|
| 527 |
+
2023-03-12 14:40:24,469 - INFO - [Train] step: 207599, loss: 0.007921, lr: 0.000020
|
| 528 |
+
2023-03-12 14:47:15,285 - INFO - [Train] step: 207999, loss: 0.003483, lr: 0.000020
|
| 529 |
+
2023-03-12 14:54:05,903 - INFO - [Train] step: 208399, loss: 0.011042, lr: 0.000020
|
| 530 |
+
2023-03-12 15:01:02,857 - INFO - [Train] step: 208799, loss: 0.003755, lr: 0.000020
|
| 531 |
+
2023-03-12 15:08:00,066 - INFO - [Train] step: 209199, loss: 0.004767, lr: 0.000020
|
| 532 |
+
2023-03-12 15:14:51,571 - INFO - [Train] step: 209599, loss: 0.003968, lr: 0.000020
|
| 533 |
+
2023-03-12 15:21:45,294 - INFO - [Train] step: 209999, loss: 0.012843, lr: 0.000020
|
| 534 |
+
2023-03-12 15:32:50,154 - INFO - [Train] step: 210399, loss: 0.002788, lr: 0.000020
|
| 535 |
+
2023-03-12 15:39:40,803 - INFO - [Train] step: 210799, loss: 0.013307, lr: 0.000020
|
| 536 |
+
2023-03-12 15:46:31,274 - INFO - [Train] step: 211199, loss: 0.008136, lr: 0.000020
|
| 537 |
+
2023-03-12 15:53:21,918 - INFO - [Train] step: 211599, loss: 0.010518, lr: 0.000020
|
| 538 |
+
2023-03-12 16:00:16,569 - INFO - [Train] step: 211999, loss: 0.007448, lr: 0.000020
|
| 539 |
+
2023-03-12 16:07:21,285 - INFO - [Train] step: 212399, loss: 0.008368, lr: 0.000020
|
| 540 |
+
2023-03-12 16:14:21,803 - INFO - [Train] step: 212799, loss: 0.006786, lr: 0.000020
|
| 541 |
+
2023-03-12 16:21:16,079 - INFO - [Train] step: 213199, loss: 0.011463, lr: 0.000020
|
| 542 |
+
2023-03-12 16:28:07,514 - INFO - [Train] step: 213599, loss: 0.003258, lr: 0.000020
|
| 543 |
+
2023-03-12 16:35:01,290 - INFO - [Train] step: 213999, loss: 0.001063, lr: 0.000020
|
| 544 |
+
2023-03-12 16:41:53,722 - INFO - [Train] step: 214399, loss: 0.002956, lr: 0.000020
|
| 545 |
+
2023-03-12 16:48:44,088 - INFO - [Train] step: 214799, loss: 0.025092, lr: 0.000020
|
| 546 |
+
2023-03-12 16:59:45,529 - INFO - [Train] step: 215199, loss: 0.003843, lr: 0.000020
|
| 547 |
+
2023-03-12 17:06:37,054 - INFO - [Train] step: 215599, loss: 0.004040, lr: 0.000020
|
| 548 |
+
2023-03-12 17:13:28,270 - INFO - [Train] step: 215999, loss: 0.021368, lr: 0.000020
|
| 549 |
+
2023-03-12 17:20:18,793 - INFO - [Train] step: 216399, loss: 0.003408, lr: 0.000020
|
| 550 |
+
2023-03-12 17:27:09,800 - INFO - [Train] step: 216799, loss: 0.018623, lr: 0.000020
|
| 551 |
+
2023-03-12 17:34:12,363 - INFO - [Train] step: 217199, loss: 0.011921, lr: 0.000020
|
| 552 |
+
2023-03-12 17:41:02,827 - INFO - [Train] step: 217599, loss: 0.010573, lr: 0.000020
|
| 553 |
+
2023-03-12 17:47:53,480 - INFO - [Train] step: 217999, loss: 0.014622, lr: 0.000020
|
| 554 |
+
2023-03-12 17:54:45,117 - INFO - [Train] step: 218399, loss: 0.019030, lr: 0.000020
|
| 555 |
+
2023-03-12 18:01:43,399 - INFO - [Train] step: 218799, loss: 0.010481, lr: 0.000020
|
| 556 |
+
2023-03-12 18:08:42,796 - INFO - [Train] step: 219199, loss: 0.007626, lr: 0.000020
|
| 557 |
+
2023-03-12 18:15:40,916 - INFO - [Train] step: 219599, loss: 0.014245, lr: 0.000020
|
| 558 |
+
2023-03-12 18:22:34,254 - INFO - [Train] step: 219999, loss: 0.003120, lr: 0.000020
|
| 559 |
+
2023-03-12 18:33:48,237 - INFO - [Train] step: 220399, loss: 0.003616, lr: 0.000020
|
| 560 |
+
2023-03-12 18:40:38,314 - INFO - [Train] step: 220799, loss: 0.005698, lr: 0.000020
|
| 561 |
+
2023-03-12 18:47:31,184 - INFO - [Train] step: 221199, loss: 0.013972, lr: 0.000020
|
| 562 |
+
2023-03-12 18:54:21,753 - INFO - [Train] step: 221599, loss: 0.003370, lr: 0.000020
|
| 563 |
+
2023-03-12 19:01:18,791 - INFO - [Train] step: 221999, loss: 0.002366, lr: 0.000020
|
| 564 |
+
2023-03-12 19:08:09,346 - INFO - [Train] step: 222399, loss: 0.004698, lr: 0.000020
|
| 565 |
+
2023-03-12 19:15:10,689 - INFO - [Train] step: 222799, loss: 0.008910, lr: 0.000020
|
| 566 |
+
2023-03-12 19:22:03,517 - INFO - [Train] step: 223199, loss: 0.021408, lr: 0.000020
|
| 567 |
+
2023-03-12 19:29:02,834 - INFO - [Train] step: 223599, loss: 0.005311, lr: 0.000020
|
| 568 |
+
2023-03-12 19:35:56,131 - INFO - [Train] step: 223999, loss: 0.000250, lr: 0.000020
|
| 569 |
+
2023-03-12 19:42:47,285 - INFO - [Train] step: 224399, loss: 0.009910, lr: 0.000020
|
| 570 |
+
2023-03-12 19:49:38,086 - INFO - [Train] step: 224799, loss: 0.028971, lr: 0.000020
|
| 571 |
+
2023-03-12 20:00:44,877 - INFO - [Train] step: 225199, loss: 0.011801, lr: 0.000020
|
| 572 |
+
2023-03-12 20:07:35,657 - INFO - [Train] step: 225599, loss: 0.023990, lr: 0.000020
|
| 573 |
+
2023-03-12 20:14:32,664 - INFO - [Train] step: 225999, loss: 0.003185, lr: 0.000020
|
| 574 |
+
2023-03-12 20:21:34,601 - INFO - [Train] step: 226399, loss: 0.020069, lr: 0.000020
|
| 575 |
+
2023-03-12 20:28:25,512 - INFO - [Train] step: 226799, loss: 0.005016, lr: 0.000020
|
| 576 |
+
2023-03-12 20:35:24,434 - INFO - [Train] step: 227199, loss: 0.006614, lr: 0.000020
|
| 577 |
+
2023-03-12 20:42:17,050 - INFO - [Train] step: 227599, loss: 0.004557, lr: 0.000020
|
| 578 |
+
2023-03-12 20:49:08,678 - INFO - [Train] step: 227999, loss: 0.007163, lr: 0.000020
|
| 579 |
+
2023-03-12 20:56:04,200 - INFO - [Train] step: 228399, loss: 0.007377, lr: 0.000020
|
| 580 |
+
2023-03-12 21:02:55,189 - INFO - [Train] step: 228799, loss: 0.006975, lr: 0.000020
|
| 581 |
+
2023-03-12 21:09:46,150 - INFO - [Train] step: 229199, loss: 0.004122, lr: 0.000020
|
| 582 |
+
2023-03-12 21:16:38,359 - INFO - [Train] step: 229599, loss: 0.006174, lr: 0.000020
|
| 583 |
+
2023-03-12 21:23:29,948 - INFO - [Train] step: 229999, loss: 0.016372, lr: 0.000020
|
| 584 |
+
2023-03-12 21:34:36,037 - INFO - [Train] step: 230399, loss: 0.004267, lr: 0.000020
|
| 585 |
+
2023-03-12 21:41:26,885 - INFO - [Train] step: 230799, loss: 0.003243, lr: 0.000020
|
| 586 |
+
2023-03-12 21:48:18,161 - INFO - [Train] step: 231199, loss: 0.009548, lr: 0.000020
|
| 587 |
+
2023-03-12 21:55:10,054 - INFO - [Train] step: 231599, loss: 0.007078, lr: 0.000020
|
| 588 |
+
2023-03-12 22:02:01,669 - INFO - [Train] step: 231999, loss: 0.003487, lr: 0.000020
|
| 589 |
+
2023-03-12 22:08:52,923 - INFO - [Train] step: 232399, loss: 0.003251, lr: 0.000020
|
| 590 |
+
2023-03-12 22:15:44,346 - INFO - [Train] step: 232799, loss: 0.010321, lr: 0.000020
|
| 591 |
+
2023-03-12 22:22:35,613 - INFO - [Train] step: 233199, loss: 0.013907, lr: 0.000020
|
| 592 |
+
2023-03-12 22:29:26,634 - INFO - [Train] step: 233599, loss: 0.008200, lr: 0.000020
|
| 593 |
+
2023-03-12 22:36:17,871 - INFO - [Train] step: 233999, loss: 0.018386, lr: 0.000020
|
| 594 |
+
2023-03-12 22:43:09,029 - INFO - [Train] step: 234399, loss: 0.012531, lr: 0.000020
|
| 595 |
+
2023-03-12 22:50:00,343 - INFO - [Train] step: 234799, loss: 0.006701, lr: 0.000020
|
| 596 |
+
2023-03-12 23:01:04,021 - INFO - [Train] step: 235199, loss: 0.017267, lr: 0.000020
|
| 597 |
+
2023-03-12 23:07:55,415 - INFO - [Train] step: 235599, loss: 0.011385, lr: 0.000020
|
| 598 |
+
2023-03-12 23:14:47,675 - INFO - [Train] step: 235999, loss: 0.006444, lr: 0.000020
|
| 599 |
+
2023-03-12 23:21:40,457 - INFO - [Train] step: 236399, loss: 0.003670, lr: 0.000020
|
| 600 |
+
2023-03-12 23:28:31,637 - INFO - [Train] step: 236799, loss: 0.007570, lr: 0.000020
|
| 601 |
+
2023-03-12 23:35:23,155 - INFO - [Train] step: 237199, loss: 0.001349, lr: 0.000020
|
| 602 |
+
2023-03-12 23:42:14,304 - INFO - [Train] step: 237599, loss: 0.008396, lr: 0.000020
|
| 603 |
+
2023-03-12 23:49:05,340 - INFO - [Train] step: 237999, loss: 0.001647, lr: 0.000020
|
| 604 |
+
2023-03-12 23:55:56,586 - INFO - [Train] step: 238399, loss: 0.002595, lr: 0.000020
|
| 605 |
+
2023-03-13 00:02:48,229 - INFO - [Train] step: 238799, loss: 0.009916, lr: 0.000020
|
| 606 |
+
2023-03-13 00:09:39,810 - INFO - [Train] step: 239199, loss: 0.008442, lr: 0.000020
|
| 607 |
+
2023-03-13 00:16:31,800 - INFO - [Train] step: 239599, loss: 0.006605, lr: 0.000020
|
| 608 |
+
2023-03-13 00:23:22,762 - INFO - [Train] step: 239999, loss: 0.006629, lr: 0.000020
|
| 609 |
+
2023-03-13 00:34:29,206 - INFO - [Train] step: 240399, loss: 0.004099, lr: 0.000020
|
| 610 |
+
2023-03-13 00:41:20,543 - INFO - [Train] step: 240799, loss: 0.013016, lr: 0.000020
|
| 611 |
+
2023-03-13 00:48:11,993 - INFO - [Train] step: 241199, loss: 0.025243, lr: 0.000020
|
| 612 |
+
2023-03-13 00:55:03,246 - INFO - [Train] step: 241599, loss: 0.030257, lr: 0.000020
|
| 613 |
+
2023-03-13 01:01:54,361 - INFO - [Train] step: 241999, loss: 0.033123, lr: 0.000020
|
| 614 |
+
2023-03-13 01:08:48,835 - INFO - [Train] step: 242399, loss: 0.001115, lr: 0.000020
|
| 615 |
+
2023-03-13 01:15:40,949 - INFO - [Train] step: 242799, loss: 0.003182, lr: 0.000020
|
| 616 |
+
2023-03-13 01:22:32,841 - INFO - [Train] step: 243199, loss: 0.012307, lr: 0.000020
|
| 617 |
+
2023-03-13 01:29:23,947 - INFO - [Train] step: 243599, loss: 0.015034, lr: 0.000020
|
| 618 |
+
2023-03-13 01:36:14,866 - INFO - [Train] step: 243999, loss: 0.007117, lr: 0.000020
|
| 619 |
+
2023-03-13 01:43:05,643 - INFO - [Train] step: 244399, loss: 0.013201, lr: 0.000020
|
| 620 |
+
2023-03-13 01:49:56,663 - INFO - [Train] step: 244799, loss: 0.003195, lr: 0.000020
|
| 621 |
+
2023-03-13 02:00:56,745 - INFO - [Train] step: 245199, loss: 0.003517, lr: 0.000020
|
| 622 |
+
2023-03-13 02:07:46,557 - INFO - [Train] step: 245599, loss: 0.002458, lr: 0.000020
|
| 623 |
+
2023-03-13 02:14:36,729 - INFO - [Train] step: 245999, loss: 0.005642, lr: 0.000020
|
| 624 |
+
2023-03-13 02:21:28,626 - INFO - [Train] step: 246399, loss: 0.002625, lr: 0.000020
|
| 625 |
+
2023-03-13 02:28:17,726 - INFO - [Train] step: 246799, loss: 0.005014, lr: 0.000020
|
| 626 |
+
2023-03-13 02:35:06,778 - INFO - [Train] step: 247199, loss: 0.005307, lr: 0.000020
|
| 627 |
+
2023-03-13 02:41:55,904 - INFO - [Train] step: 247599, loss: 0.010917, lr: 0.000020
|
| 628 |
+
2023-03-13 02:48:44,966 - INFO - [Train] step: 247999, loss: 0.018019, lr: 0.000020
|
| 629 |
+
2023-03-13 02:55:34,480 - INFO - [Train] step: 248399, loss: 0.007892, lr: 0.000020
|
| 630 |
+
2023-03-13 03:02:23,544 - INFO - [Train] step: 248799, loss: 0.002099, lr: 0.000020
|
| 631 |
+
2023-03-13 03:09:13,586 - INFO - [Train] step: 249199, loss: 0.008956, lr: 0.000020
|
| 632 |
+
2023-03-13 03:16:02,863 - INFO - [Train] step: 249599, loss: 0.007834, lr: 0.000020
|
| 633 |
+
2023-03-13 03:22:52,039 - INFO - [Train] step: 249999, loss: 0.012002, lr: 0.000020
|
| 634 |
+
2023-03-13 03:33:54,481 - INFO - [Train] step: 250399, loss: 0.003595, lr: 0.000020
|
| 635 |
+
2023-03-13 03:40:43,555 - INFO - [Train] step: 250799, loss: 0.007523, lr: 0.000020
|
| 636 |
+
2023-03-13 03:47:32,551 - INFO - [Train] step: 251199, loss: 0.001936, lr: 0.000020
|
| 637 |
+
2023-03-13 03:54:22,750 - INFO - [Train] step: 251599, loss: 0.010847, lr: 0.000020
|
| 638 |
+
2023-03-13 04:01:12,178 - INFO - [Train] step: 251999, loss: 0.011809, lr: 0.000020
|
| 639 |
+
2023-03-13 04:08:01,377 - INFO - [Train] step: 252399, loss: 0.011444, lr: 0.000020
|
| 640 |
+
2023-03-13 04:14:51,468 - INFO - [Train] step: 252799, loss: 0.004827, lr: 0.000020
|
| 641 |
+
2023-03-13 04:21:44,938 - INFO - [Train] step: 253199, loss: 0.004744, lr: 0.000020
|
| 642 |
+
2023-03-13 04:28:35,245 - INFO - [Train] step: 253599, loss: 0.004877, lr: 0.000020
|
| 643 |
+
2023-03-13 04:35:29,513 - INFO - [Train] step: 253999, loss: 0.025008, lr: 0.000020
|
| 644 |
+
2023-03-13 04:42:21,820 - INFO - [Train] step: 254399, loss: 0.001773, lr: 0.000020
|
| 645 |
+
2023-03-13 04:49:15,519 - INFO - [Train] step: 254799, loss: 0.010359, lr: 0.000020
|
| 646 |
+
2023-03-13 05:00:16,736 - INFO - [Train] step: 255199, loss: 0.016410, lr: 0.000020
|
| 647 |
+
2023-03-13 05:07:06,158 - INFO - [Train] step: 255599, loss: 0.003341, lr: 0.000020
|
| 648 |
+
2023-03-13 05:13:56,406 - INFO - [Train] step: 255999, loss: 0.021079, lr: 0.000020
|
| 649 |
+
2023-03-13 05:20:45,582 - INFO - [Train] step: 256399, loss: 0.007401, lr: 0.000020
|
| 650 |
+
2023-03-13 05:27:34,840 - INFO - [Train] step: 256799, loss: 0.007383, lr: 0.000020
|
| 651 |
+
2023-03-13 05:34:24,245 - INFO - [Train] step: 257199, loss: 0.020810, lr: 0.000020
|
| 652 |
+
2023-03-13 05:41:13,530 - INFO - [Train] step: 257599, loss: 0.004540, lr: 0.000020
|
| 653 |
+
2023-03-13 05:48:05,214 - INFO - [Train] step: 257999, loss: 0.002846, lr: 0.000020
|
| 654 |
+
2023-03-13 05:54:54,774 - INFO - [Train] step: 258399, loss: 0.013213, lr: 0.000020
|
| 655 |
+
2023-03-13 06:01:44,071 - INFO - [Train] step: 258799, loss: 0.005983, lr: 0.000020
|
| 656 |
+
2023-03-13 06:08:33,400 - INFO - [Train] step: 259199, loss: 0.024310, lr: 0.000020
|
| 657 |
+
2023-03-13 06:15:22,499 - INFO - [Train] step: 259599, loss: 0.006914, lr: 0.000020
|
| 658 |
+
2023-03-13 06:22:13,200 - INFO - [Train] step: 259999, loss: 0.001955, lr: 0.000020
|
| 659 |
+
2023-03-13 06:33:15,936 - INFO - [Train] step: 260399, loss: 0.013778, lr: 0.000020
|
| 660 |
+
2023-03-13 06:40:08,520 - INFO - [Train] step: 260799, loss: 0.014056, lr: 0.000020
|
| 661 |
+
2023-03-13 06:46:58,362 - INFO - [Train] step: 261199, loss: 0.007391, lr: 0.000020
|
| 662 |
+
2023-03-13 06:53:47,603 - INFO - [Train] step: 261599, loss: 0.004887, lr: 0.000020
|
| 663 |
+
2023-03-13 07:00:37,385 - INFO - [Train] step: 261999, loss: 0.009570, lr: 0.000020
|
| 664 |
+
2023-03-13 07:07:27,445 - INFO - [Train] step: 262399, loss: 0.014061, lr: 0.000020
|
| 665 |
+
2023-03-13 07:14:23,207 - INFO - [Train] step: 262799, loss: 0.005096, lr: 0.000020
|
| 666 |
+
2023-03-13 07:21:19,313 - INFO - [Train] step: 263199, loss: 0.009165, lr: 0.000020
|
| 667 |
+
2023-03-13 07:28:09,303 - INFO - [Train] step: 263599, loss: 0.010867, lr: 0.000020
|
| 668 |
+
2023-03-13 07:34:58,488 - INFO - [Train] step: 263999, loss: 0.004980, lr: 0.000020
|
| 669 |
+
2023-03-13 07:41:47,622 - INFO - [Train] step: 264399, loss: 0.005498, lr: 0.000020
|
| 670 |
+
2023-03-13 07:48:36,797 - INFO - [Train] step: 264799, loss: 0.023846, lr: 0.000020
|
| 671 |
+
2023-03-13 07:59:36,597 - INFO - [Train] step: 265199, loss: 0.006759, lr: 0.000020
|
| 672 |
+
2023-03-13 08:06:31,173 - INFO - [Train] step: 265599, loss: 0.002070, lr: 0.000020
|
| 673 |
+
2023-03-13 08:13:25,504 - INFO - [Train] step: 265999, loss: 0.006507, lr: 0.000020
|
| 674 |
+
2023-03-13 08:20:19,588 - INFO - [Train] step: 266399, loss: 0.016894, lr: 0.000020
|
| 675 |
+
2023-03-13 08:27:09,050 - INFO - [Train] step: 266799, loss: 0.002290, lr: 0.000020
|
| 676 |
+
2023-03-13 08:33:58,267 - INFO - [Train] step: 267199, loss: 0.005357, lr: 0.000020
|
| 677 |
+
2023-03-13 08:40:47,125 - INFO - [Train] step: 267599, loss: 0.007987, lr: 0.000020
|
| 678 |
+
2023-03-13 08:47:45,721 - INFO - [Train] step: 267999, loss: 0.004069, lr: 0.000020
|
| 679 |
+
2023-03-13 08:54:40,639 - INFO - [Train] step: 268399, loss: 0.001877, lr: 0.000020
|
| 680 |
+
2023-03-13 09:01:39,810 - INFO - [Train] step: 268799, loss: 0.001461, lr: 0.000020
|
| 681 |
+
2023-03-13 09:08:31,193 - INFO - [Train] step: 269199, loss: 0.001714, lr: 0.000020
|
| 682 |
+
2023-03-13 09:15:21,570 - INFO - [Train] step: 269599, loss: 0.005727, lr: 0.000020
|
| 683 |
+
2023-03-13 09:22:11,507 - INFO - [Train] step: 269999, loss: 0.002467, lr: 0.000020
|
| 684 |
+
2023-03-13 09:33:15,119 - INFO - [Train] step: 270399, loss: 0.007762, lr: 0.000020
|
| 685 |
+
2023-03-13 09:40:05,589 - INFO - [Train] step: 270799, loss: 0.031679, lr: 0.000020
|
| 686 |
+
2023-03-13 09:46:57,010 - INFO - [Train] step: 271199, loss: 0.002788, lr: 0.000020
|
| 687 |
+
2023-03-13 09:53:47,545 - INFO - [Train] step: 271599, loss: 0.006725, lr: 0.000020
|
| 688 |
+
2023-03-13 10:00:45,267 - INFO - [Train] step: 271999, loss: 0.003801, lr: 0.000020
|
| 689 |
+
2023-03-13 10:07:47,922 - INFO - [Train] step: 272399, loss: 0.006334, lr: 0.000020
|
| 690 |
+
2023-03-13 10:14:43,405 - INFO - [Train] step: 272799, loss: 0.006382, lr: 0.000020
|
| 691 |
+
2023-03-13 10:21:37,653 - INFO - [Train] step: 273199, loss: 0.019097, lr: 0.000020
|
| 692 |
+
2023-03-13 10:28:30,557 - INFO - [Train] step: 273599, loss: 0.005663, lr: 0.000020
|
| 693 |
+
2023-03-13 10:35:22,812 - INFO - [Train] step: 273999, loss: 0.005812, lr: 0.000020
|
| 694 |
+
2023-03-13 10:42:18,352 - INFO - [Train] step: 274399, loss: 0.010816, lr: 0.000020
|
| 695 |
+
2023-03-13 10:49:08,251 - INFO - [Train] step: 274799, loss: 0.019303, lr: 0.000020
|
| 696 |
+
2023-03-13 11:00:11,405 - INFO - [Train] step: 275199, loss: 0.024432, lr: 0.000020
|
| 697 |
+
2023-03-13 11:07:03,058 - INFO - [Train] step: 275599, loss: 0.001344, lr: 0.000020
|
| 698 |
+
2023-03-13 11:13:53,922 - INFO - [Train] step: 275999, loss: 0.003646, lr: 0.000020
|
| 699 |
+
2023-03-13 11:20:45,010 - INFO - [Train] step: 276399, loss: 0.003966, lr: 0.000020
|
| 700 |
+
2023-03-13 11:27:36,081 - INFO - [Train] step: 276799, loss: 0.002916, lr: 0.000020
|
| 701 |
+
2023-03-13 11:34:27,198 - INFO - [Train] step: 277199, loss: 0.005934, lr: 0.000020
|
| 702 |
+
2023-03-13 11:41:18,254 - INFO - [Train] step: 277599, loss: 0.002865, lr: 0.000020
|
| 703 |
+
2023-03-13 11:48:13,053 - INFO - [Train] step: 277999, loss: 0.012977, lr: 0.000020
|
| 704 |
+
2023-03-13 11:55:09,877 - INFO - [Train] step: 278399, loss: 0.005667, lr: 0.000020
|
| 705 |
+
2023-03-13 12:02:05,733 - INFO - [Train] step: 278799, loss: 0.003368, lr: 0.000020
|
| 706 |
+
2023-03-13 12:08:57,218 - INFO - [Train] step: 279199, loss: 0.002185, lr: 0.000020
|
| 707 |
+
2023-03-13 12:15:53,388 - INFO - [Train] step: 279599, loss: 0.010274, lr: 0.000020
|
| 708 |
+
2023-03-13 12:22:45,174 - INFO - [Train] step: 279999, loss: 0.002777, lr: 0.000020
|
| 709 |
+
2023-03-13 12:33:52,396 - INFO - [Train] step: 280399, loss: 0.010333, lr: 0.000020
|
| 710 |
+
2023-03-13 12:40:44,193 - INFO - [Train] step: 280799, loss: 0.029238, lr: 0.000020
|
| 711 |
+
2023-03-13 12:47:35,956 - INFO - [Train] step: 281199, loss: 0.005702, lr: 0.000020
|
| 712 |
+
2023-03-13 12:54:30,823 - INFO - [Train] step: 281599, loss: 0.004465, lr: 0.000020
|
| 713 |
+
2023-03-13 13:01:24,592 - INFO - [Train] step: 281999, loss: 0.005784, lr: 0.000020
|
| 714 |
+
2023-03-13 13:08:16,503 - INFO - [Train] step: 282399, loss: 0.003898, lr: 0.000020
|
| 715 |
+
2023-03-13 13:15:08,125 - INFO - [Train] step: 282799, loss: 0.003352, lr: 0.000020
|
| 716 |
+
2023-03-13 13:21:59,684 - INFO - [Train] step: 283199, loss: 0.003208, lr: 0.000020
|
| 717 |
+
2023-03-13 13:28:51,269 - INFO - [Train] step: 283599, loss: 0.012392, lr: 0.000020
|
| 718 |
+
2023-03-13 13:35:44,287 - INFO - [Train] step: 283999, loss: 0.017511, lr: 0.000020
|
| 719 |
+
2023-03-13 13:42:39,612 - INFO - [Train] step: 284399, loss: 0.010490, lr: 0.000020
|
| 720 |
+
2023-03-13 13:49:30,768 - INFO - [Train] step: 284799, loss: 0.012963, lr: 0.000020
|
| 721 |
+
2023-03-13 14:00:35,192 - INFO - [Train] step: 285199, loss: 0.010477, lr: 0.000020
|
| 722 |
+
2023-03-13 14:07:27,776 - INFO - [Train] step: 285599, loss: 0.003808, lr: 0.000020
|
| 723 |
+
2023-03-13 14:14:21,925 - INFO - [Train] step: 285999, loss: 0.004138, lr: 0.000020
|
| 724 |
+
2023-03-13 14:21:17,773 - INFO - [Train] step: 286399, loss: 0.003546, lr: 0.000020
|
| 725 |
+
2023-03-13 14:28:14,266 - INFO - [Train] step: 286799, loss: 0.002622, lr: 0.000020
|
| 726 |
+
2023-03-13 14:35:08,225 - INFO - [Train] step: 287199, loss: 0.006645, lr: 0.000020
|
| 727 |
+
2023-03-13 14:42:00,738 - INFO - [Train] step: 287599, loss: 0.028907, lr: 0.000020
|
| 728 |
+
2023-03-13 14:48:52,319 - INFO - [Train] step: 287999, loss: 0.005399, lr: 0.000020
|
| 729 |
+
2023-03-13 14:55:43,600 - INFO - [Train] step: 288399, loss: 0.006046, lr: 0.000020
|
| 730 |
+
2023-03-13 15:02:39,381 - INFO - [Train] step: 288799, loss: 0.028489, lr: 0.000020
|
| 731 |
+
2023-03-13 15:09:32,092 - INFO - [Train] step: 289199, loss: 0.006497, lr: 0.000020
|
| 732 |
+
2023-03-13 15:16:34,212 - INFO - [Train] step: 289599, loss: 0.001522, lr: 0.000020
|
| 733 |
+
2023-03-13 15:23:29,919 - INFO - [Train] step: 289999, loss: 0.006072, lr: 0.000020
|
| 734 |
+
2023-03-13 15:34:36,103 - INFO - [Train] step: 290399, loss: 0.013861, lr: 0.000020
|
| 735 |
+
2023-03-13 15:41:27,377 - INFO - [Train] step: 290799, loss: 0.006180, lr: 0.000020
|
| 736 |
+
2023-03-13 15:48:18,743 - INFO - [Train] step: 291199, loss: 0.002156, lr: 0.000020
|
| 737 |
+
2023-03-13 15:55:10,286 - INFO - [Train] step: 291599, loss: 0.003214, lr: 0.000020
|
| 738 |
+
2023-03-13 16:02:02,886 - INFO - [Train] step: 291999, loss: 0.011704, lr: 0.000020
|
| 739 |
+
2023-03-13 16:08:54,548 - INFO - [Train] step: 292399, loss: 0.013076, lr: 0.000020
|
| 740 |
+
2023-03-13 16:15:45,876 - INFO - [Train] step: 292799, loss: 0.008146, lr: 0.000020
|
| 741 |
+
2023-03-13 16:22:37,092 - INFO - [Train] step: 293199, loss: 0.016377, lr: 0.000020
|
| 742 |
+
2023-03-13 16:29:30,312 - INFO - [Train] step: 293599, loss: 0.004600, lr: 0.000020
|
| 743 |
+
2023-03-13 16:36:26,569 - INFO - [Train] step: 293999, loss: 0.006515, lr: 0.000020
|
| 744 |
+
2023-03-13 16:43:17,571 - INFO - [Train] step: 294399, loss: 0.009774, lr: 0.000020
|
| 745 |
+
2023-03-13 16:50:08,632 - INFO - [Train] step: 294799, loss: 0.015559, lr: 0.000020
|
| 746 |
+
2023-03-13 17:01:15,773 - INFO - [Train] step: 295199, loss: 0.007656, lr: 0.000020
|
| 747 |
+
2023-03-13 17:08:06,917 - INFO - [Train] step: 295599, loss: 0.008201, lr: 0.000020
|
| 748 |
+
2023-03-13 17:14:59,404 - INFO - [Train] step: 295999, loss: 0.023883, lr: 0.000020
|
| 749 |
+
2023-03-13 17:21:51,730 - INFO - [Train] step: 296399, loss: 0.004983, lr: 0.000020
|
| 750 |
+
2023-03-13 17:28:46,566 - INFO - [Train] step: 296799, loss: 0.003323, lr: 0.000020
|
| 751 |
+
2023-03-13 17:35:39,230 - INFO - [Train] step: 297199, loss: 0.002764, lr: 0.000020
|
| 752 |
+
2023-03-13 17:42:30,626 - INFO - [Train] step: 297599, loss: 0.006029, lr: 0.000020
|
| 753 |
+
2023-03-13 17:49:21,782 - INFO - [Train] step: 297999, loss: 0.005121, lr: 0.000020
|
| 754 |
+
2023-03-13 17:56:13,176 - INFO - [Train] step: 298399, loss: 0.003874, lr: 0.000020
|
| 755 |
+
2023-03-13 18:03:03,989 - INFO - [Train] step: 298799, loss: 0.023873, lr: 0.000020
|
| 756 |
+
2023-03-13 18:09:56,368 - INFO - [Train] step: 299199, loss: 0.002554, lr: 0.000020
|
| 757 |
+
2023-03-13 18:16:51,638 - INFO - [Train] step: 299599, loss: 0.001769, lr: 0.000020
|
| 758 |
+
2023-03-13 18:23:44,256 - INFO - [Train] step: 299999, loss: 0.018092, lr: 0.000020
|
| 759 |
+
2023-03-13 18:34:49,785 - INFO - [Train] step: 300399, loss: 0.004623, lr: 0.000020
|
| 760 |
+
2023-03-13 18:41:41,081 - INFO - [Train] step: 300799, loss: 0.004098, lr: 0.000020
|
| 761 |
+
2023-03-13 18:48:32,155 - INFO - [Train] step: 301199, loss: 0.005048, lr: 0.000020
|
| 762 |
+
2023-03-13 18:55:23,463 - INFO - [Train] step: 301599, loss: 0.009392, lr: 0.000020
|
| 763 |
+
2023-03-13 19:02:17,396 - INFO - [Train] step: 301999, loss: 0.009700, lr: 0.000020
|
| 764 |
+
2023-03-13 19:09:12,869 - INFO - [Train] step: 302399, loss: 0.009571, lr: 0.000020
|
| 765 |
+
2023-03-13 19:16:04,225 - INFO - [Train] step: 302799, loss: 0.006900, lr: 0.000020
|
| 766 |
+
2023-03-13 19:22:55,419 - INFO - [Train] step: 303199, loss: 0.004997, lr: 0.000020
|
| 767 |
+
2023-03-13 19:29:46,836 - INFO - [Train] step: 303599, loss: 0.004203, lr: 0.000020
|
| 768 |
+
2023-03-13 19:36:37,870 - INFO - [Train] step: 303999, loss: 0.001348, lr: 0.000020
|
| 769 |
+
2023-03-13 19:43:28,939 - INFO - [Train] step: 304399, loss: 0.005132, lr: 0.000020
|
| 770 |
+
2023-03-13 19:50:20,178 - INFO - [Train] step: 304799, loss: 0.008815, lr: 0.000020
|
| 771 |
+
2023-03-13 20:01:22,828 - INFO - [Train] step: 305199, loss: 0.017113, lr: 0.000020
|
| 772 |
+
2023-03-13 20:08:13,976 - INFO - [Train] step: 305599, loss: 0.007488, lr: 0.000020
|
| 773 |
+
2023-03-13 20:15:06,484 - INFO - [Train] step: 305999, loss: 0.005023, lr: 0.000020
|
| 774 |
+
2023-03-13 20:21:57,677 - INFO - [Train] step: 306399, loss: 0.002801, lr: 0.000020
|
| 775 |
+
2023-03-13 20:28:48,967 - INFO - [Train] step: 306799, loss: 0.014980, lr: 0.000020
|
| 776 |
+
2023-03-13 20:35:40,589 - INFO - [Train] step: 307199, loss: 0.007916, lr: 0.000020
|
| 777 |
+
2023-03-13 20:42:37,309 - INFO - [Train] step: 307599, loss: 0.011589, lr: 0.000020
|
| 778 |
+
2023-03-13 20:49:29,479 - INFO - [Train] step: 307999, loss: 0.007654, lr: 0.000020
|
| 779 |
+
2023-03-13 20:56:21,664 - INFO - [Train] step: 308399, loss: 0.005791, lr: 0.000020
|
| 780 |
+
2023-03-13 21:03:12,812 - INFO - [Train] step: 308799, loss: 0.003232, lr: 0.000020
|
| 781 |
+
2023-03-13 21:10:04,081 - INFO - [Train] step: 309199, loss: 0.003277, lr: 0.000020
|
| 782 |
+
2023-03-13 21:16:55,309 - INFO - [Train] step: 309599, loss: 0.003859, lr: 0.000020
|
| 783 |
+
2023-03-13 21:23:46,692 - INFO - [Train] step: 309999, loss: 0.003208, lr: 0.000020
|
| 784 |
+
2023-03-13 21:34:52,593 - INFO - [Train] step: 310399, loss: 0.008537, lr: 0.000020
|
| 785 |
+
2023-03-13 21:41:48,416 - INFO - [Train] step: 310799, loss: 0.002282, lr: 0.000020
|
| 786 |
+
2023-03-13 21:48:39,873 - INFO - [Train] step: 311199, loss: 0.010232, lr: 0.000020
|
| 787 |
+
2023-03-13 21:55:31,176 - INFO - [Train] step: 311599, loss: 0.008219, lr: 0.000020
|
| 788 |
+
2023-03-13 22:02:22,461 - INFO - [Train] step: 311999, loss: 0.012646, lr: 0.000020
|
| 789 |
+
2023-03-13 22:09:13,796 - INFO - [Train] step: 312399, loss: 0.015769, lr: 0.000020
|
| 790 |
+
2023-03-13 22:16:05,399 - INFO - [Train] step: 312799, loss: 0.003452, lr: 0.000020
|
| 791 |
+
2023-03-13 22:22:58,150 - INFO - [Train] step: 313199, loss: 0.016433, lr: 0.000020
|
| 792 |
+
2023-03-13 22:29:49,646 - INFO - [Train] step: 313599, loss: 0.003099, lr: 0.000020
|
| 793 |
+
2023-03-13 22:36:41,308 - INFO - [Train] step: 313999, loss: 0.006481, lr: 0.000020
|
| 794 |
+
2023-03-13 22:43:32,604 - INFO - [Train] step: 314399, loss: 0.005413, lr: 0.000020
|
| 795 |
+
2023-03-13 22:50:24,852 - INFO - [Train] step: 314799, loss: 0.023751, lr: 0.000020
|
| 796 |
+
2023-03-13 23:01:28,199 - INFO - [Train] step: 315199, loss: 0.008081, lr: 0.000020
|
| 797 |
+
2023-03-13 23:08:19,561 - INFO - [Train] step: 315599, loss: 0.014369, lr: 0.000020
|
| 798 |
+
2023-03-13 23:15:10,968 - INFO - [Train] step: 315999, loss: 0.003309, lr: 0.000020
|
| 799 |
+
2023-03-13 23:22:02,452 - INFO - [Train] step: 316399, loss: 0.007778, lr: 0.000020
|
| 800 |
+
2023-03-13 23:28:53,926 - INFO - [Train] step: 316799, loss: 0.012955, lr: 0.000020
|
| 801 |
+
2023-03-13 23:35:45,519 - INFO - [Train] step: 317199, loss: 0.002414, lr: 0.000020
|
| 802 |
+
2023-03-13 23:42:37,005 - INFO - [Train] step: 317599, loss: 0.006313, lr: 0.000020
|
| 803 |
+
2023-03-13 23:49:28,312 - INFO - [Train] step: 317999, loss: 0.010278, lr: 0.000020
|
| 804 |
+
2023-03-13 23:56:19,736 - INFO - [Train] step: 318399, loss: 0.003633, lr: 0.000020
|
| 805 |
+
2023-03-14 00:03:11,155 - INFO - [Train] step: 318799, loss: 0.010913, lr: 0.000020
|
| 806 |
+
2023-03-14 00:10:02,501 - INFO - [Train] step: 319199, loss: 0.000752, lr: 0.000020
|
| 807 |
+
2023-03-14 00:16:53,927 - INFO - [Train] step: 319599, loss: 0.003749, lr: 0.000020
|
| 808 |
+
2023-03-14 00:23:45,231 - INFO - [Train] step: 319999, loss: 0.005107, lr: 0.000020
|
| 809 |
+
2023-03-14 00:34:51,259 - INFO - [Train] step: 320399, loss: 0.006051, lr: 0.000020
|
| 810 |
+
2023-03-14 00:41:42,914 - INFO - [Train] step: 320799, loss: 0.003432, lr: 0.000020
|
| 811 |
+
2023-03-14 00:48:35,457 - INFO - [Train] step: 321199, loss: 0.007389, lr: 0.000020
|
| 812 |
+
2023-03-14 00:55:28,133 - INFO - [Train] step: 321599, loss: 0.007039, lr: 0.000020
|
| 813 |
+
2023-03-14 01:02:19,947 - INFO - [Train] step: 321999, loss: 0.009433, lr: 0.000020
|
| 814 |
+
2023-03-14 01:09:13,333 - INFO - [Train] step: 322399, loss: 0.030625, lr: 0.000020
|
| 815 |
+
2023-03-14 01:16:05,071 - INFO - [Train] step: 322799, loss: 0.002882, lr: 0.000020
|
| 816 |
+
2023-03-14 01:22:56,517 - INFO - [Train] step: 323199, loss: 0.014873, lr: 0.000020
|
| 817 |
+
2023-03-14 01:29:47,901 - INFO - [Train] step: 323599, loss: 0.014061, lr: 0.000020
|
| 818 |
+
2023-03-14 01:36:39,432 - INFO - [Train] step: 323999, loss: 0.012637, lr: 0.000020
|
| 819 |
+
2023-03-14 01:43:30,888 - INFO - [Train] step: 324399, loss: 0.020433, lr: 0.000020
|
| 820 |
+
2023-03-14 01:50:23,981 - INFO - [Train] step: 324799, loss: 0.004405, lr: 0.000020
|
| 821 |
+
2023-03-14 02:01:27,704 - INFO - [Train] step: 325199, loss: 0.013927, lr: 0.000020
|
| 822 |
+
2023-03-14 02:08:22,305 - INFO - [Train] step: 325599, loss: 0.004789, lr: 0.000020
|
| 823 |
+
2023-03-14 02:15:14,149 - INFO - [Train] step: 325999, loss: 0.006942, lr: 0.000020
|
| 824 |
+
2023-03-14 02:22:05,429 - INFO - [Train] step: 326399, loss: 0.004011, lr: 0.000020
|
| 825 |
+
2023-03-14 02:28:56,654 - INFO - [Train] step: 326799, loss: 0.008350, lr: 0.000020
|
| 826 |
+
2023-03-14 02:35:47,829 - INFO - [Train] step: 327199, loss: 0.003452, lr: 0.000020
|
| 827 |
+
2023-03-14 02:42:39,355 - INFO - [Train] step: 327599, loss: 0.004033, lr: 0.000020
|
| 828 |
+
2023-03-14 02:49:31,824 - INFO - [Train] step: 327999, loss: 0.006001, lr: 0.000020
|
| 829 |
+
2023-03-14 02:56:33,466 - INFO - [Train] step: 328399, loss: 0.004725, lr: 0.000020
|
| 830 |
+
2023-03-14 03:03:28,715 - INFO - [Train] step: 328799, loss: 0.006759, lr: 0.000020
|
| 831 |
+
2023-03-14 03:10:24,626 - INFO - [Train] step: 329199, loss: 0.006217, lr: 0.000020
|
| 832 |
+
2023-03-14 03:17:21,430 - INFO - [Train] step: 329599, loss: 0.003645, lr: 0.000020
|
| 833 |
+
2023-03-14 03:24:18,642 - INFO - [Train] step: 329999, loss: 0.004912, lr: 0.000020
|
| 834 |
+
2023-03-14 03:35:24,657 - INFO - [Train] step: 330399, loss: 0.017025, lr: 0.000020
|
| 835 |
+
2023-03-14 03:42:16,034 - INFO - [Train] step: 330799, loss: 0.004152, lr: 0.000020
|
| 836 |
+
2023-03-14 03:49:07,644 - INFO - [Train] step: 331199, loss: 0.010600, lr: 0.000020
|
| 837 |
+
2023-03-14 03:55:58,937 - INFO - [Train] step: 331599, loss: 0.007180, lr: 0.000020
|
| 838 |
+
2023-03-14 04:02:50,202 - INFO - [Train] step: 331999, loss: 0.014653, lr: 0.000020
|
| 839 |
+
2023-03-14 04:09:41,586 - INFO - [Train] step: 332399, loss: 0.023172, lr: 0.000020
|
| 840 |
+
2023-03-14 04:16:32,834 - INFO - [Train] step: 332799, loss: 0.006585, lr: 0.000020
|
| 841 |
+
2023-03-14 04:23:27,409 - INFO - [Train] step: 333199, loss: 0.022197, lr: 0.000020
|
| 842 |
+
2023-03-14 04:30:20,349 - INFO - [Train] step: 333599, loss: 0.004975, lr: 0.000020
|
| 843 |
+
2023-03-14 04:37:11,608 - INFO - [Train] step: 333999, loss: 0.004515, lr: 0.000020
|
| 844 |
+
2023-03-14 04:44:03,586 - INFO - [Train] step: 334399, loss: 0.009178, lr: 0.000020
|
| 845 |
+
2023-03-14 04:50:54,816 - INFO - [Train] step: 334799, loss: 0.002407, lr: 0.000020
|
| 846 |
+
2023-03-14 05:01:56,947 - INFO - [Train] step: 335199, loss: 0.006817, lr: 0.000020
|
| 847 |
+
2023-03-14 05:08:47,902 - INFO - [Train] step: 335599, loss: 0.009432, lr: 0.000020
|
| 848 |
+
2023-03-14 05:15:38,963 - INFO - [Train] step: 335999, loss: 0.002410, lr: 0.000020
|
| 849 |
+
2023-03-14 05:22:30,002 - INFO - [Train] step: 336399, loss: 0.006783, lr: 0.000020
|
| 850 |
+
2023-03-14 05:29:21,936 - INFO - [Train] step: 336799, loss: 0.005075, lr: 0.000020
|
| 851 |
+
2023-03-14 05:36:14,328 - INFO - [Train] step: 337199, loss: 0.006009, lr: 0.000020
|
| 852 |
+
2023-03-14 05:43:05,631 - INFO - [Train] step: 337599, loss: 0.011545, lr: 0.000020
|
| 853 |
+
2023-03-14 05:49:56,968 - INFO - [Train] step: 337999, loss: 0.013443, lr: 0.000020
|
| 854 |
+
2023-03-14 05:56:48,877 - INFO - [Train] step: 338399, loss: 0.016692, lr: 0.000020
|
| 855 |
+
2023-03-14 06:03:48,818 - INFO - [Train] step: 338799, loss: 0.006717, lr: 0.000020
|
| 856 |
+
2023-03-14 06:10:49,397 - INFO - [Train] step: 339199, loss: 0.016554, lr: 0.000020
|
| 857 |
+
2023-03-14 06:17:42,591 - INFO - [Train] step: 339599, loss: 0.002984, lr: 0.000020
|
| 858 |
+
2023-03-14 06:24:33,981 - INFO - [Train] step: 339999, loss: 0.009134, lr: 0.000020
|
| 859 |
+
2023-03-14 06:35:39,549 - INFO - [Train] step: 340399, loss: 0.017299, lr: 0.000020
|
| 860 |
+
2023-03-14 06:42:31,096 - INFO - [Train] step: 340799, loss: 0.008845, lr: 0.000020
|
| 861 |
+
2023-03-14 06:49:23,039 - INFO - [Train] step: 341199, loss: 0.009042, lr: 0.000020
|
| 862 |
+
2023-03-14 06:56:14,487 - INFO - [Train] step: 341599, loss: 0.009215, lr: 0.000020
|
| 863 |
+
2023-03-14 07:03:05,738 - INFO - [Train] step: 341999, loss: 0.003333, lr: 0.000020
|
| 864 |
+
2023-03-14 07:09:57,102 - INFO - [Train] step: 342399, loss: 0.004942, lr: 0.000020
|
| 865 |
+
2023-03-14 07:16:50,709 - INFO - [Train] step: 342799, loss: 0.023956, lr: 0.000020
|
| 866 |
+
2023-03-14 07:23:42,498 - INFO - [Train] step: 343199, loss: 0.004877, lr: 0.000020
|
| 867 |
+
2023-03-14 07:30:38,305 - INFO - [Train] step: 343599, loss: 0.006507, lr: 0.000020
|
| 868 |
+
2023-03-14 07:37:32,998 - INFO - [Train] step: 343999, loss: 0.006423, lr: 0.000020
|
| 869 |
+
2023-03-14 07:44:24,585 - INFO - [Train] step: 344399, loss: 0.003244, lr: 0.000020
|
| 870 |
+
2023-03-14 07:51:15,919 - INFO - [Train] step: 344799, loss: 0.003547, lr: 0.000020
|
| 871 |
+
2023-03-14 08:02:18,884 - INFO - [Train] step: 345199, loss: 0.003294, lr: 0.000020
|
| 872 |
+
2023-03-14 08:09:12,906 - INFO - [Train] step: 345599, loss: 0.006054, lr: 0.000020
|
| 873 |
+
2023-03-14 08:16:08,253 - INFO - [Train] step: 345999, loss: 0.023482, lr: 0.000020
|
| 874 |
+
2023-03-14 08:23:08,217 - INFO - [Train] step: 346399, loss: 0.006333, lr: 0.000020
|
| 875 |
+
2023-03-14 08:29:59,543 - INFO - [Train] step: 346799, loss: 0.008325, lr: 0.000020
|
| 876 |
+
2023-03-14 08:36:50,682 - INFO - [Train] step: 347199, loss: 0.012681, lr: 0.000020
|
| 877 |
+
2023-03-14 08:43:42,688 - INFO - [Train] step: 347599, loss: 0.005527, lr: 0.000020
|
| 878 |
+
2023-03-14 08:50:33,878 - INFO - [Train] step: 347999, loss: 0.006182, lr: 0.000020
|
| 879 |
+
2023-03-14 08:57:24,998 - INFO - [Train] step: 348399, loss: 0.003988, lr: 0.000020
|
| 880 |
+
2023-03-14 09:04:18,301 - INFO - [Train] step: 348799, loss: 0.007490, lr: 0.000020
|
| 881 |
+
2023-03-14 09:11:11,148 - INFO - [Train] step: 349199, loss: 0.004363, lr: 0.000020
|
| 882 |
+
2023-03-14 09:18:05,572 - INFO - [Train] step: 349599, loss: 0.014261, lr: 0.000020
|
| 883 |
+
2023-03-14 09:24:56,941 - INFO - [Train] step: 349999, loss: 0.006281, lr: 0.000020
|
| 884 |
+
2023-03-14 09:36:03,180 - INFO - [Train] step: 350399, loss: 0.035124, lr: 0.000020
|
| 885 |
+
2023-03-14 09:42:54,555 - INFO - [Train] step: 350799, loss: 0.012933, lr: 0.000020
|
| 886 |
+
2023-03-14 09:49:45,659 - INFO - [Train] step: 351199, loss: 0.009570, lr: 0.000020
|
| 887 |
+
2023-03-14 09:56:36,956 - INFO - [Train] step: 351599, loss: 0.006246, lr: 0.000020
|
| 888 |
+
2023-03-14 10:03:28,282 - INFO - [Train] step: 351999, loss: 0.004486, lr: 0.000020
|
| 889 |
+
2023-03-14 10:10:19,588 - INFO - [Train] step: 352399, loss: 0.018333, lr: 0.000020
|
| 890 |
+
2023-03-14 10:17:11,871 - INFO - [Train] step: 352799, loss: 0.002961, lr: 0.000020
|
| 891 |
+
2023-03-14 10:24:03,409 - INFO - [Train] step: 353199, loss: 0.006556, lr: 0.000020
|
| 892 |
+
2023-03-14 10:30:54,734 - INFO - [Train] step: 353599, loss: 0.012790, lr: 0.000020
|
| 893 |
+
2023-03-14 10:37:46,928 - INFO - [Train] step: 353999, loss: 0.005496, lr: 0.000020
|
ddpm_celebahq/output-2023-03-14-11-16-58.log
ADDED
|
@@ -0,0 +1,389 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
2023-03-14 11:17:00,096 - INFO - Experiment directory: ./runs/ddpm_celebahq/
|
| 2 |
+
2023-03-14 11:17:00,098 - INFO - Number of processes: 4
|
| 3 |
+
2023-03-14 11:17:00,098 - INFO - Distributed type: MULTI_GPU
|
| 4 |
+
2023-03-14 11:17:00,098 - INFO - Mixed precision: fp16
|
| 5 |
+
2023-03-14 11:17:00,267 - INFO - Size of training set: 24183
|
| 6 |
+
2023-03-14 11:17:00,267 - INFO - Batch size per process: 16
|
| 7 |
+
2023-03-14 11:17:00,267 - INFO - Total batch size: 64
|
| 8 |
+
2023-03-14 11:17:01,426 - INFO - Resume from ./runs/ddpm_celebahq/ckpt/step349999/
|
| 9 |
+
2023-03-14 11:17:02,277 - INFO - Successfully load model from ./runs/ddpm_celebahq/ckpt/step349999/
|
| 10 |
+
2023-03-14 11:17:02,418 - INFO - Successfully load ema from ./runs/ddpm_celebahq/ckpt/step349999/
|
| 11 |
+
2023-03-14 11:17:03,421 - INFO - Successfully load optimizer from ./runs/ddpm_celebahq/ckpt/step349999/
|
| 12 |
+
2023-03-14 11:17:03,424 - INFO - Restart training at step 350000
|
| 13 |
+
2023-03-14 11:17:04,704 - INFO - Start training...
|
| 14 |
+
2023-03-14 11:25:06,687 - INFO - [Train] step: 350399, loss: 0.003970, lr: 0.000020
|
| 15 |
+
2023-03-14 11:33:16,264 - INFO - [Train] step: 350799, loss: 0.000875, lr: 0.000020
|
| 16 |
+
2023-03-14 11:41:24,597 - INFO - [Train] step: 351199, loss: 0.004025, lr: 0.000020
|
| 17 |
+
2023-03-14 11:49:31,747 - INFO - [Train] step: 351599, loss: 0.004447, lr: 0.000020
|
| 18 |
+
2023-03-14 11:57:40,689 - INFO - [Train] step: 351999, loss: 0.014676, lr: 0.000020
|
| 19 |
+
2023-03-14 12:05:50,358 - INFO - [Train] step: 352399, loss: 0.008828, lr: 0.000020
|
| 20 |
+
2023-03-14 12:14:01,095 - INFO - [Train] step: 352799, loss: 0.001600, lr: 0.000020
|
| 21 |
+
2023-03-14 12:22:09,845 - INFO - [Train] step: 353199, loss: 0.012238, lr: 0.000020
|
| 22 |
+
2023-03-14 12:30:18,735 - INFO - [Train] step: 353599, loss: 0.003647, lr: 0.000020
|
| 23 |
+
2023-03-14 12:38:26,101 - INFO - [Train] step: 353999, loss: 0.009305, lr: 0.000020
|
| 24 |
+
2023-03-14 12:46:34,116 - INFO - [Train] step: 354399, loss: 0.010076, lr: 0.000020
|
| 25 |
+
2023-03-14 12:54:43,234 - INFO - [Train] step: 354799, loss: 0.004197, lr: 0.000020
|
| 26 |
+
2023-03-14 13:09:47,797 - INFO - [Train] step: 355199, loss: 0.010474, lr: 0.000020
|
| 27 |
+
2023-03-14 13:17:57,405 - INFO - [Train] step: 355599, loss: 0.004587, lr: 0.000020
|
| 28 |
+
2023-03-14 13:26:07,040 - INFO - [Train] step: 355999, loss: 0.005947, lr: 0.000020
|
| 29 |
+
2023-03-14 13:34:16,193 - INFO - [Train] step: 356399, loss: 0.002616, lr: 0.000020
|
| 30 |
+
2023-03-14 13:42:25,348 - INFO - [Train] step: 356799, loss: 0.016632, lr: 0.000020
|
| 31 |
+
2023-03-14 13:50:33,093 - INFO - [Train] step: 357199, loss: 0.009248, lr: 0.000020
|
| 32 |
+
2023-03-14 13:58:42,468 - INFO - [Train] step: 357599, loss: 0.007584, lr: 0.000020
|
| 33 |
+
2023-03-14 14:06:50,610 - INFO - [Train] step: 357999, loss: 0.005306, lr: 0.000020
|
| 34 |
+
2023-03-14 14:15:03,336 - INFO - [Train] step: 358399, loss: 0.005756, lr: 0.000020
|
| 35 |
+
2023-03-14 14:23:16,259 - INFO - [Train] step: 358799, loss: 0.001198, lr: 0.000020
|
| 36 |
+
2023-03-14 14:31:26,412 - INFO - [Train] step: 359199, loss: 0.003400, lr: 0.000020
|
| 37 |
+
2023-03-14 14:39:38,412 - INFO - [Train] step: 359599, loss: 0.004573, lr: 0.000020
|
| 38 |
+
2023-03-14 14:47:48,307 - INFO - [Train] step: 359999, loss: 0.015435, lr: 0.000020
|
| 39 |
+
2023-03-14 15:03:03,182 - INFO - [Train] step: 360399, loss: 0.008100, lr: 0.000020
|
| 40 |
+
2023-03-14 15:11:13,949 - INFO - [Train] step: 360799, loss: 0.007414, lr: 0.000020
|
| 41 |
+
2023-03-14 15:19:25,718 - INFO - [Train] step: 361199, loss: 0.005287, lr: 0.000020
|
| 42 |
+
2023-03-14 15:27:35,606 - INFO - [Train] step: 361599, loss: 0.015283, lr: 0.000020
|
| 43 |
+
2023-03-14 15:35:44,026 - INFO - [Train] step: 361999, loss: 0.002844, lr: 0.000020
|
| 44 |
+
2023-03-14 15:43:53,943 - INFO - [Train] step: 362399, loss: 0.004362, lr: 0.000020
|
| 45 |
+
2023-03-14 15:52:04,410 - INFO - [Train] step: 362799, loss: 0.001901, lr: 0.000020
|
| 46 |
+
2023-03-14 16:00:15,684 - INFO - [Train] step: 363199, loss: 0.007688, lr: 0.000020
|
| 47 |
+
2023-03-14 16:08:27,463 - INFO - [Train] step: 363599, loss: 0.008041, lr: 0.000020
|
| 48 |
+
2023-03-14 16:16:38,799 - INFO - [Train] step: 363999, loss: 0.006486, lr: 0.000020
|
| 49 |
+
2023-03-14 16:24:51,158 - INFO - [Train] step: 364399, loss: 0.002724, lr: 0.000020
|
| 50 |
+
2023-03-14 16:33:05,180 - INFO - [Train] step: 364799, loss: 0.011561, lr: 0.000020
|
| 51 |
+
2023-03-14 16:48:28,436 - INFO - [Train] step: 365199, loss: 0.005723, lr: 0.000020
|
| 52 |
+
2023-03-14 16:56:41,884 - INFO - [Train] step: 365599, loss: 0.005510, lr: 0.000020
|
| 53 |
+
2023-03-14 17:04:55,520 - INFO - [Train] step: 365999, loss: 0.005707, lr: 0.000020
|
| 54 |
+
2023-03-14 17:13:07,239 - INFO - [Train] step: 366399, loss: 0.004155, lr: 0.000020
|
| 55 |
+
2023-03-14 17:21:18,938 - INFO - [Train] step: 366799, loss: 0.004802, lr: 0.000020
|
| 56 |
+
2023-03-14 17:29:32,502 - INFO - [Train] step: 367199, loss: 0.008124, lr: 0.000020
|
| 57 |
+
2023-03-14 17:37:43,365 - INFO - [Train] step: 367599, loss: 0.003649, lr: 0.000020
|
| 58 |
+
2023-03-14 17:45:54,223 - INFO - [Train] step: 367999, loss: 0.011362, lr: 0.000020
|
| 59 |
+
2023-03-14 17:54:04,890 - INFO - [Train] step: 368399, loss: 0.031951, lr: 0.000020
|
| 60 |
+
2023-03-14 18:02:15,900 - INFO - [Train] step: 368799, loss: 0.003813, lr: 0.000020
|
| 61 |
+
2023-03-14 18:10:31,202 - INFO - [Train] step: 369199, loss: 0.002422, lr: 0.000020
|
| 62 |
+
2023-03-14 18:18:43,809 - INFO - [Train] step: 369599, loss: 0.005812, lr: 0.000020
|
| 63 |
+
2023-03-14 18:26:55,555 - INFO - [Train] step: 369999, loss: 0.018369, lr: 0.000020
|
| 64 |
+
2023-03-14 18:42:15,184 - INFO - [Train] step: 370399, loss: 0.005125, lr: 0.000020
|
| 65 |
+
2023-03-14 18:50:23,785 - INFO - [Train] step: 370799, loss: 0.021020, lr: 0.000020
|
| 66 |
+
2023-03-14 18:58:32,836 - INFO - [Train] step: 371199, loss: 0.008835, lr: 0.000020
|
| 67 |
+
2023-03-14 19:06:43,919 - INFO - [Train] step: 371599, loss: 0.002034, lr: 0.000020
|
| 68 |
+
2023-03-14 19:14:53,667 - INFO - [Train] step: 371999, loss: 0.020777, lr: 0.000020
|
| 69 |
+
2023-03-14 19:23:07,917 - INFO - [Train] step: 372399, loss: 0.002071, lr: 0.000020
|
| 70 |
+
2023-03-14 19:31:21,843 - INFO - [Train] step: 372799, loss: 0.009126, lr: 0.000020
|
| 71 |
+
2023-03-14 19:39:34,751 - INFO - [Train] step: 373199, loss: 0.003135, lr: 0.000020
|
| 72 |
+
2023-03-14 19:47:48,484 - INFO - [Train] step: 373599, loss: 0.007430, lr: 0.000020
|
| 73 |
+
2023-03-14 19:55:58,608 - INFO - [Train] step: 373999, loss: 0.009583, lr: 0.000020
|
| 74 |
+
2023-03-14 20:04:13,040 - INFO - [Train] step: 374399, loss: 0.012695, lr: 0.000020
|
| 75 |
+
2023-03-14 20:12:31,314 - INFO - [Train] step: 374799, loss: 0.002118, lr: 0.000020
|
| 76 |
+
2023-03-14 20:27:55,375 - INFO - [Train] step: 375199, loss: 0.011640, lr: 0.000020
|
| 77 |
+
2023-03-14 20:36:08,654 - INFO - [Train] step: 375599, loss: 0.001690, lr: 0.000020
|
| 78 |
+
2023-03-14 20:44:20,883 - INFO - [Train] step: 375999, loss: 0.011716, lr: 0.000020
|
| 79 |
+
2023-03-14 20:52:35,045 - INFO - [Train] step: 376399, loss: 0.002537, lr: 0.000020
|
| 80 |
+
2023-03-14 21:00:47,046 - INFO - [Train] step: 376799, loss: 0.013246, lr: 0.000020
|
| 81 |
+
2023-03-14 21:08:57,871 - INFO - [Train] step: 377199, loss: 0.004110, lr: 0.000020
|
| 82 |
+
2023-03-14 21:17:10,546 - INFO - [Train] step: 377599, loss: 0.003193, lr: 0.000020
|
| 83 |
+
2023-03-14 21:25:25,092 - INFO - [Train] step: 377999, loss: 0.003690, lr: 0.000020
|
| 84 |
+
2023-03-14 21:33:37,689 - INFO - [Train] step: 378399, loss: 0.011243, lr: 0.000020
|
| 85 |
+
2023-03-14 21:41:48,362 - INFO - [Train] step: 378799, loss: 0.004159, lr: 0.000020
|
| 86 |
+
2023-03-14 21:49:58,734 - INFO - [Train] step: 379199, loss: 0.005250, lr: 0.000020
|
| 87 |
+
2023-03-14 21:58:10,780 - INFO - [Train] step: 379599, loss: 0.013149, lr: 0.000020
|
| 88 |
+
2023-03-14 22:06:20,342 - INFO - [Train] step: 379999, loss: 0.010080, lr: 0.000020
|
| 89 |
+
2023-03-14 22:21:38,432 - INFO - [Train] step: 380399, loss: 0.008977, lr: 0.000020
|
| 90 |
+
2023-03-14 22:29:46,367 - INFO - [Train] step: 380799, loss: 0.005365, lr: 0.000020
|
| 91 |
+
2023-03-14 22:37:53,676 - INFO - [Train] step: 381199, loss: 0.006641, lr: 0.000020
|
| 92 |
+
2023-03-14 22:46:03,400 - INFO - [Train] step: 381599, loss: 0.002233, lr: 0.000020
|
| 93 |
+
2023-03-14 22:54:12,769 - INFO - [Train] step: 381999, loss: 0.015945, lr: 0.000020
|
| 94 |
+
2023-03-14 23:02:22,251 - INFO - [Train] step: 382399, loss: 0.004349, lr: 0.000020
|
| 95 |
+
2023-03-14 23:10:32,192 - INFO - [Train] step: 382799, loss: 0.008812, lr: 0.000020
|
| 96 |
+
2023-03-14 23:18:40,364 - INFO - [Train] step: 383199, loss: 0.008462, lr: 0.000020
|
| 97 |
+
2023-03-14 23:26:48,205 - INFO - [Train] step: 383599, loss: 0.003358, lr: 0.000020
|
| 98 |
+
2023-03-14 23:34:57,142 - INFO - [Train] step: 383999, loss: 0.012008, lr: 0.000020
|
| 99 |
+
2023-03-14 23:43:06,996 - INFO - [Train] step: 384399, loss: 0.004462, lr: 0.000020
|
| 100 |
+
2023-03-14 23:51:13,677 - INFO - [Train] step: 384799, loss: 0.004331, lr: 0.000020
|
| 101 |
+
2023-03-15 00:06:24,706 - INFO - [Train] step: 385199, loss: 0.009878, lr: 0.000020
|
| 102 |
+
2023-03-15 00:14:31,763 - INFO - [Train] step: 385599, loss: 0.002496, lr: 0.000020
|
| 103 |
+
2023-03-15 00:22:39,773 - INFO - [Train] step: 385999, loss: 0.005399, lr: 0.000020
|
| 104 |
+
2023-03-15 00:30:49,008 - INFO - [Train] step: 386399, loss: 0.006340, lr: 0.000020
|
| 105 |
+
2023-03-15 00:39:00,480 - INFO - [Train] step: 386799, loss: 0.007803, lr: 0.000020
|
| 106 |
+
2023-03-15 00:47:08,317 - INFO - [Train] step: 387199, loss: 0.002311, lr: 0.000020
|
| 107 |
+
2023-03-15 00:55:18,850 - INFO - [Train] step: 387599, loss: 0.005942, lr: 0.000020
|
| 108 |
+
2023-03-15 01:03:30,548 - INFO - [Train] step: 387999, loss: 0.003447, lr: 0.000020
|
| 109 |
+
2023-03-15 01:11:39,221 - INFO - [Train] step: 388399, loss: 0.017024, lr: 0.000020
|
| 110 |
+
2023-03-15 01:19:46,784 - INFO - [Train] step: 388799, loss: 0.007986, lr: 0.000020
|
| 111 |
+
2023-03-15 01:27:55,045 - INFO - [Train] step: 389199, loss: 0.004215, lr: 0.000020
|
| 112 |
+
2023-03-15 01:36:04,034 - INFO - [Train] step: 389599, loss: 0.005298, lr: 0.000020
|
| 113 |
+
2023-03-15 01:44:12,692 - INFO - [Train] step: 389999, loss: 0.007917, lr: 0.000020
|
| 114 |
+
2023-03-15 01:59:23,734 - INFO - [Train] step: 390399, loss: 0.018029, lr: 0.000020
|
| 115 |
+
2023-03-15 02:07:29,525 - INFO - [Train] step: 390799, loss: 0.005119, lr: 0.000020
|
| 116 |
+
2023-03-15 02:15:38,339 - INFO - [Train] step: 391199, loss: 0.007850, lr: 0.000020
|
| 117 |
+
2023-03-15 02:23:46,061 - INFO - [Train] step: 391599, loss: 0.002627, lr: 0.000020
|
| 118 |
+
2023-03-15 02:31:51,971 - INFO - [Train] step: 391999, loss: 0.020313, lr: 0.000020
|
| 119 |
+
2023-03-15 02:39:57,364 - INFO - [Train] step: 392399, loss: 0.010299, lr: 0.000020
|
| 120 |
+
2023-03-15 02:48:02,887 - INFO - [Train] step: 392799, loss: 0.018258, lr: 0.000020
|
| 121 |
+
2023-03-15 02:56:10,980 - INFO - [Train] step: 393199, loss: 0.020890, lr: 0.000020
|
| 122 |
+
2023-03-15 03:04:18,178 - INFO - [Train] step: 393599, loss: 0.002848, lr: 0.000020
|
| 123 |
+
2023-03-15 03:12:26,272 - INFO - [Train] step: 393999, loss: 0.006369, lr: 0.000020
|
| 124 |
+
2023-03-15 03:20:34,978 - INFO - [Train] step: 394399, loss: 0.009678, lr: 0.000020
|
| 125 |
+
2023-03-15 03:28:45,807 - INFO - [Train] step: 394799, loss: 0.007569, lr: 0.000020
|
| 126 |
+
2023-03-15 03:43:57,145 - INFO - [Train] step: 395199, loss: 0.001792, lr: 0.000020
|
| 127 |
+
2023-03-15 03:52:05,478 - INFO - [Train] step: 395599, loss: 0.019708, lr: 0.000020
|
| 128 |
+
2023-03-15 04:00:14,368 - INFO - [Train] step: 395999, loss: 0.007647, lr: 0.000020
|
| 129 |
+
2023-03-15 04:08:21,262 - INFO - [Train] step: 396399, loss: 0.009972, lr: 0.000020
|
| 130 |
+
2023-03-15 04:16:27,612 - INFO - [Train] step: 396799, loss: 0.009197, lr: 0.000020
|
| 131 |
+
2023-03-15 04:24:34,218 - INFO - [Train] step: 397199, loss: 0.003339, lr: 0.000020
|
| 132 |
+
2023-03-15 04:32:43,139 - INFO - [Train] step: 397599, loss: 0.004498, lr: 0.000020
|
| 133 |
+
2023-03-15 04:40:51,586 - INFO - [Train] step: 397999, loss: 0.005879, lr: 0.000020
|
| 134 |
+
2023-03-15 04:49:01,083 - INFO - [Train] step: 398399, loss: 0.002673, lr: 0.000020
|
| 135 |
+
2023-03-15 04:57:06,598 - INFO - [Train] step: 398799, loss: 0.004977, lr: 0.000020
|
| 136 |
+
2023-03-15 05:05:11,394 - INFO - [Train] step: 399199, loss: 0.014280, lr: 0.000020
|
| 137 |
+
2023-03-15 05:13:15,814 - INFO - [Train] step: 399599, loss: 0.003304, lr: 0.000020
|
| 138 |
+
2023-03-15 05:21:19,901 - INFO - [Train] step: 399999, loss: 0.026339, lr: 0.000020
|
| 139 |
+
2023-03-15 05:36:22,558 - INFO - [Train] step: 400399, loss: 0.020097, lr: 0.000020
|
| 140 |
+
2023-03-15 05:44:26,977 - INFO - [Train] step: 400799, loss: 0.005036, lr: 0.000020
|
| 141 |
+
2023-03-15 05:52:33,474 - INFO - [Train] step: 401199, loss: 0.012237, lr: 0.000020
|
| 142 |
+
2023-03-15 06:00:37,712 - INFO - [Train] step: 401599, loss: 0.008641, lr: 0.000020
|
| 143 |
+
2023-03-15 06:08:45,691 - INFO - [Train] step: 401999, loss: 0.004047, lr: 0.000020
|
| 144 |
+
2023-03-15 06:16:51,019 - INFO - [Train] step: 402399, loss: 0.011072, lr: 0.000020
|
| 145 |
+
2023-03-15 06:24:58,755 - INFO - [Train] step: 402799, loss: 0.003696, lr: 0.000020
|
| 146 |
+
2023-03-15 06:33:05,510 - INFO - [Train] step: 403199, loss: 0.003335, lr: 0.000020
|
| 147 |
+
2023-03-15 06:41:11,889 - INFO - [Train] step: 403599, loss: 0.005935, lr: 0.000020
|
| 148 |
+
2023-03-15 06:49:20,797 - INFO - [Train] step: 403999, loss: 0.008968, lr: 0.000020
|
| 149 |
+
2023-03-15 06:57:25,748 - INFO - [Train] step: 404399, loss: 0.005360, lr: 0.000020
|
| 150 |
+
2023-03-15 07:05:33,591 - INFO - [Train] step: 404799, loss: 0.007498, lr: 0.000020
|
| 151 |
+
2023-03-15 07:20:41,123 - INFO - [Train] step: 405199, loss: 0.013348, lr: 0.000020
|
| 152 |
+
2023-03-15 07:28:47,738 - INFO - [Train] step: 405599, loss: 0.004439, lr: 0.000020
|
| 153 |
+
2023-03-15 07:36:52,471 - INFO - [Train] step: 405999, loss: 0.005605, lr: 0.000020
|
| 154 |
+
2023-03-15 07:44:55,872 - INFO - [Train] step: 406399, loss: 0.010902, lr: 0.000020
|
| 155 |
+
2023-03-15 07:52:58,844 - INFO - [Train] step: 406799, loss: 0.014158, lr: 0.000020
|
| 156 |
+
2023-03-15 08:01:00,921 - INFO - [Train] step: 407199, loss: 0.005914, lr: 0.000020
|
| 157 |
+
2023-03-15 08:09:04,378 - INFO - [Train] step: 407599, loss: 0.003083, lr: 0.000020
|
| 158 |
+
2023-03-15 08:17:08,859 - INFO - [Train] step: 407999, loss: 0.012349, lr: 0.000020
|
| 159 |
+
2023-03-15 08:25:14,140 - INFO - [Train] step: 408399, loss: 0.003212, lr: 0.000020
|
| 160 |
+
2023-03-15 08:33:18,087 - INFO - [Train] step: 408799, loss: 0.010946, lr: 0.000020
|
| 161 |
+
2023-03-15 08:41:22,240 - INFO - [Train] step: 409199, loss: 0.005972, lr: 0.000020
|
| 162 |
+
2023-03-15 08:49:24,083 - INFO - [Train] step: 409599, loss: 0.006212, lr: 0.000020
|
| 163 |
+
2023-03-15 08:57:28,472 - INFO - [Train] step: 409999, loss: 0.017361, lr: 0.000020
|
| 164 |
+
2023-03-15 09:12:32,515 - INFO - [Train] step: 410399, loss: 0.004148, lr: 0.000020
|
| 165 |
+
2023-03-15 09:20:39,184 - INFO - [Train] step: 410799, loss: 0.017749, lr: 0.000020
|
| 166 |
+
2023-03-15 09:28:44,423 - INFO - [Train] step: 411199, loss: 0.005733, lr: 0.000020
|
| 167 |
+
2023-03-15 09:36:49,320 - INFO - [Train] step: 411599, loss: 0.011234, lr: 0.000020
|
| 168 |
+
2023-03-15 09:44:54,679 - INFO - [Train] step: 411999, loss: 0.005849, lr: 0.000020
|
| 169 |
+
2023-03-15 09:52:58,720 - INFO - [Train] step: 412399, loss: 0.017280, lr: 0.000020
|
| 170 |
+
2023-03-15 10:01:04,020 - INFO - [Train] step: 412799, loss: 0.002729, lr: 0.000020
|
| 171 |
+
2023-03-15 10:09:09,743 - INFO - [Train] step: 413199, loss: 0.003359, lr: 0.000020
|
| 172 |
+
2023-03-15 10:17:17,113 - INFO - [Train] step: 413599, loss: 0.006960, lr: 0.000020
|
| 173 |
+
2023-03-15 10:25:23,756 - INFO - [Train] step: 413999, loss: 0.012596, lr: 0.000020
|
| 174 |
+
2023-03-15 10:33:31,378 - INFO - [Train] step: 414399, loss: 0.006310, lr: 0.000020
|
| 175 |
+
2023-03-15 10:41:39,501 - INFO - [Train] step: 414799, loss: 0.002113, lr: 0.000020
|
| 176 |
+
2023-03-15 10:56:41,263 - INFO - [Train] step: 415199, loss: 0.005366, lr: 0.000020
|
| 177 |
+
2023-03-15 11:04:48,711 - INFO - [Train] step: 415599, loss: 0.002271, lr: 0.000020
|
| 178 |
+
2023-03-15 11:12:56,165 - INFO - [Train] step: 415999, loss: 0.006292, lr: 0.000020
|
| 179 |
+
2023-03-15 11:21:01,602 - INFO - [Train] step: 416399, loss: 0.017095, lr: 0.000020
|
| 180 |
+
2023-03-15 11:29:10,034 - INFO - [Train] step: 416799, loss: 0.002979, lr: 0.000020
|
| 181 |
+
2023-03-15 11:37:20,333 - INFO - [Train] step: 417199, loss: 0.009114, lr: 0.000020
|
| 182 |
+
2023-03-15 11:45:28,298 - INFO - [Train] step: 417599, loss: 0.011359, lr: 0.000020
|
| 183 |
+
2023-03-15 11:53:35,986 - INFO - [Train] step: 417999, loss: 0.002857, lr: 0.000020
|
| 184 |
+
2023-03-15 12:01:44,430 - INFO - [Train] step: 418399, loss: 0.010854, lr: 0.000020
|
| 185 |
+
2023-03-15 12:09:53,071 - INFO - [Train] step: 418799, loss: 0.004510, lr: 0.000020
|
| 186 |
+
2023-03-15 12:18:00,028 - INFO - [Train] step: 419199, loss: 0.006313, lr: 0.000020
|
| 187 |
+
2023-03-15 12:26:08,143 - INFO - [Train] step: 419599, loss: 0.035875, lr: 0.000020
|
| 188 |
+
2023-03-15 12:34:16,395 - INFO - [Train] step: 419999, loss: 0.012615, lr: 0.000020
|
| 189 |
+
2023-03-15 12:49:34,479 - INFO - [Train] step: 420399, loss: 0.005185, lr: 0.000020
|
| 190 |
+
2023-03-15 12:57:43,078 - INFO - [Train] step: 420799, loss: 0.006211, lr: 0.000020
|
| 191 |
+
2023-03-15 13:05:51,940 - INFO - [Train] step: 421199, loss: 0.001979, lr: 0.000020
|
| 192 |
+
2023-03-15 13:14:01,142 - INFO - [Train] step: 421599, loss: 0.009351, lr: 0.000020
|
| 193 |
+
2023-03-15 13:22:10,572 - INFO - [Train] step: 421999, loss: 0.002577, lr: 0.000020
|
| 194 |
+
2023-03-15 13:30:25,241 - INFO - [Train] step: 422399, loss: 0.002945, lr: 0.000020
|
| 195 |
+
2023-03-15 13:38:37,333 - INFO - [Train] step: 422799, loss: 0.018145, lr: 0.000020
|
| 196 |
+
2023-03-15 13:46:46,695 - INFO - [Train] step: 423199, loss: 0.004967, lr: 0.000020
|
| 197 |
+
2023-03-15 13:54:58,460 - INFO - [Train] step: 423599, loss: 0.006262, lr: 0.000020
|
| 198 |
+
2023-03-15 14:03:09,617 - INFO - [Train] step: 423999, loss: 0.005094, lr: 0.000020
|
| 199 |
+
2023-03-15 14:11:19,840 - INFO - [Train] step: 424399, loss: 0.006577, lr: 0.000020
|
| 200 |
+
2023-03-15 14:19:28,474 - INFO - [Train] step: 424799, loss: 0.007523, lr: 0.000020
|
| 201 |
+
2023-03-15 14:34:43,997 - INFO - [Train] step: 425199, loss: 0.011514, lr: 0.000020
|
| 202 |
+
2023-03-15 14:42:55,049 - INFO - [Train] step: 425599, loss: 0.004693, lr: 0.000020
|
| 203 |
+
2023-03-15 14:51:04,984 - INFO - [Train] step: 425999, loss: 0.004831, lr: 0.000020
|
| 204 |
+
2023-03-15 14:59:15,447 - INFO - [Train] step: 426399, loss: 0.010705, lr: 0.000020
|
| 205 |
+
2023-03-15 15:07:27,797 - INFO - [Train] step: 426799, loss: 0.003160, lr: 0.000020
|
| 206 |
+
2023-03-15 15:15:41,731 - INFO - [Train] step: 427199, loss: 0.002188, lr: 0.000020
|
| 207 |
+
2023-03-15 15:23:55,124 - INFO - [Train] step: 427599, loss: 0.004224, lr: 0.000020
|
| 208 |
+
2023-03-15 15:32:08,237 - INFO - [Train] step: 427999, loss: 0.009060, lr: 0.000020
|
| 209 |
+
2023-03-15 15:40:22,339 - INFO - [Train] step: 428399, loss: 0.011026, lr: 0.000020
|
| 210 |
+
2023-03-15 15:48:33,841 - INFO - [Train] step: 428799, loss: 0.022159, lr: 0.000020
|
| 211 |
+
2023-03-15 15:56:42,414 - INFO - [Train] step: 429199, loss: 0.013550, lr: 0.000020
|
| 212 |
+
2023-03-15 16:04:51,205 - INFO - [Train] step: 429599, loss: 0.008044, lr: 0.000020
|
| 213 |
+
2023-03-15 16:13:00,315 - INFO - [Train] step: 429999, loss: 0.011494, lr: 0.000020
|
| 214 |
+
2023-03-15 16:28:17,176 - INFO - [Train] step: 430399, loss: 0.002339, lr: 0.000020
|
| 215 |
+
2023-03-15 16:36:29,158 - INFO - [Train] step: 430799, loss: 0.011962, lr: 0.000020
|
| 216 |
+
2023-03-15 16:44:41,140 - INFO - [Train] step: 431199, loss: 0.006153, lr: 0.000020
|
| 217 |
+
2023-03-15 16:52:51,685 - INFO - [Train] step: 431599, loss: 0.002892, lr: 0.000020
|
| 218 |
+
2023-03-15 17:01:04,679 - INFO - [Train] step: 431999, loss: 0.007584, lr: 0.000020
|
| 219 |
+
2023-03-15 17:09:15,658 - INFO - [Train] step: 432399, loss: 0.005999, lr: 0.000020
|
| 220 |
+
2023-03-15 17:17:26,295 - INFO - [Train] step: 432799, loss: 0.005320, lr: 0.000020
|
| 221 |
+
2023-03-15 17:25:38,436 - INFO - [Train] step: 433199, loss: 0.003523, lr: 0.000020
|
| 222 |
+
2023-03-15 17:33:48,759 - INFO - [Train] step: 433599, loss: 0.001650, lr: 0.000020
|
| 223 |
+
2023-03-15 17:41:57,459 - INFO - [Train] step: 433999, loss: 0.002698, lr: 0.000020
|
| 224 |
+
2023-03-15 17:50:07,248 - INFO - [Train] step: 434399, loss: 0.008538, lr: 0.000020
|
| 225 |
+
2023-03-15 17:58:15,205 - INFO - [Train] step: 434799, loss: 0.007910, lr: 0.000020
|
| 226 |
+
2023-03-15 18:13:28,944 - INFO - [Train] step: 435199, loss: 0.002899, lr: 0.000020
|
| 227 |
+
2023-03-15 18:21:38,366 - INFO - [Train] step: 435599, loss: 0.004671, lr: 0.000020
|
| 228 |
+
2023-03-15 18:29:46,626 - INFO - [Train] step: 435999, loss: 0.004759, lr: 0.000020
|
| 229 |
+
2023-03-15 18:37:55,656 - INFO - [Train] step: 436399, loss: 0.003279, lr: 0.000020
|
| 230 |
+
2023-03-15 18:46:04,091 - INFO - [Train] step: 436799, loss: 0.001266, lr: 0.000020
|
| 231 |
+
2023-03-15 18:54:13,598 - INFO - [Train] step: 437199, loss: 0.002427, lr: 0.000020
|
| 232 |
+
2023-03-15 19:02:21,406 - INFO - [Train] step: 437599, loss: 0.004603, lr: 0.000020
|
| 233 |
+
2023-03-15 19:10:29,238 - INFO - [Train] step: 437999, loss: 0.017669, lr: 0.000020
|
| 234 |
+
2023-03-15 19:18:39,172 - INFO - [Train] step: 438399, loss: 0.006385, lr: 0.000020
|
| 235 |
+
2023-03-15 19:26:47,588 - INFO - [Train] step: 438799, loss: 0.005335, lr: 0.000020
|
| 236 |
+
2023-03-15 19:34:55,685 - INFO - [Train] step: 439199, loss: 0.003371, lr: 0.000020
|
| 237 |
+
2023-03-15 19:43:02,619 - INFO - [Train] step: 439599, loss: 0.004986, lr: 0.000020
|
| 238 |
+
2023-03-15 19:51:10,777 - INFO - [Train] step: 439999, loss: 0.012315, lr: 0.000020
|
| 239 |
+
2023-03-15 20:06:23,009 - INFO - [Train] step: 440399, loss: 0.008615, lr: 0.000020
|
| 240 |
+
2023-03-15 20:14:32,101 - INFO - [Train] step: 440799, loss: 0.009749, lr: 0.000020
|
| 241 |
+
2023-03-15 20:22:40,089 - INFO - [Train] step: 441199, loss: 0.021648, lr: 0.000020
|
| 242 |
+
2023-03-15 20:30:46,854 - INFO - [Train] step: 441599, loss: 0.002508, lr: 0.000020
|
| 243 |
+
2023-03-15 20:38:55,070 - INFO - [Train] step: 441999, loss: 0.004494, lr: 0.000020
|
| 244 |
+
2023-03-15 20:47:00,705 - INFO - [Train] step: 442399, loss: 0.008107, lr: 0.000020
|
| 245 |
+
2023-03-15 20:55:05,266 - INFO - [Train] step: 442799, loss: 0.013383, lr: 0.000020
|
| 246 |
+
2023-03-15 21:03:10,466 - INFO - [Train] step: 443199, loss: 0.004446, lr: 0.000020
|
| 247 |
+
2023-03-15 21:11:17,681 - INFO - [Train] step: 443599, loss: 0.006845, lr: 0.000020
|
| 248 |
+
2023-03-15 21:19:23,194 - INFO - [Train] step: 443999, loss: 0.005682, lr: 0.000020
|
| 249 |
+
2023-03-15 21:27:30,061 - INFO - [Train] step: 444399, loss: 0.003734, lr: 0.000020
|
| 250 |
+
2023-03-15 21:35:35,251 - INFO - [Train] step: 444799, loss: 0.005131, lr: 0.000020
|
| 251 |
+
2023-03-15 21:50:33,847 - INFO - [Train] step: 445199, loss: 0.002460, lr: 0.000020
|
| 252 |
+
2023-03-15 21:58:37,130 - INFO - [Train] step: 445599, loss: 0.023272, lr: 0.000020
|
| 253 |
+
2023-03-15 22:06:40,702 - INFO - [Train] step: 445999, loss: 0.010094, lr: 0.000020
|
| 254 |
+
2023-03-15 22:14:45,299 - INFO - [Train] step: 446399, loss: 0.008404, lr: 0.000020
|
| 255 |
+
2023-03-15 22:22:49,858 - INFO - [Train] step: 446799, loss: 0.020274, lr: 0.000020
|
| 256 |
+
2023-03-15 22:30:53,756 - INFO - [Train] step: 447199, loss: 0.002159, lr: 0.000020
|
| 257 |
+
2023-03-15 22:38:59,892 - INFO - [Train] step: 447599, loss: 0.005666, lr: 0.000020
|
| 258 |
+
2023-03-15 22:47:03,952 - INFO - [Train] step: 447999, loss: 0.001415, lr: 0.000020
|
| 259 |
+
2023-03-15 22:55:08,636 - INFO - [Train] step: 448399, loss: 0.005021, lr: 0.000020
|
| 260 |
+
2023-03-15 23:03:11,683 - INFO - [Train] step: 448799, loss: 0.020255, lr: 0.000020
|
| 261 |
+
2023-03-15 23:11:14,365 - INFO - [Train] step: 449199, loss: 0.004544, lr: 0.000020
|
| 262 |
+
2023-03-15 23:19:16,741 - INFO - [Train] step: 449599, loss: 0.009513, lr: 0.000020
|
| 263 |
+
2023-03-15 23:27:21,415 - INFO - [Train] step: 449999, loss: 0.009109, lr: 0.000020
|
| 264 |
+
2023-03-15 23:42:17,908 - INFO - [Train] step: 450399, loss: 0.004398, lr: 0.000020
|
| 265 |
+
2023-03-15 23:50:24,669 - INFO - [Train] step: 450799, loss: 0.002902, lr: 0.000020
|
| 266 |
+
2023-03-15 23:58:29,495 - INFO - [Train] step: 451199, loss: 0.028268, lr: 0.000020
|
| 267 |
+
2023-03-16 00:06:35,214 - INFO - [Train] step: 451599, loss: 0.004421, lr: 0.000020
|
| 268 |
+
2023-03-16 00:14:41,091 - INFO - [Train] step: 451999, loss: 0.012619, lr: 0.000020
|
| 269 |
+
2023-03-16 00:22:46,091 - INFO - [Train] step: 452399, loss: 0.005001, lr: 0.000020
|
| 270 |
+
2023-03-16 00:30:52,181 - INFO - [Train] step: 452799, loss: 0.007043, lr: 0.000020
|
| 271 |
+
2023-03-16 00:38:57,886 - INFO - [Train] step: 453199, loss: 0.003605, lr: 0.000020
|
| 272 |
+
2023-03-16 00:47:04,366 - INFO - [Train] step: 453599, loss: 0.004472, lr: 0.000020
|
| 273 |
+
2023-03-16 00:55:09,465 - INFO - [Train] step: 453999, loss: 0.004288, lr: 0.000020
|
| 274 |
+
2023-03-16 01:03:17,424 - INFO - [Train] step: 454399, loss: 0.003376, lr: 0.000020
|
| 275 |
+
2023-03-16 01:11:22,604 - INFO - [Train] step: 454799, loss: 0.006506, lr: 0.000020
|
| 276 |
+
2023-03-16 01:26:28,016 - INFO - [Train] step: 455199, loss: 0.003877, lr: 0.000020
|
| 277 |
+
2023-03-16 01:34:32,845 - INFO - [Train] step: 455599, loss: 0.005713, lr: 0.000020
|
| 278 |
+
2023-03-16 01:42:38,854 - INFO - [Train] step: 455999, loss: 0.009040, lr: 0.000020
|
| 279 |
+
2023-03-16 01:50:43,707 - INFO - [Train] step: 456399, loss: 0.021199, lr: 0.000020
|
| 280 |
+
2023-03-16 01:58:50,213 - INFO - [Train] step: 456799, loss: 0.007619, lr: 0.000020
|
| 281 |
+
2023-03-16 02:06:56,396 - INFO - [Train] step: 457199, loss: 0.004366, lr: 0.000020
|
| 282 |
+
2023-03-16 02:15:02,765 - INFO - [Train] step: 457599, loss: 0.002817, lr: 0.000020
|
| 283 |
+
2023-03-16 02:23:08,102 - INFO - [Train] step: 457999, loss: 0.012071, lr: 0.000020
|
| 284 |
+
2023-03-16 02:31:12,289 - INFO - [Train] step: 458399, loss: 0.014297, lr: 0.000020
|
| 285 |
+
2023-03-16 02:39:18,480 - INFO - [Train] step: 458799, loss: 0.006995, lr: 0.000020
|
| 286 |
+
2023-03-16 02:47:25,191 - INFO - [Train] step: 459199, loss: 0.011847, lr: 0.000020
|
| 287 |
+
2023-03-16 02:55:29,484 - INFO - [Train] step: 459599, loss: 0.003298, lr: 0.000020
|
| 288 |
+
2023-03-16 03:03:35,474 - INFO - [Train] step: 459999, loss: 0.005389, lr: 0.000020
|
| 289 |
+
2023-03-16 03:18:38,161 - INFO - [Train] step: 460399, loss: 0.003666, lr: 0.000020
|
| 290 |
+
2023-03-16 03:26:44,565 - INFO - [Train] step: 460799, loss: 0.007477, lr: 0.000020
|
| 291 |
+
2023-03-16 03:34:49,754 - INFO - [Train] step: 461199, loss: 0.010859, lr: 0.000020
|
| 292 |
+
2023-03-16 03:42:56,029 - INFO - [Train] step: 461599, loss: 0.014711, lr: 0.000020
|
| 293 |
+
2023-03-16 03:51:01,598 - INFO - [Train] step: 461999, loss: 0.009889, lr: 0.000020
|
| 294 |
+
2023-03-16 03:59:08,998 - INFO - [Train] step: 462399, loss: 0.005520, lr: 0.000020
|
| 295 |
+
2023-03-16 04:07:15,328 - INFO - [Train] step: 462799, loss: 0.002978, lr: 0.000020
|
| 296 |
+
2023-03-16 04:15:20,104 - INFO - [Train] step: 463199, loss: 0.022579, lr: 0.000020
|
| 297 |
+
2023-03-16 04:23:26,253 - INFO - [Train] step: 463599, loss: 0.009094, lr: 0.000020
|
| 298 |
+
2023-03-16 04:31:33,296 - INFO - [Train] step: 463999, loss: 0.002004, lr: 0.000020
|
| 299 |
+
2023-03-16 04:39:38,288 - INFO - [Train] step: 464399, loss: 0.007121, lr: 0.000020
|
| 300 |
+
2023-03-16 04:47:44,263 - INFO - [Train] step: 464799, loss: 0.003139, lr: 0.000020
|
| 301 |
+
2023-03-16 05:02:47,348 - INFO - [Train] step: 465199, loss: 0.010771, lr: 0.000020
|
| 302 |
+
2023-03-16 05:10:51,966 - INFO - [Train] step: 465599, loss: 0.003798, lr: 0.000020
|
| 303 |
+
2023-03-16 05:18:56,638 - INFO - [Train] step: 465999, loss: 0.004734, lr: 0.000020
|
| 304 |
+
2023-03-16 05:27:00,918 - INFO - [Train] step: 466399, loss: 0.004762, lr: 0.000020
|
| 305 |
+
2023-03-16 05:35:06,063 - INFO - [Train] step: 466799, loss: 0.023615, lr: 0.000020
|
| 306 |
+
2023-03-16 05:43:10,137 - INFO - [Train] step: 467199, loss: 0.024223, lr: 0.000020
|
| 307 |
+
2023-03-16 05:51:17,410 - INFO - [Train] step: 467599, loss: 0.001413, lr: 0.000020
|
| 308 |
+
2023-03-16 05:59:24,144 - INFO - [Train] step: 467999, loss: 0.001962, lr: 0.000020
|
| 309 |
+
2023-03-16 06:07:28,937 - INFO - [Train] step: 468399, loss: 0.007139, lr: 0.000020
|
| 310 |
+
2023-03-16 06:15:33,704 - INFO - [Train] step: 468799, loss: 0.003064, lr: 0.000020
|
| 311 |
+
2023-03-16 06:23:39,295 - INFO - [Train] step: 469199, loss: 0.003509, lr: 0.000020
|
| 312 |
+
2023-03-16 06:31:45,330 - INFO - [Train] step: 469599, loss: 0.013686, lr: 0.000020
|
| 313 |
+
2023-03-16 06:39:50,019 - INFO - [Train] step: 469999, loss: 0.005026, lr: 0.000020
|
| 314 |
+
2023-03-16 06:54:51,967 - INFO - [Train] step: 470399, loss: 0.004154, lr: 0.000020
|
| 315 |
+
2023-03-16 07:02:55,661 - INFO - [Train] step: 470799, loss: 0.009365, lr: 0.000020
|
| 316 |
+
2023-03-16 07:11:00,984 - INFO - [Train] step: 471199, loss: 0.003761, lr: 0.000020
|
| 317 |
+
2023-03-16 07:19:06,055 - INFO - [Train] step: 471599, loss: 0.010502, lr: 0.000020
|
| 318 |
+
2023-03-16 07:27:10,143 - INFO - [Train] step: 471999, loss: 0.003188, lr: 0.000020
|
| 319 |
+
2023-03-16 07:35:16,150 - INFO - [Train] step: 472399, loss: 0.002208, lr: 0.000020
|
| 320 |
+
2023-03-16 07:43:22,565 - INFO - [Train] step: 472799, loss: 0.004596, lr: 0.000020
|
| 321 |
+
2023-03-16 07:51:26,933 - INFO - [Train] step: 473199, loss: 0.014811, lr: 0.000020
|
| 322 |
+
2023-03-16 07:59:30,771 - INFO - [Train] step: 473599, loss: 0.002519, lr: 0.000020
|
| 323 |
+
2023-03-16 08:07:33,520 - INFO - [Train] step: 473999, loss: 0.010115, lr: 0.000020
|
| 324 |
+
2023-03-16 08:15:36,006 - INFO - [Train] step: 474399, loss: 0.025330, lr: 0.000020
|
| 325 |
+
2023-03-16 08:23:40,546 - INFO - [Train] step: 474799, loss: 0.010386, lr: 0.000020
|
| 326 |
+
2023-03-16 08:38:37,687 - INFO - [Train] step: 475199, loss: 0.002988, lr: 0.000020
|
| 327 |
+
2023-03-16 08:46:41,108 - INFO - [Train] step: 475599, loss: 0.010869, lr: 0.000020
|
| 328 |
+
2023-03-16 08:54:46,347 - INFO - [Train] step: 475999, loss: 0.002006, lr: 0.000020
|
| 329 |
+
2023-03-16 09:02:50,254 - INFO - [Train] step: 476399, loss: 0.017607, lr: 0.000020
|
| 330 |
+
2023-03-16 09:10:56,286 - INFO - [Train] step: 476799, loss: 0.002159, lr: 0.000020
|
| 331 |
+
2023-03-16 09:19:04,915 - INFO - [Train] step: 477199, loss: 0.011133, lr: 0.000020
|
| 332 |
+
2023-03-16 09:27:09,084 - INFO - [Train] step: 477599, loss: 0.009595, lr: 0.000020
|
| 333 |
+
2023-03-16 09:35:14,138 - INFO - [Train] step: 477999, loss: 0.023977, lr: 0.000020
|
| 334 |
+
2023-03-16 09:43:21,207 - INFO - [Train] step: 478399, loss: 0.003024, lr: 0.000020
|
| 335 |
+
2023-03-16 09:51:25,775 - INFO - [Train] step: 478799, loss: 0.003689, lr: 0.000020
|
| 336 |
+
2023-03-16 09:59:30,078 - INFO - [Train] step: 479199, loss: 0.006851, lr: 0.000020
|
| 337 |
+
2023-03-16 10:07:35,651 - INFO - [Train] step: 479599, loss: 0.005116, lr: 0.000020
|
| 338 |
+
2023-03-16 10:15:43,199 - INFO - [Train] step: 479999, loss: 0.002679, lr: 0.000020
|
| 339 |
+
2023-03-16 10:30:48,719 - INFO - [Train] step: 480399, loss: 0.002056, lr: 0.000020
|
| 340 |
+
2023-03-16 10:38:55,526 - INFO - [Train] step: 480799, loss: 0.005128, lr: 0.000020
|
| 341 |
+
2023-03-16 10:47:02,721 - INFO - [Train] step: 481199, loss: 0.009313, lr: 0.000020
|
| 342 |
+
2023-03-16 10:55:10,669 - INFO - [Train] step: 481599, loss: 0.004314, lr: 0.000020
|
| 343 |
+
2023-03-16 11:03:16,956 - INFO - [Train] step: 481999, loss: 0.003181, lr: 0.000020
|
| 344 |
+
2023-03-16 11:11:22,962 - INFO - [Train] step: 482399, loss: 0.010505, lr: 0.000020
|
| 345 |
+
2023-03-16 11:19:31,245 - INFO - [Train] step: 482799, loss: 0.005266, lr: 0.000020
|
| 346 |
+
2023-03-16 11:27:37,856 - INFO - [Train] step: 483199, loss: 0.001510, lr: 0.000020
|
| 347 |
+
2023-03-16 11:35:44,707 - INFO - [Train] step: 483599, loss: 0.017742, lr: 0.000020
|
| 348 |
+
2023-03-16 11:43:51,238 - INFO - [Train] step: 483999, loss: 0.014786, lr: 0.000020
|
| 349 |
+
2023-03-16 11:51:57,390 - INFO - [Train] step: 484399, loss: 0.003933, lr: 0.000020
|
| 350 |
+
2023-03-16 12:00:03,887 - INFO - [Train] step: 484799, loss: 0.004019, lr: 0.000020
|
| 351 |
+
2023-03-16 12:15:11,303 - INFO - [Train] step: 485199, loss: 0.014691, lr: 0.000020
|
| 352 |
+
2023-03-16 12:23:18,857 - INFO - [Train] step: 485599, loss: 0.012383, lr: 0.000020
|
| 353 |
+
2023-03-16 12:31:26,071 - INFO - [Train] step: 485999, loss: 0.034199, lr: 0.000020
|
| 354 |
+
2023-03-16 12:39:33,907 - INFO - [Train] step: 486399, loss: 0.009188, lr: 0.000020
|
| 355 |
+
2023-03-16 12:47:38,516 - INFO - [Train] step: 486799, loss: 0.007082, lr: 0.000020
|
| 356 |
+
2023-03-16 12:55:46,119 - INFO - [Train] step: 487199, loss: 0.008936, lr: 0.000020
|
| 357 |
+
2023-03-16 13:03:53,511 - INFO - [Train] step: 487599, loss: 0.007188, lr: 0.000020
|
| 358 |
+
2023-03-16 13:12:04,124 - INFO - [Train] step: 487999, loss: 0.026298, lr: 0.000020
|
| 359 |
+
2023-03-16 13:20:14,252 - INFO - [Train] step: 488399, loss: 0.001881, lr: 0.000020
|
| 360 |
+
2023-03-16 13:28:23,608 - INFO - [Train] step: 488799, loss: 0.011646, lr: 0.000020
|
| 361 |
+
2023-03-16 13:36:32,847 - INFO - [Train] step: 489199, loss: 0.006114, lr: 0.000020
|
| 362 |
+
2023-03-16 13:44:43,367 - INFO - [Train] step: 489599, loss: 0.011290, lr: 0.000020
|
| 363 |
+
2023-03-16 13:52:51,516 - INFO - [Train] step: 489999, loss: 0.004321, lr: 0.000020
|
| 364 |
+
2023-03-16 14:08:05,248 - INFO - [Train] step: 490399, loss: 0.001973, lr: 0.000020
|
| 365 |
+
2023-03-16 14:16:14,008 - INFO - [Train] step: 490799, loss: 0.009576, lr: 0.000020
|
| 366 |
+
2023-03-16 14:24:24,099 - INFO - [Train] step: 491199, loss: 0.011714, lr: 0.000020
|
| 367 |
+
2023-03-16 14:32:32,477 - INFO - [Train] step: 491599, loss: 0.017272, lr: 0.000020
|
| 368 |
+
2023-03-16 14:40:41,678 - INFO - [Train] step: 491999, loss: 0.003934, lr: 0.000020
|
| 369 |
+
2023-03-16 14:48:50,762 - INFO - [Train] step: 492399, loss: 0.004091, lr: 0.000020
|
| 370 |
+
2023-03-16 14:56:59,591 - INFO - [Train] step: 492799, loss: 0.002672, lr: 0.000020
|
| 371 |
+
2023-03-16 15:05:06,727 - INFO - [Train] step: 493199, loss: 0.002655, lr: 0.000020
|
| 372 |
+
2023-03-16 15:13:12,498 - INFO - [Train] step: 493599, loss: 0.003188, lr: 0.000020
|
| 373 |
+
2023-03-16 15:21:20,084 - INFO - [Train] step: 493999, loss: 0.008530, lr: 0.000020
|
| 374 |
+
2023-03-16 15:29:29,480 - INFO - [Train] step: 494399, loss: 0.037123, lr: 0.000020
|
| 375 |
+
2023-03-16 15:37:37,057 - INFO - [Train] step: 494799, loss: 0.017426, lr: 0.000020
|
| 376 |
+
2023-03-16 15:52:40,043 - INFO - [Train] step: 495199, loss: 0.007916, lr: 0.000020
|
| 377 |
+
2023-03-16 16:00:46,491 - INFO - [Train] step: 495599, loss: 0.006630, lr: 0.000020
|
| 378 |
+
2023-03-16 16:08:50,627 - INFO - [Train] step: 495999, loss: 0.006427, lr: 0.000020
|
| 379 |
+
2023-03-16 16:16:55,630 - INFO - [Train] step: 496399, loss: 0.002252, lr: 0.000020
|
| 380 |
+
2023-03-16 16:24:59,976 - INFO - [Train] step: 496799, loss: 0.003233, lr: 0.000020
|
| 381 |
+
2023-03-16 16:33:04,518 - INFO - [Train] step: 497199, loss: 0.002532, lr: 0.000020
|
| 382 |
+
2023-03-16 16:41:08,627 - INFO - [Train] step: 497599, loss: 0.013784, lr: 0.000020
|
| 383 |
+
2023-03-16 16:49:12,418 - INFO - [Train] step: 497999, loss: 0.007258, lr: 0.000020
|
| 384 |
+
2023-03-16 16:57:17,603 - INFO - [Train] step: 498399, loss: 0.008979, lr: 0.000020
|
| 385 |
+
2023-03-16 17:05:23,798 - INFO - [Train] step: 498799, loss: 0.005983, lr: 0.000020
|
| 386 |
+
2023-03-16 17:13:31,231 - INFO - [Train] step: 499199, loss: 0.003115, lr: 0.000020
|
| 387 |
+
2023-03-16 17:21:37,292 - INFO - [Train] step: 499599, loss: 0.006188, lr: 0.000020
|
| 388 |
+
2023-03-16 17:29:44,844 - INFO - [Train] step: 499999, loss: 0.004860, lr: 0.000020
|
| 389 |
+
2023-03-16 17:36:45,308 - INFO - End of training
|
ddpm_celebahq/samples.zip
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:fce4d45547ea72a6286292611e531a833432d46f1874aa0e0b1ae5df412f24a5
|
| 3 |
+
size 318653278
|
ddpm_celebahq/tensorboard/events.out.tfevents.1678369865.boot-SYS-4029GP-TRT.222806.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2bb546a4bc82156df7c06008d590ed080e5d165f245107c14c11785eb1bbf2e4
|
| 3 |
+
size 34673516
|
ddpm_celebahq/tensorboard/events.out.tfevents.1678763820.admin.cluster.local.39024.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1f65aebea20d74f6d53b214b6663d2fca4bcd04c47a18f1db1625970113ac2f3
|
| 3 |
+
size 14700040
|