craa commited on
Commit
2740eb9
·
verified ·
1 Parent(s): 40b9871

Training in progress, step 20000, checkpoint

Browse files
checkpoint-20000/config.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "activation_function": "gelu_new",
3
+ "architectures": [
4
+ "GPT2LMHeadModel"
5
+ ],
6
+ "attn_pdrop": 0.1,
7
+ "bos_token_id": 50256,
8
+ "embd_pdrop": 0.1,
9
+ "eos_token_id": 50256,
10
+ "initializer_range": 0.02,
11
+ "layer_norm_epsilon": 1e-05,
12
+ "model_type": "gpt2",
13
+ "n_embd": 768,
14
+ "n_head": 12,
15
+ "n_inner": null,
16
+ "n_layer": 12,
17
+ "n_positions": 1024,
18
+ "reorder_and_upcast_attn": false,
19
+ "resid_pdrop": 0.1,
20
+ "scale_attn_by_inverse_layer_idx": false,
21
+ "scale_attn_weights": true,
22
+ "summary_activation": null,
23
+ "summary_first_dropout": 0.1,
24
+ "summary_proj_to_labels": true,
25
+ "summary_type": "cls_index",
26
+ "summary_use_proj": true,
27
+ "torch_dtype": "float32",
28
+ "transformers_version": "4.47.0.dev0",
29
+ "use_cache": true,
30
+ "vocab_size": 52000
31
+ }
checkpoint-20000/generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 50256,
4
+ "eos_token_id": 50256,
5
+ "transformers_version": "4.47.0.dev0"
6
+ }
checkpoint-20000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:57c1c4bf48756413da7129d8d5dfc187fa90a4491440d1b1a26e1e8d09a995a6
3
+ size 503128704
checkpoint-20000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:015f03ab4a2ff0123b56a2dcfe16eac2a8a05cf660124eae3dcf8ebc804f3ff8
3
+ size 1006351290
checkpoint-20000/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6f445e5735bba08c88e65ab96aae78ec298bb49c43b3cf66e98cb0d63aad8d72
3
+ size 14244
checkpoint-20000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4081d50f1887f3a982b2f3cabb2e58590020a0424181d8b5b776143e4ebaaf72
3
+ size 1064
checkpoint-20000/special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {}
checkpoint-20000/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-20000/tokenizer_config.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-20000/trainer_state.json ADDED
@@ -0,0 +1,3013 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 3.573341131210327,
3
+ "best_model_checkpoint": "/scratch/cl5625/exceptions/models/100M__8397/checkpoint-20000",
4
+ "epoch": 2.1563342318059298,
5
+ "eval_steps": 1000,
6
+ "global_step": 20000,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.005390835579514825,
13
+ "grad_norm": 1.9401150941848755,
14
+ "learning_rate": 0.0003,
15
+ "loss": 8.6351,
16
+ "step": 50
17
+ },
18
+ {
19
+ "epoch": 0.01078167115902965,
20
+ "grad_norm": 1.2623108625411987,
21
+ "learning_rate": 0.0006,
22
+ "loss": 6.8864,
23
+ "step": 100
24
+ },
25
+ {
26
+ "epoch": 0.016172506738544475,
27
+ "grad_norm": 3.276719808578491,
28
+ "learning_rate": 0.0005996762007555315,
29
+ "loss": 6.4636,
30
+ "step": 150
31
+ },
32
+ {
33
+ "epoch": 0.0215633423180593,
34
+ "grad_norm": 2.706618309020996,
35
+ "learning_rate": 0.000599352401511063,
36
+ "loss": 6.2398,
37
+ "step": 200
38
+ },
39
+ {
40
+ "epoch": 0.026954177897574125,
41
+ "grad_norm": 1.0366517305374146,
42
+ "learning_rate": 0.0005990286022665946,
43
+ "loss": 6.0879,
44
+ "step": 250
45
+ },
46
+ {
47
+ "epoch": 0.03234501347708895,
48
+ "grad_norm": 2.069812297821045,
49
+ "learning_rate": 0.0005987048030221263,
50
+ "loss": 5.9796,
51
+ "step": 300
52
+ },
53
+ {
54
+ "epoch": 0.03773584905660377,
55
+ "grad_norm": 1.1950629949569702,
56
+ "learning_rate": 0.0005983810037776578,
57
+ "loss": 5.8688,
58
+ "step": 350
59
+ },
60
+ {
61
+ "epoch": 0.0431266846361186,
62
+ "grad_norm": 1.0007357597351074,
63
+ "learning_rate": 0.0005980572045331894,
64
+ "loss": 5.8233,
65
+ "step": 400
66
+ },
67
+ {
68
+ "epoch": 0.04851752021563342,
69
+ "grad_norm": 2.0554916858673096,
70
+ "learning_rate": 0.0005977334052887209,
71
+ "loss": 5.7452,
72
+ "step": 450
73
+ },
74
+ {
75
+ "epoch": 0.05390835579514825,
76
+ "grad_norm": 1.996433973312378,
77
+ "learning_rate": 0.0005974096060442526,
78
+ "loss": 5.6484,
79
+ "step": 500
80
+ },
81
+ {
82
+ "epoch": 0.05929919137466307,
83
+ "grad_norm": 1.439592719078064,
84
+ "learning_rate": 0.0005970858067997841,
85
+ "loss": 5.5734,
86
+ "step": 550
87
+ },
88
+ {
89
+ "epoch": 0.0646900269541779,
90
+ "grad_norm": 1.0077263116836548,
91
+ "learning_rate": 0.0005967620075553157,
92
+ "loss": 5.4903,
93
+ "step": 600
94
+ },
95
+ {
96
+ "epoch": 0.07008086253369272,
97
+ "grad_norm": 1.460601568222046,
98
+ "learning_rate": 0.0005964382083108472,
99
+ "loss": 5.4273,
100
+ "step": 650
101
+ },
102
+ {
103
+ "epoch": 0.07547169811320754,
104
+ "grad_norm": 1.5479509830474854,
105
+ "learning_rate": 0.0005961144090663788,
106
+ "loss": 5.3803,
107
+ "step": 700
108
+ },
109
+ {
110
+ "epoch": 0.08086253369272237,
111
+ "grad_norm": 1.2044662237167358,
112
+ "learning_rate": 0.0005957906098219104,
113
+ "loss": 5.305,
114
+ "step": 750
115
+ },
116
+ {
117
+ "epoch": 0.0862533692722372,
118
+ "grad_norm": 1.4456267356872559,
119
+ "learning_rate": 0.0005954668105774419,
120
+ "loss": 5.2577,
121
+ "step": 800
122
+ },
123
+ {
124
+ "epoch": 0.09164420485175202,
125
+ "grad_norm": 1.4306107759475708,
126
+ "learning_rate": 0.0005951430113329735,
127
+ "loss": 5.2168,
128
+ "step": 850
129
+ },
130
+ {
131
+ "epoch": 0.09703504043126684,
132
+ "grad_norm": 1.5723680257797241,
133
+ "learning_rate": 0.0005948192120885051,
134
+ "loss": 5.1595,
135
+ "step": 900
136
+ },
137
+ {
138
+ "epoch": 0.10242587601078167,
139
+ "grad_norm": 1.4030689001083374,
140
+ "learning_rate": 0.0005944954128440366,
141
+ "loss": 5.1322,
142
+ "step": 950
143
+ },
144
+ {
145
+ "epoch": 0.1078167115902965,
146
+ "grad_norm": 0.9765119552612305,
147
+ "learning_rate": 0.0005941716135995682,
148
+ "loss": 5.0934,
149
+ "step": 1000
150
+ },
151
+ {
152
+ "epoch": 0.1078167115902965,
153
+ "eval_accuracy": 0.22754229014862737,
154
+ "eval_loss": 5.022522926330566,
155
+ "eval_runtime": 185.9242,
156
+ "eval_samples_per_second": 96.873,
157
+ "eval_steps_per_second": 6.056,
158
+ "step": 1000
159
+ },
160
+ {
161
+ "epoch": 0.11320754716981132,
162
+ "grad_norm": 0.8278682827949524,
163
+ "learning_rate": 0.0005938478143550997,
164
+ "loss": 5.0442,
165
+ "step": 1050
166
+ },
167
+ {
168
+ "epoch": 0.11859838274932614,
169
+ "grad_norm": 1.0204647779464722,
170
+ "learning_rate": 0.0005935240151106314,
171
+ "loss": 5.0209,
172
+ "step": 1100
173
+ },
174
+ {
175
+ "epoch": 0.12398921832884097,
176
+ "grad_norm": 1.004250407218933,
177
+ "learning_rate": 0.0005932002158661629,
178
+ "loss": 4.9915,
179
+ "step": 1150
180
+ },
181
+ {
182
+ "epoch": 0.1293800539083558,
183
+ "grad_norm": 1.2949087619781494,
184
+ "learning_rate": 0.0005928764166216945,
185
+ "loss": 4.9493,
186
+ "step": 1200
187
+ },
188
+ {
189
+ "epoch": 0.1347708894878706,
190
+ "grad_norm": 0.8809250593185425,
191
+ "learning_rate": 0.000592552617377226,
192
+ "loss": 4.9039,
193
+ "step": 1250
194
+ },
195
+ {
196
+ "epoch": 0.14016172506738545,
197
+ "grad_norm": 1.0270975828170776,
198
+ "learning_rate": 0.0005922288181327577,
199
+ "loss": 4.8792,
200
+ "step": 1300
201
+ },
202
+ {
203
+ "epoch": 0.14555256064690028,
204
+ "grad_norm": 0.9649767279624939,
205
+ "learning_rate": 0.0005919050188882893,
206
+ "loss": 4.8622,
207
+ "step": 1350
208
+ },
209
+ {
210
+ "epoch": 0.1509433962264151,
211
+ "grad_norm": 0.9902446866035461,
212
+ "learning_rate": 0.0005915812196438207,
213
+ "loss": 4.8481,
214
+ "step": 1400
215
+ },
216
+ {
217
+ "epoch": 0.15633423180592992,
218
+ "grad_norm": 0.7882956862449646,
219
+ "learning_rate": 0.0005912574203993524,
220
+ "loss": 4.7968,
221
+ "step": 1450
222
+ },
223
+ {
224
+ "epoch": 0.16172506738544473,
225
+ "grad_norm": 0.8941258788108826,
226
+ "learning_rate": 0.0005909336211548839,
227
+ "loss": 4.8149,
228
+ "step": 1500
229
+ },
230
+ {
231
+ "epoch": 0.16711590296495957,
232
+ "grad_norm": 1.3297462463378906,
233
+ "learning_rate": 0.0005906098219104155,
234
+ "loss": 4.7656,
235
+ "step": 1550
236
+ },
237
+ {
238
+ "epoch": 0.1725067385444744,
239
+ "grad_norm": 1.497783899307251,
240
+ "learning_rate": 0.000590286022665947,
241
+ "loss": 4.7642,
242
+ "step": 1600
243
+ },
244
+ {
245
+ "epoch": 0.1778975741239892,
246
+ "grad_norm": 1.0149500370025635,
247
+ "learning_rate": 0.0005899622234214787,
248
+ "loss": 4.7241,
249
+ "step": 1650
250
+ },
251
+ {
252
+ "epoch": 0.18328840970350405,
253
+ "grad_norm": 1.067275047302246,
254
+ "learning_rate": 0.0005896384241770102,
255
+ "loss": 4.7115,
256
+ "step": 1700
257
+ },
258
+ {
259
+ "epoch": 0.18867924528301888,
260
+ "grad_norm": 0.7510073781013489,
261
+ "learning_rate": 0.0005893146249325418,
262
+ "loss": 4.6734,
263
+ "step": 1750
264
+ },
265
+ {
266
+ "epoch": 0.1940700808625337,
267
+ "grad_norm": 1.0857677459716797,
268
+ "learning_rate": 0.0005889908256880733,
269
+ "loss": 4.6802,
270
+ "step": 1800
271
+ },
272
+ {
273
+ "epoch": 0.19946091644204852,
274
+ "grad_norm": 0.9117745161056519,
275
+ "learning_rate": 0.0005886670264436049,
276
+ "loss": 4.6481,
277
+ "step": 1850
278
+ },
279
+ {
280
+ "epoch": 0.20485175202156333,
281
+ "grad_norm": 0.9455132484436035,
282
+ "learning_rate": 0.0005883432271991365,
283
+ "loss": 4.6385,
284
+ "step": 1900
285
+ },
286
+ {
287
+ "epoch": 0.21024258760107817,
288
+ "grad_norm": 0.8344852924346924,
289
+ "learning_rate": 0.0005880194279546681,
290
+ "loss": 4.6113,
291
+ "step": 1950
292
+ },
293
+ {
294
+ "epoch": 0.215633423180593,
295
+ "grad_norm": 1.0357882976531982,
296
+ "learning_rate": 0.0005876956287101996,
297
+ "loss": 4.5937,
298
+ "step": 2000
299
+ },
300
+ {
301
+ "epoch": 0.215633423180593,
302
+ "eval_accuracy": 0.27015801715433524,
303
+ "eval_loss": 4.516484260559082,
304
+ "eval_runtime": 185.6539,
305
+ "eval_samples_per_second": 97.014,
306
+ "eval_steps_per_second": 6.065,
307
+ "step": 2000
308
+ },
309
+ {
310
+ "epoch": 0.2210242587601078,
311
+ "grad_norm": 0.6943921446800232,
312
+ "learning_rate": 0.0005873718294657312,
313
+ "loss": 4.5747,
314
+ "step": 2050
315
+ },
316
+ {
317
+ "epoch": 0.22641509433962265,
318
+ "grad_norm": 0.8243718147277832,
319
+ "learning_rate": 0.0005870480302212628,
320
+ "loss": 4.5582,
321
+ "step": 2100
322
+ },
323
+ {
324
+ "epoch": 0.23180592991913745,
325
+ "grad_norm": 1.0496394634246826,
326
+ "learning_rate": 0.0005867242309767943,
327
+ "loss": 4.5315,
328
+ "step": 2150
329
+ },
330
+ {
331
+ "epoch": 0.2371967654986523,
332
+ "grad_norm": 0.8335784077644348,
333
+ "learning_rate": 0.0005864004317323259,
334
+ "loss": 4.5086,
335
+ "step": 2200
336
+ },
337
+ {
338
+ "epoch": 0.24258760107816713,
339
+ "grad_norm": 0.8698639273643494,
340
+ "learning_rate": 0.0005860766324878575,
341
+ "loss": 4.4958,
342
+ "step": 2250
343
+ },
344
+ {
345
+ "epoch": 0.24797843665768193,
346
+ "grad_norm": 0.7829259634017944,
347
+ "learning_rate": 0.000585752833243389,
348
+ "loss": 4.5067,
349
+ "step": 2300
350
+ },
351
+ {
352
+ "epoch": 0.25336927223719674,
353
+ "grad_norm": 0.8259062767028809,
354
+ "learning_rate": 0.0005854290339989206,
355
+ "loss": 4.4569,
356
+ "step": 2350
357
+ },
358
+ {
359
+ "epoch": 0.2587601078167116,
360
+ "grad_norm": 0.8445196151733398,
361
+ "learning_rate": 0.0005851052347544521,
362
+ "loss": 4.4611,
363
+ "step": 2400
364
+ },
365
+ {
366
+ "epoch": 0.2641509433962264,
367
+ "grad_norm": 0.8523595929145813,
368
+ "learning_rate": 0.0005847814355099838,
369
+ "loss": 4.4402,
370
+ "step": 2450
371
+ },
372
+ {
373
+ "epoch": 0.2695417789757412,
374
+ "grad_norm": 0.8830829858779907,
375
+ "learning_rate": 0.0005844576362655154,
376
+ "loss": 4.4235,
377
+ "step": 2500
378
+ },
379
+ {
380
+ "epoch": 0.2749326145552561,
381
+ "grad_norm": 0.8365247845649719,
382
+ "learning_rate": 0.0005841338370210469,
383
+ "loss": 4.4101,
384
+ "step": 2550
385
+ },
386
+ {
387
+ "epoch": 0.2803234501347709,
388
+ "grad_norm": 0.853497326374054,
389
+ "learning_rate": 0.0005838100377765785,
390
+ "loss": 4.3826,
391
+ "step": 2600
392
+ },
393
+ {
394
+ "epoch": 0.2857142857142857,
395
+ "grad_norm": 0.9241225719451904,
396
+ "learning_rate": 0.0005834862385321101,
397
+ "loss": 4.3931,
398
+ "step": 2650
399
+ },
400
+ {
401
+ "epoch": 0.29110512129380056,
402
+ "grad_norm": 0.8265724778175354,
403
+ "learning_rate": 0.0005831624392876417,
404
+ "loss": 4.3887,
405
+ "step": 2700
406
+ },
407
+ {
408
+ "epoch": 0.29649595687331537,
409
+ "grad_norm": 0.8722224831581116,
410
+ "learning_rate": 0.0005828386400431731,
411
+ "loss": 4.3563,
412
+ "step": 2750
413
+ },
414
+ {
415
+ "epoch": 0.3018867924528302,
416
+ "grad_norm": 0.7775854468345642,
417
+ "learning_rate": 0.0005825148407987048,
418
+ "loss": 4.3271,
419
+ "step": 2800
420
+ },
421
+ {
422
+ "epoch": 0.30727762803234504,
423
+ "grad_norm": 0.7281326055526733,
424
+ "learning_rate": 0.0005821910415542363,
425
+ "loss": 4.3363,
426
+ "step": 2850
427
+ },
428
+ {
429
+ "epoch": 0.31266846361185985,
430
+ "grad_norm": 0.8523993492126465,
431
+ "learning_rate": 0.0005818672423097679,
432
+ "loss": 4.3284,
433
+ "step": 2900
434
+ },
435
+ {
436
+ "epoch": 0.31805929919137466,
437
+ "grad_norm": 0.8152425289154053,
438
+ "learning_rate": 0.0005815434430652994,
439
+ "loss": 4.313,
440
+ "step": 2950
441
+ },
442
+ {
443
+ "epoch": 0.32345013477088946,
444
+ "grad_norm": 0.9258395433425903,
445
+ "learning_rate": 0.0005812196438208311,
446
+ "loss": 4.3092,
447
+ "step": 3000
448
+ },
449
+ {
450
+ "epoch": 0.32345013477088946,
451
+ "eval_accuracy": 0.29783223363934697,
452
+ "eval_loss": 4.246051788330078,
453
+ "eval_runtime": 185.609,
454
+ "eval_samples_per_second": 97.037,
455
+ "eval_steps_per_second": 6.067,
456
+ "step": 3000
457
+ },
458
+ {
459
+ "epoch": 0.3288409703504043,
460
+ "grad_norm": 0.78269362449646,
461
+ "learning_rate": 0.0005808958445763626,
462
+ "loss": 4.304,
463
+ "step": 3050
464
+ },
465
+ {
466
+ "epoch": 0.33423180592991913,
467
+ "grad_norm": 0.7207179069519043,
468
+ "learning_rate": 0.0005805720453318942,
469
+ "loss": 4.3041,
470
+ "step": 3100
471
+ },
472
+ {
473
+ "epoch": 0.33962264150943394,
474
+ "grad_norm": 0.791796088218689,
475
+ "learning_rate": 0.0005802482460874257,
476
+ "loss": 4.2633,
477
+ "step": 3150
478
+ },
479
+ {
480
+ "epoch": 0.3450134770889488,
481
+ "grad_norm": 0.6603590250015259,
482
+ "learning_rate": 0.0005799244468429573,
483
+ "loss": 4.2613,
484
+ "step": 3200
485
+ },
486
+ {
487
+ "epoch": 0.3504043126684636,
488
+ "grad_norm": 0.922822892665863,
489
+ "learning_rate": 0.0005796006475984889,
490
+ "loss": 4.2617,
491
+ "step": 3250
492
+ },
493
+ {
494
+ "epoch": 0.3557951482479784,
495
+ "grad_norm": 0.7605053782463074,
496
+ "learning_rate": 0.0005792768483540205,
497
+ "loss": 4.2446,
498
+ "step": 3300
499
+ },
500
+ {
501
+ "epoch": 0.3611859838274933,
502
+ "grad_norm": 0.7679175734519958,
503
+ "learning_rate": 0.000578953049109552,
504
+ "loss": 4.26,
505
+ "step": 3350
506
+ },
507
+ {
508
+ "epoch": 0.3665768194070081,
509
+ "grad_norm": 0.7325921654701233,
510
+ "learning_rate": 0.0005786292498650836,
511
+ "loss": 4.2337,
512
+ "step": 3400
513
+ },
514
+ {
515
+ "epoch": 0.3719676549865229,
516
+ "grad_norm": 0.6152936816215515,
517
+ "learning_rate": 0.0005783054506206152,
518
+ "loss": 4.21,
519
+ "step": 3450
520
+ },
521
+ {
522
+ "epoch": 0.37735849056603776,
523
+ "grad_norm": 0.7148654460906982,
524
+ "learning_rate": 0.0005779816513761467,
525
+ "loss": 4.2203,
526
+ "step": 3500
527
+ },
528
+ {
529
+ "epoch": 0.38274932614555257,
530
+ "grad_norm": 0.8136675953865051,
531
+ "learning_rate": 0.0005776578521316782,
532
+ "loss": 4.2143,
533
+ "step": 3550
534
+ },
535
+ {
536
+ "epoch": 0.3881401617250674,
537
+ "grad_norm": 0.6423671245574951,
538
+ "learning_rate": 0.0005773340528872099,
539
+ "loss": 4.2164,
540
+ "step": 3600
541
+ },
542
+ {
543
+ "epoch": 0.3935309973045822,
544
+ "grad_norm": 0.8761276602745056,
545
+ "learning_rate": 0.0005770102536427414,
546
+ "loss": 4.1887,
547
+ "step": 3650
548
+ },
549
+ {
550
+ "epoch": 0.39892183288409705,
551
+ "grad_norm": 0.7816224694252014,
552
+ "learning_rate": 0.000576686454398273,
553
+ "loss": 4.1937,
554
+ "step": 3700
555
+ },
556
+ {
557
+ "epoch": 0.40431266846361186,
558
+ "grad_norm": 0.7834933400154114,
559
+ "learning_rate": 0.0005763626551538045,
560
+ "loss": 4.1965,
561
+ "step": 3750
562
+ },
563
+ {
564
+ "epoch": 0.40970350404312667,
565
+ "grad_norm": 0.6208221912384033,
566
+ "learning_rate": 0.0005760388559093362,
567
+ "loss": 4.1864,
568
+ "step": 3800
569
+ },
570
+ {
571
+ "epoch": 0.41509433962264153,
572
+ "grad_norm": 0.6872389316558838,
573
+ "learning_rate": 0.0005757150566648678,
574
+ "loss": 4.1732,
575
+ "step": 3850
576
+ },
577
+ {
578
+ "epoch": 0.42048517520215634,
579
+ "grad_norm": 0.7426034808158875,
580
+ "learning_rate": 0.0005753912574203993,
581
+ "loss": 4.1682,
582
+ "step": 3900
583
+ },
584
+ {
585
+ "epoch": 0.42587601078167114,
586
+ "grad_norm": 0.7201637625694275,
587
+ "learning_rate": 0.0005750674581759309,
588
+ "loss": 4.157,
589
+ "step": 3950
590
+ },
591
+ {
592
+ "epoch": 0.431266846361186,
593
+ "grad_norm": 0.7274760007858276,
594
+ "learning_rate": 0.0005747436589314624,
595
+ "loss": 4.1586,
596
+ "step": 4000
597
+ },
598
+ {
599
+ "epoch": 0.431266846361186,
600
+ "eval_accuracy": 0.3127459290207626,
601
+ "eval_loss": 4.088518142700195,
602
+ "eval_runtime": 185.6568,
603
+ "eval_samples_per_second": 97.012,
604
+ "eval_steps_per_second": 6.065,
605
+ "step": 4000
606
+ },
607
+ {
608
+ "epoch": 0.4366576819407008,
609
+ "grad_norm": 0.5464039444923401,
610
+ "learning_rate": 0.0005744198596869941,
611
+ "loss": 4.149,
612
+ "step": 4050
613
+ },
614
+ {
615
+ "epoch": 0.4420485175202156,
616
+ "grad_norm": 0.7344082593917847,
617
+ "learning_rate": 0.0005740960604425255,
618
+ "loss": 4.1532,
619
+ "step": 4100
620
+ },
621
+ {
622
+ "epoch": 0.4474393530997305,
623
+ "grad_norm": 0.6581411957740784,
624
+ "learning_rate": 0.0005737722611980572,
625
+ "loss": 4.1547,
626
+ "step": 4150
627
+ },
628
+ {
629
+ "epoch": 0.4528301886792453,
630
+ "grad_norm": 0.6639128923416138,
631
+ "learning_rate": 0.0005734484619535887,
632
+ "loss": 4.1367,
633
+ "step": 4200
634
+ },
635
+ {
636
+ "epoch": 0.4582210242587601,
637
+ "grad_norm": 0.734761655330658,
638
+ "learning_rate": 0.0005731246627091203,
639
+ "loss": 4.1221,
640
+ "step": 4250
641
+ },
642
+ {
643
+ "epoch": 0.4636118598382749,
644
+ "grad_norm": 0.8860589265823364,
645
+ "learning_rate": 0.0005728008634646518,
646
+ "loss": 4.1218,
647
+ "step": 4300
648
+ },
649
+ {
650
+ "epoch": 0.46900269541778977,
651
+ "grad_norm": 0.5995060205459595,
652
+ "learning_rate": 0.0005724770642201835,
653
+ "loss": 4.1167,
654
+ "step": 4350
655
+ },
656
+ {
657
+ "epoch": 0.4743935309973046,
658
+ "grad_norm": 0.7229450345039368,
659
+ "learning_rate": 0.000572153264975715,
660
+ "loss": 4.1148,
661
+ "step": 4400
662
+ },
663
+ {
664
+ "epoch": 0.4797843665768194,
665
+ "grad_norm": 0.731573224067688,
666
+ "learning_rate": 0.0005718294657312466,
667
+ "loss": 4.1018,
668
+ "step": 4450
669
+ },
670
+ {
671
+ "epoch": 0.48517520215633425,
672
+ "grad_norm": 0.7716799974441528,
673
+ "learning_rate": 0.0005715056664867781,
674
+ "loss": 4.109,
675
+ "step": 4500
676
+ },
677
+ {
678
+ "epoch": 0.49056603773584906,
679
+ "grad_norm": 0.542386531829834,
680
+ "learning_rate": 0.0005711818672423097,
681
+ "loss": 4.1017,
682
+ "step": 4550
683
+ },
684
+ {
685
+ "epoch": 0.49595687331536387,
686
+ "grad_norm": 0.794573187828064,
687
+ "learning_rate": 0.0005708580679978413,
688
+ "loss": 4.086,
689
+ "step": 4600
690
+ },
691
+ {
692
+ "epoch": 0.5013477088948787,
693
+ "grad_norm": 0.6180745959281921,
694
+ "learning_rate": 0.0005705342687533729,
695
+ "loss": 4.0787,
696
+ "step": 4650
697
+ },
698
+ {
699
+ "epoch": 0.5067385444743935,
700
+ "grad_norm": 0.6277428269386292,
701
+ "learning_rate": 0.0005702104695089044,
702
+ "loss": 4.0932,
703
+ "step": 4700
704
+ },
705
+ {
706
+ "epoch": 0.5121293800539084,
707
+ "grad_norm": 0.6428430676460266,
708
+ "learning_rate": 0.000569886670264436,
709
+ "loss": 4.0812,
710
+ "step": 4750
711
+ },
712
+ {
713
+ "epoch": 0.5175202156334232,
714
+ "grad_norm": 0.6423416137695312,
715
+ "learning_rate": 0.0005695628710199675,
716
+ "loss": 4.0791,
717
+ "step": 4800
718
+ },
719
+ {
720
+ "epoch": 0.522911051212938,
721
+ "grad_norm": 0.6095171570777893,
722
+ "learning_rate": 0.0005692390717754991,
723
+ "loss": 4.0655,
724
+ "step": 4850
725
+ },
726
+ {
727
+ "epoch": 0.5283018867924528,
728
+ "grad_norm": 0.6891219615936279,
729
+ "learning_rate": 0.0005689152725310306,
730
+ "loss": 4.0527,
731
+ "step": 4900
732
+ },
733
+ {
734
+ "epoch": 0.5336927223719676,
735
+ "grad_norm": 0.6701120734214783,
736
+ "learning_rate": 0.0005685914732865623,
737
+ "loss": 4.0714,
738
+ "step": 4950
739
+ },
740
+ {
741
+ "epoch": 0.5390835579514824,
742
+ "grad_norm": 0.5252517461776733,
743
+ "learning_rate": 0.0005682676740420939,
744
+ "loss": 4.0492,
745
+ "step": 5000
746
+ },
747
+ {
748
+ "epoch": 0.5390835579514824,
749
+ "eval_accuracy": 0.32119879773406573,
750
+ "eval_loss": 3.9914467334747314,
751
+ "eval_runtime": 185.5695,
752
+ "eval_samples_per_second": 97.058,
753
+ "eval_steps_per_second": 6.068,
754
+ "step": 5000
755
+ },
756
+ {
757
+ "epoch": 0.5444743935309974,
758
+ "grad_norm": 0.6187731623649597,
759
+ "learning_rate": 0.0005679438747976254,
760
+ "loss": 4.0591,
761
+ "step": 5050
762
+ },
763
+ {
764
+ "epoch": 0.5498652291105122,
765
+ "grad_norm": 0.6102878451347351,
766
+ "learning_rate": 0.000567620075553157,
767
+ "loss": 4.0513,
768
+ "step": 5100
769
+ },
770
+ {
771
+ "epoch": 0.555256064690027,
772
+ "grad_norm": 0.708706259727478,
773
+ "learning_rate": 0.0005672962763086886,
774
+ "loss": 4.0489,
775
+ "step": 5150
776
+ },
777
+ {
778
+ "epoch": 0.5606469002695418,
779
+ "grad_norm": 0.6461382508277893,
780
+ "learning_rate": 0.0005669724770642202,
781
+ "loss": 4.0661,
782
+ "step": 5200
783
+ },
784
+ {
785
+ "epoch": 0.5660377358490566,
786
+ "grad_norm": 0.7211845517158508,
787
+ "learning_rate": 0.0005666486778197517,
788
+ "loss": 4.0418,
789
+ "step": 5250
790
+ },
791
+ {
792
+ "epoch": 0.5714285714285714,
793
+ "grad_norm": 0.7692673802375793,
794
+ "learning_rate": 0.0005663248785752833,
795
+ "loss": 4.0325,
796
+ "step": 5300
797
+ },
798
+ {
799
+ "epoch": 0.5768194070080862,
800
+ "grad_norm": 0.5925557613372803,
801
+ "learning_rate": 0.0005660010793308148,
802
+ "loss": 4.0266,
803
+ "step": 5350
804
+ },
805
+ {
806
+ "epoch": 0.5822102425876011,
807
+ "grad_norm": 0.5735204815864563,
808
+ "learning_rate": 0.0005656772800863465,
809
+ "loss": 4.0372,
810
+ "step": 5400
811
+ },
812
+ {
813
+ "epoch": 0.5876010781671159,
814
+ "grad_norm": 0.5482949018478394,
815
+ "learning_rate": 0.0005653534808418779,
816
+ "loss": 4.036,
817
+ "step": 5450
818
+ },
819
+ {
820
+ "epoch": 0.5929919137466307,
821
+ "grad_norm": 0.7778071165084839,
822
+ "learning_rate": 0.0005650296815974096,
823
+ "loss": 4.0203,
824
+ "step": 5500
825
+ },
826
+ {
827
+ "epoch": 0.5983827493261455,
828
+ "grad_norm": 0.603145956993103,
829
+ "learning_rate": 0.0005647058823529411,
830
+ "loss": 4.009,
831
+ "step": 5550
832
+ },
833
+ {
834
+ "epoch": 0.6037735849056604,
835
+ "grad_norm": 0.5814369916915894,
836
+ "learning_rate": 0.0005643820831084727,
837
+ "loss": 4.0256,
838
+ "step": 5600
839
+ },
840
+ {
841
+ "epoch": 0.6091644204851752,
842
+ "grad_norm": 0.6685661673545837,
843
+ "learning_rate": 0.0005640582838640042,
844
+ "loss": 4.0055,
845
+ "step": 5650
846
+ },
847
+ {
848
+ "epoch": 0.6145552560646901,
849
+ "grad_norm": 0.5345515608787537,
850
+ "learning_rate": 0.0005637344846195358,
851
+ "loss": 4.0021,
852
+ "step": 5700
853
+ },
854
+ {
855
+ "epoch": 0.6199460916442049,
856
+ "grad_norm": 0.6599013209342957,
857
+ "learning_rate": 0.0005634106853750674,
858
+ "loss": 3.9773,
859
+ "step": 5750
860
+ },
861
+ {
862
+ "epoch": 0.6253369272237197,
863
+ "grad_norm": 0.6814241409301758,
864
+ "learning_rate": 0.000563086886130599,
865
+ "loss": 3.985,
866
+ "step": 5800
867
+ },
868
+ {
869
+ "epoch": 0.6307277628032345,
870
+ "grad_norm": 0.6166688203811646,
871
+ "learning_rate": 0.0005627630868861305,
872
+ "loss": 3.9948,
873
+ "step": 5850
874
+ },
875
+ {
876
+ "epoch": 0.6361185983827493,
877
+ "grad_norm": 0.6375031471252441,
878
+ "learning_rate": 0.0005624392876416621,
879
+ "loss": 3.995,
880
+ "step": 5900
881
+ },
882
+ {
883
+ "epoch": 0.6415094339622641,
884
+ "grad_norm": 0.6021779179573059,
885
+ "learning_rate": 0.0005621154883971937,
886
+ "loss": 3.9838,
887
+ "step": 5950
888
+ },
889
+ {
890
+ "epoch": 0.6469002695417789,
891
+ "grad_norm": 0.6350939273834229,
892
+ "learning_rate": 0.0005617916891527253,
893
+ "loss": 3.9962,
894
+ "step": 6000
895
+ },
896
+ {
897
+ "epoch": 0.6469002695417789,
898
+ "eval_accuracy": 0.3278974655736041,
899
+ "eval_loss": 3.919321060180664,
900
+ "eval_runtime": 185.6272,
901
+ "eval_samples_per_second": 97.028,
902
+ "eval_steps_per_second": 6.066,
903
+ "step": 6000
904
+ },
905
+ {
906
+ "epoch": 0.6522911051212938,
907
+ "grad_norm": 0.6063998937606812,
908
+ "learning_rate": 0.0005614678899082568,
909
+ "loss": 3.9623,
910
+ "step": 6050
911
+ },
912
+ {
913
+ "epoch": 0.6576819407008087,
914
+ "grad_norm": 0.6236124634742737,
915
+ "learning_rate": 0.0005611440906637884,
916
+ "loss": 3.9998,
917
+ "step": 6100
918
+ },
919
+ {
920
+ "epoch": 0.6630727762803235,
921
+ "grad_norm": 0.7479600310325623,
922
+ "learning_rate": 0.0005608267674042094,
923
+ "loss": 3.9792,
924
+ "step": 6150
925
+ },
926
+ {
927
+ "epoch": 0.6684636118598383,
928
+ "grad_norm": 0.5813738703727722,
929
+ "learning_rate": 0.0005605029681597409,
930
+ "loss": 3.953,
931
+ "step": 6200
932
+ },
933
+ {
934
+ "epoch": 0.6738544474393531,
935
+ "grad_norm": 0.5313798785209656,
936
+ "learning_rate": 0.0005601791689152725,
937
+ "loss": 3.972,
938
+ "step": 6250
939
+ },
940
+ {
941
+ "epoch": 0.6792452830188679,
942
+ "grad_norm": 0.6238154768943787,
943
+ "learning_rate": 0.000559855369670804,
944
+ "loss": 3.9769,
945
+ "step": 6300
946
+ },
947
+ {
948
+ "epoch": 0.6846361185983828,
949
+ "grad_norm": 0.5220109224319458,
950
+ "learning_rate": 0.0005595315704263357,
951
+ "loss": 3.9613,
952
+ "step": 6350
953
+ },
954
+ {
955
+ "epoch": 0.6900269541778976,
956
+ "grad_norm": 0.673324704170227,
957
+ "learning_rate": 0.0005592077711818672,
958
+ "loss": 3.9706,
959
+ "step": 6400
960
+ },
961
+ {
962
+ "epoch": 0.6954177897574124,
963
+ "grad_norm": 0.6563860177993774,
964
+ "learning_rate": 0.0005588839719373988,
965
+ "loss": 3.969,
966
+ "step": 6450
967
+ },
968
+ {
969
+ "epoch": 0.7008086253369272,
970
+ "grad_norm": 0.6665467023849487,
971
+ "learning_rate": 0.0005585601726929303,
972
+ "loss": 3.9606,
973
+ "step": 6500
974
+ },
975
+ {
976
+ "epoch": 0.706199460916442,
977
+ "grad_norm": 0.5975165367126465,
978
+ "learning_rate": 0.0005582363734484619,
979
+ "loss": 3.959,
980
+ "step": 6550
981
+ },
982
+ {
983
+ "epoch": 0.7115902964959568,
984
+ "grad_norm": 0.6370920538902283,
985
+ "learning_rate": 0.0005579125742039935,
986
+ "loss": 3.9606,
987
+ "step": 6600
988
+ },
989
+ {
990
+ "epoch": 0.7169811320754716,
991
+ "grad_norm": 0.6234825849533081,
992
+ "learning_rate": 0.0005575887749595251,
993
+ "loss": 3.97,
994
+ "step": 6650
995
+ },
996
+ {
997
+ "epoch": 0.7223719676549866,
998
+ "grad_norm": 0.6267138123512268,
999
+ "learning_rate": 0.0005572649757150566,
1000
+ "loss": 3.9344,
1001
+ "step": 6700
1002
+ },
1003
+ {
1004
+ "epoch": 0.7277628032345014,
1005
+ "grad_norm": 0.5611273050308228,
1006
+ "learning_rate": 0.0005569411764705882,
1007
+ "loss": 3.9631,
1008
+ "step": 6750
1009
+ },
1010
+ {
1011
+ "epoch": 0.7331536388140162,
1012
+ "grad_norm": 0.6250568628311157,
1013
+ "learning_rate": 0.0005566173772261198,
1014
+ "loss": 3.9344,
1015
+ "step": 6800
1016
+ },
1017
+ {
1018
+ "epoch": 0.738544474393531,
1019
+ "grad_norm": 0.5902653932571411,
1020
+ "learning_rate": 0.0005562935779816513,
1021
+ "loss": 3.9345,
1022
+ "step": 6850
1023
+ },
1024
+ {
1025
+ "epoch": 0.7439353099730458,
1026
+ "grad_norm": 0.5684419274330139,
1027
+ "learning_rate": 0.0005559697787371828,
1028
+ "loss": 3.9367,
1029
+ "step": 6900
1030
+ },
1031
+ {
1032
+ "epoch": 0.7493261455525606,
1033
+ "grad_norm": 0.5950194001197815,
1034
+ "learning_rate": 0.0005556459794927145,
1035
+ "loss": 3.9418,
1036
+ "step": 6950
1037
+ },
1038
+ {
1039
+ "epoch": 0.7547169811320755,
1040
+ "grad_norm": 0.6283013820648193,
1041
+ "learning_rate": 0.000555322180248246,
1042
+ "loss": 3.9219,
1043
+ "step": 7000
1044
+ },
1045
+ {
1046
+ "epoch": 0.7547169811320755,
1047
+ "eval_accuracy": 0.33346299244612526,
1048
+ "eval_loss": 3.8631539344787598,
1049
+ "eval_runtime": 185.3774,
1050
+ "eval_samples_per_second": 97.159,
1051
+ "eval_steps_per_second": 6.074,
1052
+ "step": 7000
1053
+ },
1054
+ {
1055
+ "epoch": 0.7601078167115903,
1056
+ "grad_norm": 0.6419661641120911,
1057
+ "learning_rate": 0.0005549983810037776,
1058
+ "loss": 3.9341,
1059
+ "step": 7050
1060
+ },
1061
+ {
1062
+ "epoch": 0.7654986522911051,
1063
+ "grad_norm": 0.8758557438850403,
1064
+ "learning_rate": 0.0005546745817593091,
1065
+ "loss": 3.9093,
1066
+ "step": 7100
1067
+ },
1068
+ {
1069
+ "epoch": 0.77088948787062,
1070
+ "grad_norm": 0.6024784445762634,
1071
+ "learning_rate": 0.0005543507825148408,
1072
+ "loss": 3.9347,
1073
+ "step": 7150
1074
+ },
1075
+ {
1076
+ "epoch": 0.7762803234501348,
1077
+ "grad_norm": 0.5663654208183289,
1078
+ "learning_rate": 0.0005540269832703723,
1079
+ "loss": 3.9072,
1080
+ "step": 7200
1081
+ },
1082
+ {
1083
+ "epoch": 0.7816711590296496,
1084
+ "grad_norm": 0.6945359110832214,
1085
+ "learning_rate": 0.0005537031840259039,
1086
+ "loss": 3.899,
1087
+ "step": 7250
1088
+ },
1089
+ {
1090
+ "epoch": 0.7870619946091644,
1091
+ "grad_norm": 0.5187634229660034,
1092
+ "learning_rate": 0.0005533793847814354,
1093
+ "loss": 3.9124,
1094
+ "step": 7300
1095
+ },
1096
+ {
1097
+ "epoch": 0.7924528301886793,
1098
+ "grad_norm": 0.6671504378318787,
1099
+ "learning_rate": 0.000553055585536967,
1100
+ "loss": 3.9133,
1101
+ "step": 7350
1102
+ },
1103
+ {
1104
+ "epoch": 0.7978436657681941,
1105
+ "grad_norm": 0.5317527055740356,
1106
+ "learning_rate": 0.0005527317862924987,
1107
+ "loss": 3.9103,
1108
+ "step": 7400
1109
+ },
1110
+ {
1111
+ "epoch": 0.8032345013477089,
1112
+ "grad_norm": 0.567512571811676,
1113
+ "learning_rate": 0.0005524079870480301,
1114
+ "loss": 3.8879,
1115
+ "step": 7450
1116
+ },
1117
+ {
1118
+ "epoch": 0.8086253369272237,
1119
+ "grad_norm": 0.5598046183586121,
1120
+ "learning_rate": 0.0005520841878035618,
1121
+ "loss": 3.9029,
1122
+ "step": 7500
1123
+ },
1124
+ {
1125
+ "epoch": 0.8140161725067385,
1126
+ "grad_norm": 0.6032840013504028,
1127
+ "learning_rate": 0.0005517603885590933,
1128
+ "loss": 3.8964,
1129
+ "step": 7550
1130
+ },
1131
+ {
1132
+ "epoch": 0.8194070080862533,
1133
+ "grad_norm": 0.5689701437950134,
1134
+ "learning_rate": 0.0005514365893146249,
1135
+ "loss": 3.8911,
1136
+ "step": 7600
1137
+ },
1138
+ {
1139
+ "epoch": 0.8247978436657682,
1140
+ "grad_norm": 0.5858896374702454,
1141
+ "learning_rate": 0.0005511127900701564,
1142
+ "loss": 3.8974,
1143
+ "step": 7650
1144
+ },
1145
+ {
1146
+ "epoch": 0.8301886792452831,
1147
+ "grad_norm": 0.6379143595695496,
1148
+ "learning_rate": 0.000550788990825688,
1149
+ "loss": 3.8913,
1150
+ "step": 7700
1151
+ },
1152
+ {
1153
+ "epoch": 0.8355795148247979,
1154
+ "grad_norm": 0.6346825361251831,
1155
+ "learning_rate": 0.0005504651915812196,
1156
+ "loss": 3.8938,
1157
+ "step": 7750
1158
+ },
1159
+ {
1160
+ "epoch": 0.8409703504043127,
1161
+ "grad_norm": 0.651748776435852,
1162
+ "learning_rate": 0.0005501413923367512,
1163
+ "loss": 3.8768,
1164
+ "step": 7800
1165
+ },
1166
+ {
1167
+ "epoch": 0.8463611859838275,
1168
+ "grad_norm": 0.5529057383537292,
1169
+ "learning_rate": 0.0005498175930922827,
1170
+ "loss": 3.8814,
1171
+ "step": 7850
1172
+ },
1173
+ {
1174
+ "epoch": 0.8517520215633423,
1175
+ "grad_norm": 0.5781732797622681,
1176
+ "learning_rate": 0.0005494937938478143,
1177
+ "loss": 3.8778,
1178
+ "step": 7900
1179
+ },
1180
+ {
1181
+ "epoch": 0.8571428571428571,
1182
+ "grad_norm": 0.599770188331604,
1183
+ "learning_rate": 0.0005491699946033459,
1184
+ "loss": 3.9077,
1185
+ "step": 7950
1186
+ },
1187
+ {
1188
+ "epoch": 0.862533692722372,
1189
+ "grad_norm": 0.6126531958580017,
1190
+ "learning_rate": 0.0005488461953588775,
1191
+ "loss": 3.8566,
1192
+ "step": 8000
1193
+ },
1194
+ {
1195
+ "epoch": 0.862533692722372,
1196
+ "eval_accuracy": 0.3376529737589151,
1197
+ "eval_loss": 3.815525531768799,
1198
+ "eval_runtime": 186.958,
1199
+ "eval_samples_per_second": 96.337,
1200
+ "eval_steps_per_second": 6.023,
1201
+ "step": 8000
1202
+ },
1203
+ {
1204
+ "epoch": 0.8679245283018868,
1205
+ "grad_norm": 0.6914803981781006,
1206
+ "learning_rate": 0.000548522396114409,
1207
+ "loss": 3.8954,
1208
+ "step": 8050
1209
+ },
1210
+ {
1211
+ "epoch": 0.8733153638814016,
1212
+ "grad_norm": 0.6908184289932251,
1213
+ "learning_rate": 0.0005481985968699406,
1214
+ "loss": 3.8747,
1215
+ "step": 8100
1216
+ },
1217
+ {
1218
+ "epoch": 0.8787061994609164,
1219
+ "grad_norm": 0.5247132778167725,
1220
+ "learning_rate": 0.0005478747976254721,
1221
+ "loss": 3.8785,
1222
+ "step": 8150
1223
+ },
1224
+ {
1225
+ "epoch": 0.8840970350404312,
1226
+ "grad_norm": 0.585969090461731,
1227
+ "learning_rate": 0.0005475509983810037,
1228
+ "loss": 3.8734,
1229
+ "step": 8200
1230
+ },
1231
+ {
1232
+ "epoch": 0.889487870619946,
1233
+ "grad_norm": 0.5993847250938416,
1234
+ "learning_rate": 0.0005472271991365352,
1235
+ "loss": 3.8685,
1236
+ "step": 8250
1237
+ },
1238
+ {
1239
+ "epoch": 0.894878706199461,
1240
+ "grad_norm": 0.6115968227386475,
1241
+ "learning_rate": 0.0005469033998920669,
1242
+ "loss": 3.8867,
1243
+ "step": 8300
1244
+ },
1245
+ {
1246
+ "epoch": 0.9002695417789758,
1247
+ "grad_norm": 0.5844521522521973,
1248
+ "learning_rate": 0.0005465860766324878,
1249
+ "loss": 3.8566,
1250
+ "step": 8350
1251
+ },
1252
+ {
1253
+ "epoch": 0.9056603773584906,
1254
+ "grad_norm": 0.5750303864479065,
1255
+ "learning_rate": 0.0005462622773880194,
1256
+ "loss": 3.8639,
1257
+ "step": 8400
1258
+ },
1259
+ {
1260
+ "epoch": 0.9110512129380054,
1261
+ "grad_norm": 0.5799904465675354,
1262
+ "learning_rate": 0.000545938478143551,
1263
+ "loss": 3.859,
1264
+ "step": 8450
1265
+ },
1266
+ {
1267
+ "epoch": 0.9164420485175202,
1268
+ "grad_norm": 0.6240545511245728,
1269
+ "learning_rate": 0.0005456146788990825,
1270
+ "loss": 3.8648,
1271
+ "step": 8500
1272
+ },
1273
+ {
1274
+ "epoch": 0.921832884097035,
1275
+ "grad_norm": 0.5578630566596985,
1276
+ "learning_rate": 0.000545290879654614,
1277
+ "loss": 3.8597,
1278
+ "step": 8550
1279
+ },
1280
+ {
1281
+ "epoch": 0.9272237196765498,
1282
+ "grad_norm": 0.5732514262199402,
1283
+ "learning_rate": 0.0005449670804101457,
1284
+ "loss": 3.8652,
1285
+ "step": 8600
1286
+ },
1287
+ {
1288
+ "epoch": 0.9326145552560647,
1289
+ "grad_norm": 0.614829957485199,
1290
+ "learning_rate": 0.0005446432811656773,
1291
+ "loss": 3.845,
1292
+ "step": 8650
1293
+ },
1294
+ {
1295
+ "epoch": 0.9380053908355795,
1296
+ "grad_norm": 0.6199792623519897,
1297
+ "learning_rate": 0.0005443194819212088,
1298
+ "loss": 3.8484,
1299
+ "step": 8700
1300
+ },
1301
+ {
1302
+ "epoch": 0.9433962264150944,
1303
+ "grad_norm": 0.5770998597145081,
1304
+ "learning_rate": 0.0005439956826767404,
1305
+ "loss": 3.8456,
1306
+ "step": 8750
1307
+ },
1308
+ {
1309
+ "epoch": 0.9487870619946092,
1310
+ "grad_norm": 0.5756552815437317,
1311
+ "learning_rate": 0.000543671883432272,
1312
+ "loss": 3.8492,
1313
+ "step": 8800
1314
+ },
1315
+ {
1316
+ "epoch": 0.954177897574124,
1317
+ "grad_norm": 0.5448219180107117,
1318
+ "learning_rate": 0.0005433480841878035,
1319
+ "loss": 3.8518,
1320
+ "step": 8850
1321
+ },
1322
+ {
1323
+ "epoch": 0.9595687331536388,
1324
+ "grad_norm": 0.5587062239646912,
1325
+ "learning_rate": 0.000543024284943335,
1326
+ "loss": 3.8389,
1327
+ "step": 8900
1328
+ },
1329
+ {
1330
+ "epoch": 0.9649595687331537,
1331
+ "grad_norm": 0.5609190464019775,
1332
+ "learning_rate": 0.0005427004856988667,
1333
+ "loss": 3.8303,
1334
+ "step": 8950
1335
+ },
1336
+ {
1337
+ "epoch": 0.9703504043126685,
1338
+ "grad_norm": 0.6230307817459106,
1339
+ "learning_rate": 0.0005423766864543982,
1340
+ "loss": 3.856,
1341
+ "step": 9000
1342
+ },
1343
+ {
1344
+ "epoch": 0.9703504043126685,
1345
+ "eval_accuracy": 0.34079347682830485,
1346
+ "eval_loss": 3.7813315391540527,
1347
+ "eval_runtime": 186.6989,
1348
+ "eval_samples_per_second": 96.471,
1349
+ "eval_steps_per_second": 6.031,
1350
+ "step": 9000
1351
+ },
1352
+ {
1353
+ "epoch": 0.9757412398921833,
1354
+ "grad_norm": 0.6406134963035583,
1355
+ "learning_rate": 0.0005420528872099298,
1356
+ "loss": 3.871,
1357
+ "step": 9050
1358
+ },
1359
+ {
1360
+ "epoch": 0.9811320754716981,
1361
+ "grad_norm": 0.49706608057022095,
1362
+ "learning_rate": 0.0005417290879654613,
1363
+ "loss": 3.8273,
1364
+ "step": 9100
1365
+ },
1366
+ {
1367
+ "epoch": 0.9865229110512129,
1368
+ "grad_norm": 0.5765336751937866,
1369
+ "learning_rate": 0.000541405288720993,
1370
+ "loss": 3.8418,
1371
+ "step": 9150
1372
+ },
1373
+ {
1374
+ "epoch": 0.9919137466307277,
1375
+ "grad_norm": 0.5545505881309509,
1376
+ "learning_rate": 0.0005410814894765245,
1377
+ "loss": 3.8285,
1378
+ "step": 9200
1379
+ },
1380
+ {
1381
+ "epoch": 0.9973045822102425,
1382
+ "grad_norm": 0.5445045232772827,
1383
+ "learning_rate": 0.0005407576902320561,
1384
+ "loss": 3.8485,
1385
+ "step": 9250
1386
+ },
1387
+ {
1388
+ "epoch": 1.0026954177897573,
1389
+ "grad_norm": 0.5650607347488403,
1390
+ "learning_rate": 0.0005404338909875876,
1391
+ "loss": 3.7963,
1392
+ "step": 9300
1393
+ },
1394
+ {
1395
+ "epoch": 1.0080862533692723,
1396
+ "grad_norm": 0.5752312541007996,
1397
+ "learning_rate": 0.0005401100917431192,
1398
+ "loss": 3.7605,
1399
+ "step": 9350
1400
+ },
1401
+ {
1402
+ "epoch": 1.013477088948787,
1403
+ "grad_norm": 0.5649228096008301,
1404
+ "learning_rate": 0.0005397862924986508,
1405
+ "loss": 3.7564,
1406
+ "step": 9400
1407
+ },
1408
+ {
1409
+ "epoch": 1.0188679245283019,
1410
+ "grad_norm": 0.5715786218643188,
1411
+ "learning_rate": 0.0005394624932541824,
1412
+ "loss": 3.7554,
1413
+ "step": 9450
1414
+ },
1415
+ {
1416
+ "epoch": 1.0242587601078168,
1417
+ "grad_norm": 0.5207133889198303,
1418
+ "learning_rate": 0.0005391386940097139,
1419
+ "loss": 3.7592,
1420
+ "step": 9500
1421
+ },
1422
+ {
1423
+ "epoch": 1.0296495956873315,
1424
+ "grad_norm": 0.5603737235069275,
1425
+ "learning_rate": 0.0005388148947652455,
1426
+ "loss": 3.7704,
1427
+ "step": 9550
1428
+ },
1429
+ {
1430
+ "epoch": 1.0350404312668464,
1431
+ "grad_norm": 0.6097581386566162,
1432
+ "learning_rate": 0.000538491095520777,
1433
+ "loss": 3.7686,
1434
+ "step": 9600
1435
+ },
1436
+ {
1437
+ "epoch": 1.0404312668463611,
1438
+ "grad_norm": 0.6030071377754211,
1439
+ "learning_rate": 0.0005381672962763086,
1440
+ "loss": 3.7726,
1441
+ "step": 9650
1442
+ },
1443
+ {
1444
+ "epoch": 1.045822102425876,
1445
+ "grad_norm": 0.546688437461853,
1446
+ "learning_rate": 0.0005378434970318403,
1447
+ "loss": 3.7567,
1448
+ "step": 9700
1449
+ },
1450
+ {
1451
+ "epoch": 1.0512129380053907,
1452
+ "grad_norm": 0.5595793724060059,
1453
+ "learning_rate": 0.0005375196977873718,
1454
+ "loss": 3.755,
1455
+ "step": 9750
1456
+ },
1457
+ {
1458
+ "epoch": 1.0566037735849056,
1459
+ "grad_norm": 0.5992977619171143,
1460
+ "learning_rate": 0.0005371958985429034,
1461
+ "loss": 3.7821,
1462
+ "step": 9800
1463
+ },
1464
+ {
1465
+ "epoch": 1.0619946091644206,
1466
+ "grad_norm": 0.5221476554870605,
1467
+ "learning_rate": 0.0005368720992984349,
1468
+ "loss": 3.7845,
1469
+ "step": 9850
1470
+ },
1471
+ {
1472
+ "epoch": 1.0673854447439353,
1473
+ "grad_norm": 0.5831781029701233,
1474
+ "learning_rate": 0.0005365483000539665,
1475
+ "loss": 3.772,
1476
+ "step": 9900
1477
+ },
1478
+ {
1479
+ "epoch": 1.0727762803234502,
1480
+ "grad_norm": 0.5980172753334045,
1481
+ "learning_rate": 0.0005362245008094981,
1482
+ "loss": 3.7854,
1483
+ "step": 9950
1484
+ },
1485
+ {
1486
+ "epoch": 1.0781671159029649,
1487
+ "grad_norm": 0.5590375661849976,
1488
+ "learning_rate": 0.0005359007015650297,
1489
+ "loss": 3.7649,
1490
+ "step": 10000
1491
+ },
1492
+ {
1493
+ "epoch": 1.0781671159029649,
1494
+ "eval_accuracy": 0.34435424926776104,
1495
+ "eval_loss": 3.7518138885498047,
1496
+ "eval_runtime": 186.0381,
1497
+ "eval_samples_per_second": 96.814,
1498
+ "eval_steps_per_second": 6.053,
1499
+ "step": 10000
1500
+ },
1501
+ {
1502
+ "epoch": 1.0835579514824798,
1503
+ "grad_norm": 0.5709946751594543,
1504
+ "learning_rate": 0.0005355769023205612,
1505
+ "loss": 3.7581,
1506
+ "step": 10050
1507
+ },
1508
+ {
1509
+ "epoch": 1.0889487870619945,
1510
+ "grad_norm": 0.618277370929718,
1511
+ "learning_rate": 0.0005352531030760928,
1512
+ "loss": 3.7592,
1513
+ "step": 10100
1514
+ },
1515
+ {
1516
+ "epoch": 1.0943396226415094,
1517
+ "grad_norm": 0.6374980211257935,
1518
+ "learning_rate": 0.0005349293038316244,
1519
+ "loss": 3.7555,
1520
+ "step": 10150
1521
+ },
1522
+ {
1523
+ "epoch": 1.0997304582210243,
1524
+ "grad_norm": 0.5958013534545898,
1525
+ "learning_rate": 0.0005346055045871559,
1526
+ "loss": 3.7666,
1527
+ "step": 10200
1528
+ },
1529
+ {
1530
+ "epoch": 1.105121293800539,
1531
+ "grad_norm": 0.6115665435791016,
1532
+ "learning_rate": 0.0005342817053426874,
1533
+ "loss": 3.7524,
1534
+ "step": 10250
1535
+ },
1536
+ {
1537
+ "epoch": 1.110512129380054,
1538
+ "grad_norm": 0.585763692855835,
1539
+ "learning_rate": 0.0005339579060982191,
1540
+ "loss": 3.764,
1541
+ "step": 10300
1542
+ },
1543
+ {
1544
+ "epoch": 1.1159029649595686,
1545
+ "grad_norm": 0.653211772441864,
1546
+ "learning_rate": 0.0005336341068537506,
1547
+ "loss": 3.7568,
1548
+ "step": 10350
1549
+ },
1550
+ {
1551
+ "epoch": 1.1212938005390836,
1552
+ "grad_norm": 0.5737949013710022,
1553
+ "learning_rate": 0.0005333103076092822,
1554
+ "loss": 3.7638,
1555
+ "step": 10400
1556
+ },
1557
+ {
1558
+ "epoch": 1.1266846361185983,
1559
+ "grad_norm": 0.5587084293365479,
1560
+ "learning_rate": 0.0005329865083648137,
1561
+ "loss": 3.7609,
1562
+ "step": 10450
1563
+ },
1564
+ {
1565
+ "epoch": 1.1320754716981132,
1566
+ "grad_norm": 0.6286759972572327,
1567
+ "learning_rate": 0.0005326627091203454,
1568
+ "loss": 3.7635,
1569
+ "step": 10500
1570
+ },
1571
+ {
1572
+ "epoch": 1.137466307277628,
1573
+ "grad_norm": 0.5914649367332458,
1574
+ "learning_rate": 0.0005323453858607662,
1575
+ "loss": 3.7659,
1576
+ "step": 10550
1577
+ },
1578
+ {
1579
+ "epoch": 1.1428571428571428,
1580
+ "grad_norm": 0.574865460395813,
1581
+ "learning_rate": 0.0005320215866162979,
1582
+ "loss": 3.7552,
1583
+ "step": 10600
1584
+ },
1585
+ {
1586
+ "epoch": 1.1482479784366577,
1587
+ "grad_norm": 0.570719838142395,
1588
+ "learning_rate": 0.0005316977873718294,
1589
+ "loss": 3.7441,
1590
+ "step": 10650
1591
+ },
1592
+ {
1593
+ "epoch": 1.1536388140161726,
1594
+ "grad_norm": 0.5437813997268677,
1595
+ "learning_rate": 0.000531373988127361,
1596
+ "loss": 3.7548,
1597
+ "step": 10700
1598
+ },
1599
+ {
1600
+ "epoch": 1.1590296495956873,
1601
+ "grad_norm": 0.575222909450531,
1602
+ "learning_rate": 0.0005310501888828925,
1603
+ "loss": 3.7359,
1604
+ "step": 10750
1605
+ },
1606
+ {
1607
+ "epoch": 1.1644204851752022,
1608
+ "grad_norm": 0.5911815762519836,
1609
+ "learning_rate": 0.0005307263896384242,
1610
+ "loss": 3.7398,
1611
+ "step": 10800
1612
+ },
1613
+ {
1614
+ "epoch": 1.169811320754717,
1615
+ "grad_norm": 0.5724892020225525,
1616
+ "learning_rate": 0.0005304025903939556,
1617
+ "loss": 3.7319,
1618
+ "step": 10850
1619
+ },
1620
+ {
1621
+ "epoch": 1.1752021563342319,
1622
+ "grad_norm": 0.6011896729469299,
1623
+ "learning_rate": 0.0005300787911494873,
1624
+ "loss": 3.7465,
1625
+ "step": 10900
1626
+ },
1627
+ {
1628
+ "epoch": 1.1805929919137466,
1629
+ "grad_norm": 0.5386224985122681,
1630
+ "learning_rate": 0.0005297549919050189,
1631
+ "loss": 3.7593,
1632
+ "step": 10950
1633
+ },
1634
+ {
1635
+ "epoch": 1.1859838274932615,
1636
+ "grad_norm": 0.5424002408981323,
1637
+ "learning_rate": 0.0005294311926605504,
1638
+ "loss": 3.769,
1639
+ "step": 11000
1640
+ },
1641
+ {
1642
+ "epoch": 1.1859838274932615,
1643
+ "eval_accuracy": 0.3473514391781235,
1644
+ "eval_loss": 3.723193883895874,
1645
+ "eval_runtime": 186.4484,
1646
+ "eval_samples_per_second": 96.6,
1647
+ "eval_steps_per_second": 6.039,
1648
+ "step": 11000
1649
+ },
1650
+ {
1651
+ "epoch": 1.1913746630727764,
1652
+ "grad_norm": 0.5702701210975647,
1653
+ "learning_rate": 0.000529107393416082,
1654
+ "loss": 3.7567,
1655
+ "step": 11050
1656
+ },
1657
+ {
1658
+ "epoch": 1.196765498652291,
1659
+ "grad_norm": 0.5534523129463196,
1660
+ "learning_rate": 0.0005287835941716135,
1661
+ "loss": 3.7412,
1662
+ "step": 11100
1663
+ },
1664
+ {
1665
+ "epoch": 1.202156334231806,
1666
+ "grad_norm": 0.5500940084457397,
1667
+ "learning_rate": 0.0005284597949271452,
1668
+ "loss": 3.7407,
1669
+ "step": 11150
1670
+ },
1671
+ {
1672
+ "epoch": 1.2075471698113207,
1673
+ "grad_norm": 0.5379878878593445,
1674
+ "learning_rate": 0.0005281359956826767,
1675
+ "loss": 3.7631,
1676
+ "step": 11200
1677
+ },
1678
+ {
1679
+ "epoch": 1.2129380053908356,
1680
+ "grad_norm": 0.5732437372207642,
1681
+ "learning_rate": 0.0005278121964382083,
1682
+ "loss": 3.7498,
1683
+ "step": 11250
1684
+ },
1685
+ {
1686
+ "epoch": 1.2183288409703503,
1687
+ "grad_norm": 0.5423194766044617,
1688
+ "learning_rate": 0.0005274883971937398,
1689
+ "loss": 3.7371,
1690
+ "step": 11300
1691
+ },
1692
+ {
1693
+ "epoch": 1.2237196765498652,
1694
+ "grad_norm": 0.542545735836029,
1695
+ "learning_rate": 0.0005271645979492714,
1696
+ "loss": 3.7301,
1697
+ "step": 11350
1698
+ },
1699
+ {
1700
+ "epoch": 1.2291105121293802,
1701
+ "grad_norm": 0.8030261397361755,
1702
+ "learning_rate": 0.0005268472746896923,
1703
+ "loss": 3.7452,
1704
+ "step": 11400
1705
+ },
1706
+ {
1707
+ "epoch": 1.2345013477088949,
1708
+ "grad_norm": 0.5232232809066772,
1709
+ "learning_rate": 0.000526523475445224,
1710
+ "loss": 3.755,
1711
+ "step": 11450
1712
+ },
1713
+ {
1714
+ "epoch": 1.2398921832884098,
1715
+ "grad_norm": 0.5808193683624268,
1716
+ "learning_rate": 0.0005261996762007554,
1717
+ "loss": 3.7452,
1718
+ "step": 11500
1719
+ },
1720
+ {
1721
+ "epoch": 1.2452830188679245,
1722
+ "grad_norm": 0.5781920552253723,
1723
+ "learning_rate": 0.0005258758769562871,
1724
+ "loss": 3.7392,
1725
+ "step": 11550
1726
+ },
1727
+ {
1728
+ "epoch": 1.2506738544474394,
1729
+ "grad_norm": 0.5637895464897156,
1730
+ "learning_rate": 0.0005255520777118186,
1731
+ "loss": 3.7512,
1732
+ "step": 11600
1733
+ },
1734
+ {
1735
+ "epoch": 1.256064690026954,
1736
+ "grad_norm": 0.5855022668838501,
1737
+ "learning_rate": 0.0005252282784673502,
1738
+ "loss": 3.7318,
1739
+ "step": 11650
1740
+ },
1741
+ {
1742
+ "epoch": 1.261455525606469,
1743
+ "grad_norm": 0.5060122013092041,
1744
+ "learning_rate": 0.0005249044792228817,
1745
+ "loss": 3.7424,
1746
+ "step": 11700
1747
+ },
1748
+ {
1749
+ "epoch": 1.266846361185984,
1750
+ "grad_norm": 0.5578184723854065,
1751
+ "learning_rate": 0.0005245806799784133,
1752
+ "loss": 3.7421,
1753
+ "step": 11750
1754
+ },
1755
+ {
1756
+ "epoch": 1.2722371967654986,
1757
+ "grad_norm": 0.5692754983901978,
1758
+ "learning_rate": 0.0005242568807339449,
1759
+ "loss": 3.7541,
1760
+ "step": 11800
1761
+ },
1762
+ {
1763
+ "epoch": 1.2776280323450135,
1764
+ "grad_norm": 0.5293059349060059,
1765
+ "learning_rate": 0.0005239330814894765,
1766
+ "loss": 3.7389,
1767
+ "step": 11850
1768
+ },
1769
+ {
1770
+ "epoch": 1.2830188679245282,
1771
+ "grad_norm": 0.5823394060134888,
1772
+ "learning_rate": 0.000523609282245008,
1773
+ "loss": 3.7336,
1774
+ "step": 11900
1775
+ },
1776
+ {
1777
+ "epoch": 1.2884097035040432,
1778
+ "grad_norm": 0.5567950010299683,
1779
+ "learning_rate": 0.0005232854830005396,
1780
+ "loss": 3.7176,
1781
+ "step": 11950
1782
+ },
1783
+ {
1784
+ "epoch": 1.2938005390835579,
1785
+ "grad_norm": 0.5795886516571045,
1786
+ "learning_rate": 0.0005229616837560712,
1787
+ "loss": 3.7288,
1788
+ "step": 12000
1789
+ },
1790
+ {
1791
+ "epoch": 1.2938005390835579,
1792
+ "eval_accuracy": 0.34931164592718456,
1793
+ "eval_loss": 3.697058916091919,
1794
+ "eval_runtime": 186.8168,
1795
+ "eval_samples_per_second": 96.41,
1796
+ "eval_steps_per_second": 6.027,
1797
+ "step": 12000
1798
+ },
1799
+ {
1800
+ "epoch": 1.2991913746630728,
1801
+ "grad_norm": 0.6153177618980408,
1802
+ "learning_rate": 0.0005226378845116028,
1803
+ "loss": 3.7453,
1804
+ "step": 12050
1805
+ },
1806
+ {
1807
+ "epoch": 1.3045822102425877,
1808
+ "grad_norm": 0.6005005836486816,
1809
+ "learning_rate": 0.0005223140852671344,
1810
+ "loss": 3.7398,
1811
+ "step": 12100
1812
+ },
1813
+ {
1814
+ "epoch": 1.3099730458221024,
1815
+ "grad_norm": 0.5839811563491821,
1816
+ "learning_rate": 0.0005219902860226659,
1817
+ "loss": 3.7295,
1818
+ "step": 12150
1819
+ },
1820
+ {
1821
+ "epoch": 1.3153638814016173,
1822
+ "grad_norm": 0.54124516248703,
1823
+ "learning_rate": 0.0005216664867781975,
1824
+ "loss": 3.7451,
1825
+ "step": 12200
1826
+ },
1827
+ {
1828
+ "epoch": 1.320754716981132,
1829
+ "grad_norm": 0.5872597098350525,
1830
+ "learning_rate": 0.000521342687533729,
1831
+ "loss": 3.7195,
1832
+ "step": 12250
1833
+ },
1834
+ {
1835
+ "epoch": 1.326145552560647,
1836
+ "grad_norm": 0.6764320731163025,
1837
+ "learning_rate": 0.0005210188882892606,
1838
+ "loss": 3.7273,
1839
+ "step": 12300
1840
+ },
1841
+ {
1842
+ "epoch": 1.3315363881401616,
1843
+ "grad_norm": 0.5127867460250854,
1844
+ "learning_rate": 0.0005206950890447922,
1845
+ "loss": 3.7354,
1846
+ "step": 12350
1847
+ },
1848
+ {
1849
+ "epoch": 1.3369272237196765,
1850
+ "grad_norm": 0.5766441822052002,
1851
+ "learning_rate": 0.0005203712898003238,
1852
+ "loss": 3.7329,
1853
+ "step": 12400
1854
+ },
1855
+ {
1856
+ "epoch": 1.3423180592991915,
1857
+ "grad_norm": 0.5768564343452454,
1858
+ "learning_rate": 0.0005200474905558553,
1859
+ "loss": 3.7196,
1860
+ "step": 12450
1861
+ },
1862
+ {
1863
+ "epoch": 1.3477088948787062,
1864
+ "grad_norm": 0.5573644638061523,
1865
+ "learning_rate": 0.0005197236913113869,
1866
+ "loss": 3.7398,
1867
+ "step": 12500
1868
+ },
1869
+ {
1870
+ "epoch": 1.353099730458221,
1871
+ "grad_norm": 0.519527018070221,
1872
+ "learning_rate": 0.0005193998920669184,
1873
+ "loss": 3.7141,
1874
+ "step": 12550
1875
+ },
1876
+ {
1877
+ "epoch": 1.3584905660377358,
1878
+ "grad_norm": 0.6151216626167297,
1879
+ "learning_rate": 0.0005190760928224501,
1880
+ "loss": 3.7121,
1881
+ "step": 12600
1882
+ },
1883
+ {
1884
+ "epoch": 1.3638814016172507,
1885
+ "grad_norm": 0.6031010746955872,
1886
+ "learning_rate": 0.0005187522935779816,
1887
+ "loss": 3.7035,
1888
+ "step": 12650
1889
+ },
1890
+ {
1891
+ "epoch": 1.3692722371967654,
1892
+ "grad_norm": 0.534813642501831,
1893
+ "learning_rate": 0.0005184284943335132,
1894
+ "loss": 3.7193,
1895
+ "step": 12700
1896
+ },
1897
+ {
1898
+ "epoch": 1.3746630727762803,
1899
+ "grad_norm": 0.5405371785163879,
1900
+ "learning_rate": 0.0005181046950890447,
1901
+ "loss": 3.7074,
1902
+ "step": 12750
1903
+ },
1904
+ {
1905
+ "epoch": 1.3800539083557952,
1906
+ "grad_norm": 0.537695050239563,
1907
+ "learning_rate": 0.0005177808958445764,
1908
+ "loss": 3.7067,
1909
+ "step": 12800
1910
+ },
1911
+ {
1912
+ "epoch": 1.38544474393531,
1913
+ "grad_norm": 0.6358603239059448,
1914
+ "learning_rate": 0.0005174570966001078,
1915
+ "loss": 3.7174,
1916
+ "step": 12850
1917
+ },
1918
+ {
1919
+ "epoch": 1.3908355795148248,
1920
+ "grad_norm": 0.6034652590751648,
1921
+ "learning_rate": 0.0005171332973556395,
1922
+ "loss": 3.7168,
1923
+ "step": 12900
1924
+ },
1925
+ {
1926
+ "epoch": 1.3962264150943398,
1927
+ "grad_norm": 0.553503692150116,
1928
+ "learning_rate": 0.000516809498111171,
1929
+ "loss": 3.7269,
1930
+ "step": 12950
1931
+ },
1932
+ {
1933
+ "epoch": 1.4016172506738545,
1934
+ "grad_norm": 0.5765125155448914,
1935
+ "learning_rate": 0.0005164856988667026,
1936
+ "loss": 3.7086,
1937
+ "step": 13000
1938
+ },
1939
+ {
1940
+ "epoch": 1.4016172506738545,
1941
+ "eval_accuracy": 0.35153598784652257,
1942
+ "eval_loss": 3.674765110015869,
1943
+ "eval_runtime": 186.3343,
1944
+ "eval_samples_per_second": 96.66,
1945
+ "eval_steps_per_second": 6.043,
1946
+ "step": 13000
1947
+ },
1948
+ {
1949
+ "epoch": 1.4070080862533692,
1950
+ "grad_norm": 0.6072783470153809,
1951
+ "learning_rate": 0.0005161618996222341,
1952
+ "loss": 3.7063,
1953
+ "step": 13050
1954
+ },
1955
+ {
1956
+ "epoch": 1.412398921832884,
1957
+ "grad_norm": 0.5731549859046936,
1958
+ "learning_rate": 0.0005158381003777657,
1959
+ "loss": 3.6997,
1960
+ "step": 13100
1961
+ },
1962
+ {
1963
+ "epoch": 1.417789757412399,
1964
+ "grad_norm": 0.5856156945228577,
1965
+ "learning_rate": 0.0005155143011332973,
1966
+ "loss": 3.7057,
1967
+ "step": 13150
1968
+ },
1969
+ {
1970
+ "epoch": 1.4231805929919137,
1971
+ "grad_norm": 0.566673755645752,
1972
+ "learning_rate": 0.0005151905018888289,
1973
+ "loss": 3.7061,
1974
+ "step": 13200
1975
+ },
1976
+ {
1977
+ "epoch": 1.4285714285714286,
1978
+ "grad_norm": 0.527152419090271,
1979
+ "learning_rate": 0.0005148667026443604,
1980
+ "loss": 3.7152,
1981
+ "step": 13250
1982
+ },
1983
+ {
1984
+ "epoch": 1.4339622641509435,
1985
+ "grad_norm": 0.5248644948005676,
1986
+ "learning_rate": 0.000514542903399892,
1987
+ "loss": 3.6882,
1988
+ "step": 13300
1989
+ },
1990
+ {
1991
+ "epoch": 1.4393530997304582,
1992
+ "grad_norm": 0.5552476048469543,
1993
+ "learning_rate": 0.0005142191041554237,
1994
+ "loss": 3.708,
1995
+ "step": 13350
1996
+ },
1997
+ {
1998
+ "epoch": 1.444743935309973,
1999
+ "grad_norm": 0.5420221090316772,
2000
+ "learning_rate": 0.0005138953049109552,
2001
+ "loss": 3.723,
2002
+ "step": 13400
2003
+ },
2004
+ {
2005
+ "epoch": 1.4501347708894878,
2006
+ "grad_norm": 0.546251654624939,
2007
+ "learning_rate": 0.0005135715056664868,
2008
+ "loss": 3.7274,
2009
+ "step": 13450
2010
+ },
2011
+ {
2012
+ "epoch": 1.4555256064690028,
2013
+ "grad_norm": 0.656278669834137,
2014
+ "learning_rate": 0.0005132477064220183,
2015
+ "loss": 3.719,
2016
+ "step": 13500
2017
+ },
2018
+ {
2019
+ "epoch": 1.4609164420485174,
2020
+ "grad_norm": 0.5882102847099304,
2021
+ "learning_rate": 0.0005129239071775499,
2022
+ "loss": 3.7077,
2023
+ "step": 13550
2024
+ },
2025
+ {
2026
+ "epoch": 1.4663072776280324,
2027
+ "grad_norm": 0.5904348492622375,
2028
+ "learning_rate": 0.0005126001079330814,
2029
+ "loss": 3.7233,
2030
+ "step": 13600
2031
+ },
2032
+ {
2033
+ "epoch": 1.4716981132075473,
2034
+ "grad_norm": 0.5616297721862793,
2035
+ "learning_rate": 0.000512276308688613,
2036
+ "loss": 3.7126,
2037
+ "step": 13650
2038
+ },
2039
+ {
2040
+ "epoch": 1.477088948787062,
2041
+ "grad_norm": 0.588398277759552,
2042
+ "learning_rate": 0.0005119525094441446,
2043
+ "loss": 3.6976,
2044
+ "step": 13700
2045
+ },
2046
+ {
2047
+ "epoch": 1.482479784366577,
2048
+ "grad_norm": 0.6392883062362671,
2049
+ "learning_rate": 0.0005116287101996762,
2050
+ "loss": 3.701,
2051
+ "step": 13750
2052
+ },
2053
+ {
2054
+ "epoch": 1.4878706199460916,
2055
+ "grad_norm": 0.5546556711196899,
2056
+ "learning_rate": 0.0005113049109552077,
2057
+ "loss": 3.7046,
2058
+ "step": 13800
2059
+ },
2060
+ {
2061
+ "epoch": 1.4932614555256065,
2062
+ "grad_norm": 0.5846448540687561,
2063
+ "learning_rate": 0.0005109811117107393,
2064
+ "loss": 3.7087,
2065
+ "step": 13850
2066
+ },
2067
+ {
2068
+ "epoch": 1.4986522911051212,
2069
+ "grad_norm": 0.6359493732452393,
2070
+ "learning_rate": 0.0005106573124662708,
2071
+ "loss": 3.6907,
2072
+ "step": 13900
2073
+ },
2074
+ {
2075
+ "epoch": 1.5040431266846361,
2076
+ "grad_norm": 0.5580483078956604,
2077
+ "learning_rate": 0.0005103335132218025,
2078
+ "loss": 3.688,
2079
+ "step": 13950
2080
+ },
2081
+ {
2082
+ "epoch": 1.509433962264151,
2083
+ "grad_norm": 0.5794188976287842,
2084
+ "learning_rate": 0.000510009713977334,
2085
+ "loss": 3.703,
2086
+ "step": 14000
2087
+ },
2088
+ {
2089
+ "epoch": 1.509433962264151,
2090
+ "eval_accuracy": 0.35349782438890087,
2091
+ "eval_loss": 3.6583781242370605,
2092
+ "eval_runtime": 185.6312,
2093
+ "eval_samples_per_second": 97.026,
2094
+ "eval_steps_per_second": 6.066,
2095
+ "step": 14000
2096
+ },
2097
+ {
2098
+ "epoch": 1.5148247978436657,
2099
+ "grad_norm": 0.5856167078018188,
2100
+ "learning_rate": 0.0005096859147328656,
2101
+ "loss": 3.6999,
2102
+ "step": 14050
2103
+ },
2104
+ {
2105
+ "epoch": 1.5202156334231804,
2106
+ "grad_norm": 0.6054800152778625,
2107
+ "learning_rate": 0.0005093621154883971,
2108
+ "loss": 3.6984,
2109
+ "step": 14100
2110
+ },
2111
+ {
2112
+ "epoch": 1.5256064690026954,
2113
+ "grad_norm": 0.5391426086425781,
2114
+ "learning_rate": 0.0005090383162439288,
2115
+ "loss": 3.695,
2116
+ "step": 14150
2117
+ },
2118
+ {
2119
+ "epoch": 1.5309973045822103,
2120
+ "grad_norm": 0.5351890921592712,
2121
+ "learning_rate": 0.0005087145169994602,
2122
+ "loss": 3.701,
2123
+ "step": 14200
2124
+ },
2125
+ {
2126
+ "epoch": 1.536388140161725,
2127
+ "grad_norm": 0.5853469371795654,
2128
+ "learning_rate": 0.0005083907177549918,
2129
+ "loss": 3.7001,
2130
+ "step": 14250
2131
+ },
2132
+ {
2133
+ "epoch": 1.54177897574124,
2134
+ "grad_norm": 0.5472508668899536,
2135
+ "learning_rate": 0.0005080669185105234,
2136
+ "loss": 3.697,
2137
+ "step": 14300
2138
+ },
2139
+ {
2140
+ "epoch": 1.5471698113207548,
2141
+ "grad_norm": 0.5463733673095703,
2142
+ "learning_rate": 0.000507743119266055,
2143
+ "loss": 3.6871,
2144
+ "step": 14350
2145
+ },
2146
+ {
2147
+ "epoch": 1.5525606469002695,
2148
+ "grad_norm": 0.549136757850647,
2149
+ "learning_rate": 0.0005074193200215865,
2150
+ "loss": 3.6895,
2151
+ "step": 14400
2152
+ },
2153
+ {
2154
+ "epoch": 1.5579514824797842,
2155
+ "grad_norm": 0.5161502957344055,
2156
+ "learning_rate": 0.0005070955207771181,
2157
+ "loss": 3.693,
2158
+ "step": 14450
2159
+ },
2160
+ {
2161
+ "epoch": 1.5633423180592994,
2162
+ "grad_norm": 0.5291047096252441,
2163
+ "learning_rate": 0.0005067717215326498,
2164
+ "loss": 3.6909,
2165
+ "step": 14500
2166
+ },
2167
+ {
2168
+ "epoch": 1.568733153638814,
2169
+ "grad_norm": 0.5388016700744629,
2170
+ "learning_rate": 0.0005064479222881813,
2171
+ "loss": 3.6906,
2172
+ "step": 14550
2173
+ },
2174
+ {
2175
+ "epoch": 1.5741239892183287,
2176
+ "grad_norm": 0.5703091621398926,
2177
+ "learning_rate": 0.0005061241230437129,
2178
+ "loss": 3.6751,
2179
+ "step": 14600
2180
+ },
2181
+ {
2182
+ "epoch": 1.5795148247978437,
2183
+ "grad_norm": 0.5753449201583862,
2184
+ "learning_rate": 0.0005058003237992444,
2185
+ "loss": 3.6803,
2186
+ "step": 14650
2187
+ },
2188
+ {
2189
+ "epoch": 1.5849056603773586,
2190
+ "grad_norm": 0.5543888807296753,
2191
+ "learning_rate": 0.000505476524554776,
2192
+ "loss": 3.6864,
2193
+ "step": 14700
2194
+ },
2195
+ {
2196
+ "epoch": 1.5902964959568733,
2197
+ "grad_norm": 0.5601723790168762,
2198
+ "learning_rate": 0.0005051527253103076,
2199
+ "loss": 3.6654,
2200
+ "step": 14750
2201
+ },
2202
+ {
2203
+ "epoch": 1.595687331536388,
2204
+ "grad_norm": 0.5522553324699402,
2205
+ "learning_rate": 0.0005048289260658392,
2206
+ "loss": 3.6576,
2207
+ "step": 14800
2208
+ },
2209
+ {
2210
+ "epoch": 1.6010781671159031,
2211
+ "grad_norm": 0.5866943001747131,
2212
+ "learning_rate": 0.0005045051268213707,
2213
+ "loss": 3.6834,
2214
+ "step": 14850
2215
+ },
2216
+ {
2217
+ "epoch": 1.6064690026954178,
2218
+ "grad_norm": 0.5801655650138855,
2219
+ "learning_rate": 0.0005041813275769023,
2220
+ "loss": 3.6865,
2221
+ "step": 14900
2222
+ },
2223
+ {
2224
+ "epoch": 1.6118598382749325,
2225
+ "grad_norm": 0.5756445527076721,
2226
+ "learning_rate": 0.0005038575283324338,
2227
+ "loss": 3.6724,
2228
+ "step": 14950
2229
+ },
2230
+ {
2231
+ "epoch": 1.6172506738544474,
2232
+ "grad_norm": 0.4960193932056427,
2233
+ "learning_rate": 0.0005035337290879654,
2234
+ "loss": 3.664,
2235
+ "step": 15000
2236
+ },
2237
+ {
2238
+ "epoch": 1.6172506738544474,
2239
+ "eval_accuracy": 0.35552974204391946,
2240
+ "eval_loss": 3.637425184249878,
2241
+ "eval_runtime": 185.9925,
2242
+ "eval_samples_per_second": 96.837,
2243
+ "eval_steps_per_second": 6.054,
2244
+ "step": 15000
2245
+ },
2246
+ {
2247
+ "epoch": 1.6226415094339623,
2248
+ "grad_norm": 0.5649005770683289,
2249
+ "learning_rate": 0.000503209929843497,
2250
+ "loss": 3.6632,
2251
+ "step": 15050
2252
+ },
2253
+ {
2254
+ "epoch": 1.628032345013477,
2255
+ "grad_norm": 0.5567517876625061,
2256
+ "learning_rate": 0.0005028861305990286,
2257
+ "loss": 3.6753,
2258
+ "step": 15100
2259
+ },
2260
+ {
2261
+ "epoch": 1.633423180592992,
2262
+ "grad_norm": 0.5907676219940186,
2263
+ "learning_rate": 0.0005025623313545601,
2264
+ "loss": 3.6967,
2265
+ "step": 15150
2266
+ },
2267
+ {
2268
+ "epoch": 1.6388140161725069,
2269
+ "grad_norm": 0.5697629451751709,
2270
+ "learning_rate": 0.0005022385321100917,
2271
+ "loss": 3.6818,
2272
+ "step": 15200
2273
+ },
2274
+ {
2275
+ "epoch": 1.6442048517520216,
2276
+ "grad_norm": 0.5234145522117615,
2277
+ "learning_rate": 0.0005019147328656232,
2278
+ "loss": 3.6737,
2279
+ "step": 15250
2280
+ },
2281
+ {
2282
+ "epoch": 1.6495956873315363,
2283
+ "grad_norm": 0.5655122995376587,
2284
+ "learning_rate": 0.0005015909336211549,
2285
+ "loss": 3.6731,
2286
+ "step": 15300
2287
+ },
2288
+ {
2289
+ "epoch": 1.6549865229110512,
2290
+ "grad_norm": 0.5556879639625549,
2291
+ "learning_rate": 0.0005012671343766864,
2292
+ "loss": 3.6583,
2293
+ "step": 15350
2294
+ },
2295
+ {
2296
+ "epoch": 1.6603773584905661,
2297
+ "grad_norm": 0.5384864211082458,
2298
+ "learning_rate": 0.000500943335132218,
2299
+ "loss": 3.6766,
2300
+ "step": 15400
2301
+ },
2302
+ {
2303
+ "epoch": 1.6657681940700808,
2304
+ "grad_norm": 0.6065835952758789,
2305
+ "learning_rate": 0.0005006195358877495,
2306
+ "loss": 3.6646,
2307
+ "step": 15450
2308
+ },
2309
+ {
2310
+ "epoch": 1.6711590296495957,
2311
+ "grad_norm": 0.5641660690307617,
2312
+ "learning_rate": 0.0005002957366432812,
2313
+ "loss": 3.666,
2314
+ "step": 15500
2315
+ },
2316
+ {
2317
+ "epoch": 1.6765498652291106,
2318
+ "grad_norm": 0.5256580710411072,
2319
+ "learning_rate": 0.0004999719373988127,
2320
+ "loss": 3.6515,
2321
+ "step": 15550
2322
+ },
2323
+ {
2324
+ "epoch": 1.6819407008086253,
2325
+ "grad_norm": 0.5791158676147461,
2326
+ "learning_rate": 0.0004996481381543442,
2327
+ "loss": 3.6708,
2328
+ "step": 15600
2329
+ },
2330
+ {
2331
+ "epoch": 1.68733153638814,
2332
+ "grad_norm": 0.5627703666687012,
2333
+ "learning_rate": 0.0004993308148947651,
2334
+ "loss": 3.6779,
2335
+ "step": 15650
2336
+ },
2337
+ {
2338
+ "epoch": 1.692722371967655,
2339
+ "grad_norm": 0.6009765863418579,
2340
+ "learning_rate": 0.0004990070156502968,
2341
+ "loss": 3.6717,
2342
+ "step": 15700
2343
+ },
2344
+ {
2345
+ "epoch": 1.6981132075471699,
2346
+ "grad_norm": 0.6165124177932739,
2347
+ "learning_rate": 0.0004986832164058284,
2348
+ "loss": 3.6558,
2349
+ "step": 15750
2350
+ },
2351
+ {
2352
+ "epoch": 1.7035040431266846,
2353
+ "grad_norm": 0.573556661605835,
2354
+ "learning_rate": 0.0004983594171613599,
2355
+ "loss": 3.6659,
2356
+ "step": 15800
2357
+ },
2358
+ {
2359
+ "epoch": 1.7088948787061995,
2360
+ "grad_norm": 0.5521446466445923,
2361
+ "learning_rate": 0.0004980356179168915,
2362
+ "loss": 3.6825,
2363
+ "step": 15850
2364
+ },
2365
+ {
2366
+ "epoch": 1.7142857142857144,
2367
+ "grad_norm": 0.5886697769165039,
2368
+ "learning_rate": 0.000497711818672423,
2369
+ "loss": 3.6819,
2370
+ "step": 15900
2371
+ },
2372
+ {
2373
+ "epoch": 1.719676549865229,
2374
+ "grad_norm": 0.6262629628181458,
2375
+ "learning_rate": 0.0004973880194279547,
2376
+ "loss": 3.6689,
2377
+ "step": 15950
2378
+ },
2379
+ {
2380
+ "epoch": 1.7250673854447438,
2381
+ "grad_norm": 0.6005555391311646,
2382
+ "learning_rate": 0.0004970642201834862,
2383
+ "loss": 3.6639,
2384
+ "step": 16000
2385
+ },
2386
+ {
2387
+ "epoch": 1.7250673854447438,
2388
+ "eval_accuracy": 0.3572108195241851,
2389
+ "eval_loss": 3.6206066608428955,
2390
+ "eval_runtime": 185.4999,
2391
+ "eval_samples_per_second": 97.094,
2392
+ "eval_steps_per_second": 6.07,
2393
+ "step": 16000
2394
+ },
2395
+ {
2396
+ "epoch": 1.7304582210242587,
2397
+ "grad_norm": 0.5764271020889282,
2398
+ "learning_rate": 0.0004967404209390178,
2399
+ "loss": 3.6523,
2400
+ "step": 16050
2401
+ },
2402
+ {
2403
+ "epoch": 1.7358490566037736,
2404
+ "grad_norm": 0.5741286277770996,
2405
+ "learning_rate": 0.0004964166216945493,
2406
+ "loss": 3.653,
2407
+ "step": 16100
2408
+ },
2409
+ {
2410
+ "epoch": 1.7412398921832883,
2411
+ "grad_norm": 0.6296107172966003,
2412
+ "learning_rate": 0.000496092822450081,
2413
+ "loss": 3.6606,
2414
+ "step": 16150
2415
+ },
2416
+ {
2417
+ "epoch": 1.7466307277628033,
2418
+ "grad_norm": 0.6621016263961792,
2419
+ "learning_rate": 0.0004957690232056125,
2420
+ "loss": 3.667,
2421
+ "step": 16200
2422
+ },
2423
+ {
2424
+ "epoch": 1.7520215633423182,
2425
+ "grad_norm": 0.6163436770439148,
2426
+ "learning_rate": 0.0004954452239611441,
2427
+ "loss": 3.6514,
2428
+ "step": 16250
2429
+ },
2430
+ {
2431
+ "epoch": 1.7574123989218329,
2432
+ "grad_norm": 0.5336316823959351,
2433
+ "learning_rate": 0.0004951214247166756,
2434
+ "loss": 3.645,
2435
+ "step": 16300
2436
+ },
2437
+ {
2438
+ "epoch": 1.7628032345013476,
2439
+ "grad_norm": 0.5538309812545776,
2440
+ "learning_rate": 0.0004947976254722072,
2441
+ "loss": 3.6681,
2442
+ "step": 16350
2443
+ },
2444
+ {
2445
+ "epoch": 1.7681940700808625,
2446
+ "grad_norm": 0.5325199365615845,
2447
+ "learning_rate": 0.0004944738262277387,
2448
+ "loss": 3.6466,
2449
+ "step": 16400
2450
+ },
2451
+ {
2452
+ "epoch": 1.7735849056603774,
2453
+ "grad_norm": 0.576173722743988,
2454
+ "learning_rate": 0.0004941500269832703,
2455
+ "loss": 3.6569,
2456
+ "step": 16450
2457
+ },
2458
+ {
2459
+ "epoch": 1.778975741239892,
2460
+ "grad_norm": 0.5102316737174988,
2461
+ "learning_rate": 0.0004938262277388019,
2462
+ "loss": 3.6425,
2463
+ "step": 16500
2464
+ },
2465
+ {
2466
+ "epoch": 1.784366576819407,
2467
+ "grad_norm": 0.5195789933204651,
2468
+ "learning_rate": 0.0004935024284943335,
2469
+ "loss": 3.6499,
2470
+ "step": 16550
2471
+ },
2472
+ {
2473
+ "epoch": 1.789757412398922,
2474
+ "grad_norm": 0.5175392031669617,
2475
+ "learning_rate": 0.000493178629249865,
2476
+ "loss": 3.6622,
2477
+ "step": 16600
2478
+ },
2479
+ {
2480
+ "epoch": 1.7951482479784366,
2481
+ "grad_norm": 0.5237419605255127,
2482
+ "learning_rate": 0.0004928548300053966,
2483
+ "loss": 3.6531,
2484
+ "step": 16650
2485
+ },
2486
+ {
2487
+ "epoch": 1.8005390835579513,
2488
+ "grad_norm": 0.6199319958686829,
2489
+ "learning_rate": 0.0004925310307609282,
2490
+ "loss": 3.66,
2491
+ "step": 16700
2492
+ },
2493
+ {
2494
+ "epoch": 1.8059299191374663,
2495
+ "grad_norm": 0.5926080346107483,
2496
+ "learning_rate": 0.0004922072315164598,
2497
+ "loss": 3.659,
2498
+ "step": 16750
2499
+ },
2500
+ {
2501
+ "epoch": 1.8113207547169812,
2502
+ "grad_norm": 0.5379118919372559,
2503
+ "learning_rate": 0.0004918834322719913,
2504
+ "loss": 3.6662,
2505
+ "step": 16800
2506
+ },
2507
+ {
2508
+ "epoch": 1.8167115902964959,
2509
+ "grad_norm": 0.6137577295303345,
2510
+ "learning_rate": 0.0004915596330275229,
2511
+ "loss": 3.6746,
2512
+ "step": 16850
2513
+ },
2514
+ {
2515
+ "epoch": 1.8221024258760108,
2516
+ "grad_norm": 0.5653969645500183,
2517
+ "learning_rate": 0.0004912358337830544,
2518
+ "loss": 3.6474,
2519
+ "step": 16900
2520
+ },
2521
+ {
2522
+ "epoch": 1.8274932614555257,
2523
+ "grad_norm": 0.56174635887146,
2524
+ "learning_rate": 0.000490912034538586,
2525
+ "loss": 3.6592,
2526
+ "step": 16950
2527
+ },
2528
+ {
2529
+ "epoch": 1.8328840970350404,
2530
+ "grad_norm": 0.5398900508880615,
2531
+ "learning_rate": 0.0004905882352941175,
2532
+ "loss": 3.6383,
2533
+ "step": 17000
2534
+ },
2535
+ {
2536
+ "epoch": 1.8328840970350404,
2537
+ "eval_accuracy": 0.3587294609371681,
2538
+ "eval_loss": 3.6051440238952637,
2539
+ "eval_runtime": 185.6742,
2540
+ "eval_samples_per_second": 97.003,
2541
+ "eval_steps_per_second": 6.064,
2542
+ "step": 17000
2543
+ },
2544
+ {
2545
+ "epoch": 1.838274932614555,
2546
+ "grad_norm": 0.5739971399307251,
2547
+ "learning_rate": 0.0004902644360496492,
2548
+ "loss": 3.6431,
2549
+ "step": 17050
2550
+ },
2551
+ {
2552
+ "epoch": 1.8436657681940702,
2553
+ "grad_norm": 0.5605867505073547,
2554
+ "learning_rate": 0.0004899406368051808,
2555
+ "loss": 3.6362,
2556
+ "step": 17100
2557
+ },
2558
+ {
2559
+ "epoch": 1.849056603773585,
2560
+ "grad_norm": 0.5735852718353271,
2561
+ "learning_rate": 0.0004896168375607123,
2562
+ "loss": 3.6416,
2563
+ "step": 17150
2564
+ },
2565
+ {
2566
+ "epoch": 1.8544474393530996,
2567
+ "grad_norm": 0.5647099614143372,
2568
+ "learning_rate": 0.0004892930383162439,
2569
+ "loss": 3.6467,
2570
+ "step": 17200
2571
+ },
2572
+ {
2573
+ "epoch": 1.8598382749326146,
2574
+ "grad_norm": 0.5226011276245117,
2575
+ "learning_rate": 0.0004889692390717754,
2576
+ "loss": 3.6534,
2577
+ "step": 17250
2578
+ },
2579
+ {
2580
+ "epoch": 1.8652291105121295,
2581
+ "grad_norm": 0.5790987610816956,
2582
+ "learning_rate": 0.0004886454398273071,
2583
+ "loss": 3.6564,
2584
+ "step": 17300
2585
+ },
2586
+ {
2587
+ "epoch": 1.8706199460916442,
2588
+ "grad_norm": 0.5586245059967041,
2589
+ "learning_rate": 0.0004883216405828386,
2590
+ "loss": 3.6568,
2591
+ "step": 17350
2592
+ },
2593
+ {
2594
+ "epoch": 1.8760107816711589,
2595
+ "grad_norm": 0.5110759139060974,
2596
+ "learning_rate": 0.00048799784133837017,
2597
+ "loss": 3.6508,
2598
+ "step": 17400
2599
+ },
2600
+ {
2601
+ "epoch": 1.881401617250674,
2602
+ "grad_norm": 0.5973682999610901,
2603
+ "learning_rate": 0.0004876740420939017,
2604
+ "loss": 3.6405,
2605
+ "step": 17450
2606
+ },
2607
+ {
2608
+ "epoch": 1.8867924528301887,
2609
+ "grad_norm": 0.5328883528709412,
2610
+ "learning_rate": 0.0004873502428494333,
2611
+ "loss": 3.6423,
2612
+ "step": 17500
2613
+ },
2614
+ {
2615
+ "epoch": 1.8921832884097034,
2616
+ "grad_norm": 0.6016663312911987,
2617
+ "learning_rate": 0.0004870264436049649,
2618
+ "loss": 3.6395,
2619
+ "step": 17550
2620
+ },
2621
+ {
2622
+ "epoch": 1.8975741239892183,
2623
+ "grad_norm": 0.5924084186553955,
2624
+ "learning_rate": 0.00048670264436049643,
2625
+ "loss": 3.6529,
2626
+ "step": 17600
2627
+ },
2628
+ {
2629
+ "epoch": 1.9029649595687332,
2630
+ "grad_norm": 0.5941994190216064,
2631
+ "learning_rate": 0.00048637884511602803,
2632
+ "loss": 3.6441,
2633
+ "step": 17650
2634
+ },
2635
+ {
2636
+ "epoch": 1.908355795148248,
2637
+ "grad_norm": 0.5356640219688416,
2638
+ "learning_rate": 0.0004860550458715596,
2639
+ "loss": 3.638,
2640
+ "step": 17700
2641
+ },
2642
+ {
2643
+ "epoch": 1.9137466307277629,
2644
+ "grad_norm": 0.6006718277931213,
2645
+ "learning_rate": 0.0004857312466270912,
2646
+ "loss": 3.6598,
2647
+ "step": 17750
2648
+ },
2649
+ {
2650
+ "epoch": 1.9191374663072778,
2651
+ "grad_norm": 0.5845404863357544,
2652
+ "learning_rate": 0.00048540744738262274,
2653
+ "loss": 3.6449,
2654
+ "step": 17800
2655
+ },
2656
+ {
2657
+ "epoch": 1.9245283018867925,
2658
+ "grad_norm": 0.5526080131530762,
2659
+ "learning_rate": 0.00048508364813815434,
2660
+ "loss": 3.6354,
2661
+ "step": 17850
2662
+ },
2663
+ {
2664
+ "epoch": 1.9299191374663072,
2665
+ "grad_norm": 0.5421946048736572,
2666
+ "learning_rate": 0.0004847663248785753,
2667
+ "loss": 3.6465,
2668
+ "step": 17900
2669
+ },
2670
+ {
2671
+ "epoch": 1.935309973045822,
2672
+ "grad_norm": 0.5944849848747253,
2673
+ "learning_rate": 0.0004844425256341068,
2674
+ "loss": 3.6269,
2675
+ "step": 17950
2676
+ },
2677
+ {
2678
+ "epoch": 1.940700808625337,
2679
+ "grad_norm": 0.5479409694671631,
2680
+ "learning_rate": 0.00048411872638963834,
2681
+ "loss": 3.6321,
2682
+ "step": 18000
2683
+ },
2684
+ {
2685
+ "epoch": 1.940700808625337,
2686
+ "eval_accuracy": 0.35996886442846787,
2687
+ "eval_loss": 3.5911636352539062,
2688
+ "eval_runtime": 185.6549,
2689
+ "eval_samples_per_second": 97.013,
2690
+ "eval_steps_per_second": 6.065,
2691
+ "step": 18000
2692
+ },
2693
+ {
2694
+ "epoch": 1.9460916442048517,
2695
+ "grad_norm": 0.5909312963485718,
2696
+ "learning_rate": 0.00048379492714516995,
2697
+ "loss": 3.6343,
2698
+ "step": 18050
2699
+ },
2700
+ {
2701
+ "epoch": 1.9514824797843666,
2702
+ "grad_norm": 0.5868159532546997,
2703
+ "learning_rate": 0.0004834711279007015,
2704
+ "loss": 3.6358,
2705
+ "step": 18100
2706
+ },
2707
+ {
2708
+ "epoch": 1.9568733153638815,
2709
+ "grad_norm": 0.5208358764648438,
2710
+ "learning_rate": 0.0004831473286562331,
2711
+ "loss": 3.6319,
2712
+ "step": 18150
2713
+ },
2714
+ {
2715
+ "epoch": 1.9622641509433962,
2716
+ "grad_norm": 0.5705773234367371,
2717
+ "learning_rate": 0.00048282352941176465,
2718
+ "loss": 3.6214,
2719
+ "step": 18200
2720
+ },
2721
+ {
2722
+ "epoch": 1.967654986522911,
2723
+ "grad_norm": 0.5833699703216553,
2724
+ "learning_rate": 0.00048249973016729626,
2725
+ "loss": 3.6428,
2726
+ "step": 18250
2727
+ },
2728
+ {
2729
+ "epoch": 1.9730458221024259,
2730
+ "grad_norm": 0.594480574131012,
2731
+ "learning_rate": 0.0004821759309228278,
2732
+ "loss": 3.6519,
2733
+ "step": 18300
2734
+ },
2735
+ {
2736
+ "epoch": 1.9784366576819408,
2737
+ "grad_norm": 0.5162302851676941,
2738
+ "learning_rate": 0.00048185213167835936,
2739
+ "loss": 3.6368,
2740
+ "step": 18350
2741
+ },
2742
+ {
2743
+ "epoch": 1.9838274932614555,
2744
+ "grad_norm": 0.6090728640556335,
2745
+ "learning_rate": 0.00048152833243389096,
2746
+ "loss": 3.6373,
2747
+ "step": 18400
2748
+ },
2749
+ {
2750
+ "epoch": 1.9892183288409704,
2751
+ "grad_norm": 0.5224543809890747,
2752
+ "learning_rate": 0.0004812045331894225,
2753
+ "loss": 3.6452,
2754
+ "step": 18450
2755
+ },
2756
+ {
2757
+ "epoch": 1.9946091644204853,
2758
+ "grad_norm": 0.6286558508872986,
2759
+ "learning_rate": 0.0004808807339449541,
2760
+ "loss": 3.6383,
2761
+ "step": 18500
2762
+ },
2763
+ {
2764
+ "epoch": 2.0,
2765
+ "grad_norm": 1.1893088817596436,
2766
+ "learning_rate": 0.0004805569347004856,
2767
+ "loss": 3.6417,
2768
+ "step": 18550
2769
+ },
2770
+ {
2771
+ "epoch": 2.0053908355795147,
2772
+ "grad_norm": 0.5484069585800171,
2773
+ "learning_rate": 0.0004802331354560173,
2774
+ "loss": 3.5707,
2775
+ "step": 18600
2776
+ },
2777
+ {
2778
+ "epoch": 2.01078167115903,
2779
+ "grad_norm": 0.5843707323074341,
2780
+ "learning_rate": 0.00047990933621154877,
2781
+ "loss": 3.5399,
2782
+ "step": 18650
2783
+ },
2784
+ {
2785
+ "epoch": 2.0161725067385445,
2786
+ "grad_norm": 0.5230047106742859,
2787
+ "learning_rate": 0.0004795855369670804,
2788
+ "loss": 3.5512,
2789
+ "step": 18700
2790
+ },
2791
+ {
2792
+ "epoch": 2.0215633423180592,
2793
+ "grad_norm": 0.5233069658279419,
2794
+ "learning_rate": 0.0004792617377226119,
2795
+ "loss": 3.5319,
2796
+ "step": 18750
2797
+ },
2798
+ {
2799
+ "epoch": 2.026954177897574,
2800
+ "grad_norm": 0.5597031712532043,
2801
+ "learning_rate": 0.0004789379384781435,
2802
+ "loss": 3.5518,
2803
+ "step": 18800
2804
+ },
2805
+ {
2806
+ "epoch": 2.032345013477089,
2807
+ "grad_norm": 0.5443568825721741,
2808
+ "learning_rate": 0.0004786141392336751,
2809
+ "loss": 3.5586,
2810
+ "step": 18850
2811
+ },
2812
+ {
2813
+ "epoch": 2.0377358490566038,
2814
+ "grad_norm": 0.5472878217697144,
2815
+ "learning_rate": 0.00047829033998920663,
2816
+ "loss": 3.5469,
2817
+ "step": 18900
2818
+ },
2819
+ {
2820
+ "epoch": 2.0431266846361185,
2821
+ "grad_norm": 0.5626387000083923,
2822
+ "learning_rate": 0.00047796654074473824,
2823
+ "loss": 3.5408,
2824
+ "step": 18950
2825
+ },
2826
+ {
2827
+ "epoch": 2.0485175202156336,
2828
+ "grad_norm": 0.5888155698776245,
2829
+ "learning_rate": 0.0004776427415002698,
2830
+ "loss": 3.5672,
2831
+ "step": 19000
2832
+ },
2833
+ {
2834
+ "epoch": 2.0485175202156336,
2835
+ "eval_accuracy": 0.361499131700447,
2836
+ "eval_loss": 3.5821456909179688,
2837
+ "eval_runtime": 185.3913,
2838
+ "eval_samples_per_second": 97.151,
2839
+ "eval_steps_per_second": 6.074,
2840
+ "step": 19000
2841
+ },
2842
+ {
2843
+ "epoch": 2.0539083557951483,
2844
+ "grad_norm": 0.5674854516983032,
2845
+ "learning_rate": 0.0004773189422558014,
2846
+ "loss": 3.5436,
2847
+ "step": 19050
2848
+ },
2849
+ {
2850
+ "epoch": 2.059299191374663,
2851
+ "grad_norm": 0.5584875345230103,
2852
+ "learning_rate": 0.00047699514301133294,
2853
+ "loss": 3.567,
2854
+ "step": 19100
2855
+ },
2856
+ {
2857
+ "epoch": 2.0646900269541777,
2858
+ "grad_norm": 0.5911595821380615,
2859
+ "learning_rate": 0.00047667134376686455,
2860
+ "loss": 3.5623,
2861
+ "step": 19150
2862
+ },
2863
+ {
2864
+ "epoch": 2.070080862533693,
2865
+ "grad_norm": 0.5857378840446472,
2866
+ "learning_rate": 0.0004763475445223961,
2867
+ "loss": 3.5702,
2868
+ "step": 19200
2869
+ },
2870
+ {
2871
+ "epoch": 2.0754716981132075,
2872
+ "grad_norm": 0.5929583311080933,
2873
+ "learning_rate": 0.0004760237452779276,
2874
+ "loss": 3.5495,
2875
+ "step": 19250
2876
+ },
2877
+ {
2878
+ "epoch": 2.0808625336927222,
2879
+ "grad_norm": 0.5647702813148499,
2880
+ "learning_rate": 0.0004756999460334592,
2881
+ "loss": 3.5433,
2882
+ "step": 19300
2883
+ },
2884
+ {
2885
+ "epoch": 2.0862533692722374,
2886
+ "grad_norm": 0.5857425332069397,
2887
+ "learning_rate": 0.00047537614678899075,
2888
+ "loss": 3.547,
2889
+ "step": 19350
2890
+ },
2891
+ {
2892
+ "epoch": 2.091644204851752,
2893
+ "grad_norm": 0.5703374743461609,
2894
+ "learning_rate": 0.00047505234754452235,
2895
+ "loss": 3.5461,
2896
+ "step": 19400
2897
+ },
2898
+ {
2899
+ "epoch": 2.0970350404312668,
2900
+ "grad_norm": 0.8500975966453552,
2901
+ "learning_rate": 0.0004747285483000539,
2902
+ "loss": 3.5538,
2903
+ "step": 19450
2904
+ },
2905
+ {
2906
+ "epoch": 2.1024258760107815,
2907
+ "grad_norm": 0.6051983833312988,
2908
+ "learning_rate": 0.0004744047490555855,
2909
+ "loss": 3.5454,
2910
+ "step": 19500
2911
+ },
2912
+ {
2913
+ "epoch": 2.1078167115902966,
2914
+ "grad_norm": 0.7593019008636475,
2915
+ "learning_rate": 0.00047408094981111706,
2916
+ "loss": 3.5555,
2917
+ "step": 19550
2918
+ },
2919
+ {
2920
+ "epoch": 2.1132075471698113,
2921
+ "grad_norm": 0.5925678014755249,
2922
+ "learning_rate": 0.000473763626551538,
2923
+ "loss": 3.5514,
2924
+ "step": 19600
2925
+ },
2926
+ {
2927
+ "epoch": 2.118598382749326,
2928
+ "grad_norm": 0.5761629343032837,
2929
+ "learning_rate": 0.00047343982730706956,
2930
+ "loss": 3.5705,
2931
+ "step": 19650
2932
+ },
2933
+ {
2934
+ "epoch": 2.123989218328841,
2935
+ "grad_norm": 0.5896238088607788,
2936
+ "learning_rate": 0.00047311602806260117,
2937
+ "loss": 3.5423,
2938
+ "step": 19700
2939
+ },
2940
+ {
2941
+ "epoch": 2.129380053908356,
2942
+ "grad_norm": 0.5506173968315125,
2943
+ "learning_rate": 0.0004727922288181327,
2944
+ "loss": 3.5619,
2945
+ "step": 19750
2946
+ },
2947
+ {
2948
+ "epoch": 2.1347708894878705,
2949
+ "grad_norm": 0.5747953653335571,
2950
+ "learning_rate": 0.0004724684295736643,
2951
+ "loss": 3.5549,
2952
+ "step": 19800
2953
+ },
2954
+ {
2955
+ "epoch": 2.1401617250673857,
2956
+ "grad_norm": 0.6055583953857422,
2957
+ "learning_rate": 0.0004721446303291959,
2958
+ "loss": 3.5466,
2959
+ "step": 19850
2960
+ },
2961
+ {
2962
+ "epoch": 2.1455525606469004,
2963
+ "grad_norm": 0.5854771733283997,
2964
+ "learning_rate": 0.0004718208310847275,
2965
+ "loss": 3.5586,
2966
+ "step": 19900
2967
+ },
2968
+ {
2969
+ "epoch": 2.150943396226415,
2970
+ "grad_norm": 0.5524225234985352,
2971
+ "learning_rate": 0.000471497031840259,
2972
+ "loss": 3.5506,
2973
+ "step": 19950
2974
+ },
2975
+ {
2976
+ "epoch": 2.1563342318059298,
2977
+ "grad_norm": 0.6449847221374512,
2978
+ "learning_rate": 0.0004711732325957905,
2979
+ "loss": 3.5754,
2980
+ "step": 20000
2981
+ },
2982
+ {
2983
+ "epoch": 2.1563342318059298,
2984
+ "eval_accuracy": 0.3623671596211969,
2985
+ "eval_loss": 3.573341131210327,
2986
+ "eval_runtime": 185.7691,
2987
+ "eval_samples_per_second": 96.954,
2988
+ "eval_steps_per_second": 6.061,
2989
+ "step": 20000
2990
+ }
2991
+ ],
2992
+ "logging_steps": 50,
2993
+ "max_steps": 92750,
2994
+ "num_input_tokens_seen": 0,
2995
+ "num_train_epochs": 10,
2996
+ "save_steps": 10000,
2997
+ "stateful_callbacks": {
2998
+ "TrainerControl": {
2999
+ "args": {
3000
+ "should_epoch_stop": false,
3001
+ "should_evaluate": false,
3002
+ "should_log": false,
3003
+ "should_save": true,
3004
+ "should_training_stop": false
3005
+ },
3006
+ "attributes": {}
3007
+ }
3008
+ },
3009
+ "total_flos": 1.672138358784e+17,
3010
+ "train_batch_size": 32,
3011
+ "trial_name": null,
3012
+ "trial_params": null
3013
+ }
checkpoint-20000/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4128eda3e9c8c97041ff17245f83cc1f26210a065e6d51188d1e8a5776110948
3
+ size 5304