DANGDOCAO commited on
Commit
49bed59
·
verified ·
1 Parent(s): 4476b4e
Files changed (49) hide show
  1. .gitattributes +12 -0
  2. HVU_QA/t5-viet-qg-finetuned/checkpoint-2000/added_tokens.json +98 -0
  3. HVU_QA/t5-viet-qg-finetuned/checkpoint-2000/config.json +31 -0
  4. HVU_QA/t5-viet-qg-finetuned/checkpoint-2000/generation_config.json +7 -0
  5. HVU_QA/t5-viet-qg-finetuned/checkpoint-2000/model.safetensors +3 -0
  6. HVU_QA/t5-viet-qg-finetuned/checkpoint-2000/optimizer.pt +3 -0
  7. HVU_QA/t5-viet-qg-finetuned/checkpoint-2000/rng_state.pth +0 -0
  8. HVU_QA/t5-viet-qg-finetuned/checkpoint-2000/scheduler.pt +0 -0
  9. HVU_QA/t5-viet-qg-finetuned/checkpoint-2000/special_tokens_map.json +121 -0
  10. HVU_QA/t5-viet-qg-finetuned/checkpoint-2000/spiece.model +3 -0
  11. HVU_QA/t5-viet-qg-finetuned/checkpoint-2000/tokenizer_config.json +905 -0
  12. HVU_QA/t5-viet-qg-finetuned/checkpoint-2000/trainer_state.json +1434 -0
  13. HVU_QA/t5-viet-qg-finetuned/checkpoint-2000/training_args.bin +0 -0
  14. HVU_QA/t5-viet-qg-finetuned/checkpoint-2500/added_tokens.json +98 -0
  15. HVU_QA/t5-viet-qg-finetuned/checkpoint-2500/config.json +31 -0
  16. HVU_QA/t5-viet-qg-finetuned/checkpoint-2500/generation_config.json +7 -0
  17. HVU_QA/t5-viet-qg-finetuned/checkpoint-2500/model.safetensors +3 -0
  18. HVU_QA/t5-viet-qg-finetuned/checkpoint-2500/optimizer.pt +3 -0
  19. HVU_QA/t5-viet-qg-finetuned/checkpoint-2500/rng_state.pth +0 -0
  20. HVU_QA/t5-viet-qg-finetuned/checkpoint-2500/scheduler.pt +0 -0
  21. HVU_QA/t5-viet-qg-finetuned/checkpoint-2500/special_tokens_map.json +121 -0
  22. HVU_QA/t5-viet-qg-finetuned/checkpoint-2500/spiece.model +3 -0
  23. HVU_QA/t5-viet-qg-finetuned/checkpoint-2500/tokenizer_config.json +905 -0
  24. HVU_QA/t5-viet-qg-finetuned/checkpoint-2500/trainer_state.json +1784 -0
  25. HVU_QA/t5-viet-qg-finetuned/checkpoint-2500/training_args.bin +0 -0
  26. HVU_QA/t5-viet-qg-finetuned/checkpoint-3000/added_tokens.json +98 -0
  27. HVU_QA/t5-viet-qg-finetuned/checkpoint-3000/config.json +31 -0
  28. HVU_QA/t5-viet-qg-finetuned/checkpoint-3000/generation_config.json +7 -0
  29. HVU_QA/t5-viet-qg-finetuned/checkpoint-3000/model.safetensors +3 -0
  30. HVU_QA/t5-viet-qg-finetuned/checkpoint-3000/optimizer.pt +3 -0
  31. HVU_QA/t5-viet-qg-finetuned/checkpoint-3000/rng_state.pth +0 -0
  32. HVU_QA/t5-viet-qg-finetuned/checkpoint-3000/scheduler.pt +0 -0
  33. HVU_QA/t5-viet-qg-finetuned/checkpoint-3000/special_tokens_map.json +121 -0
  34. HVU_QA/t5-viet-qg-finetuned/checkpoint-3000/spiece.model +3 -0
  35. HVU_QA/t5-viet-qg-finetuned/checkpoint-3000/tokenizer_config.json +905 -0
  36. HVU_QA/t5-viet-qg-finetuned/checkpoint-3000/trainer_state.json +2134 -0
  37. HVU_QA/t5-viet-qg-finetuned/checkpoint-3000/training_args.bin +0 -0
  38. HVU_QA/t5-viet-qg-finetuned/checkpoint-3500/added_tokens.json +98 -0
  39. HVU_QA/t5-viet-qg-finetuned/checkpoint-3500/config.json +31 -0
  40. HVU_QA/t5-viet-qg-finetuned/checkpoint-3500/generation_config.json +7 -0
  41. HVU_QA/t5-viet-qg-finetuned/checkpoint-3500/model.safetensors +3 -0
  42. HVU_QA/t5-viet-qg-finetuned/checkpoint-3500/optimizer.pt +3 -0
  43. HVU_QA/t5-viet-qg-finetuned/checkpoint-3500/rng_state.pth +0 -0
  44. HVU_QA/t5-viet-qg-finetuned/checkpoint-3500/scheduler.pt +0 -0
  45. HVU_QA/t5-viet-qg-finetuned/checkpoint-3500/special_tokens_map.json +121 -0
  46. HVU_QA/t5-viet-qg-finetuned/checkpoint-3500/spiece.model +3 -0
  47. HVU_QA/t5-viet-qg-finetuned/checkpoint-3500/tokenizer_config.json +905 -0
  48. HVU_QA/t5-viet-qg-finetuned/checkpoint-3500/trainer_state.json +2484 -0
  49. HVU_QA/t5-viet-qg-finetuned/checkpoint-3500/training_args.bin +0 -0
.gitattributes CHANGED
@@ -10,3 +10,15 @@ HVU_QA/t5-viet-qg-finetuned/checkpoint-1000/spiece.model filter=lfs diff=lfs mer
10
  HVU_QA/t5-viet-qg-finetuned/checkpoint-1500/model.safetensors filter=lfs diff=lfs merge=lfs -text
11
  HVU_QA/t5-viet-qg-finetuned/checkpoint-1500/optimizer.pt filter=lfs diff=lfs merge=lfs -text
12
  HVU_QA/t5-viet-qg-finetuned/checkpoint-1500/spiece.model filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
10
  HVU_QA/t5-viet-qg-finetuned/checkpoint-1500/model.safetensors filter=lfs diff=lfs merge=lfs -text
11
  HVU_QA/t5-viet-qg-finetuned/checkpoint-1500/optimizer.pt filter=lfs diff=lfs merge=lfs -text
12
  HVU_QA/t5-viet-qg-finetuned/checkpoint-1500/spiece.model filter=lfs diff=lfs merge=lfs -text
13
+ HVU_QA/t5-viet-qg-finetuned/checkpoint-2000/model.safetensors filter=lfs diff=lfs merge=lfs -text
14
+ HVU_QA/t5-viet-qg-finetuned/checkpoint-2000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
15
+ HVU_QA/t5-viet-qg-finetuned/checkpoint-2000/spiece.model filter=lfs diff=lfs merge=lfs -text
16
+ HVU_QA/t5-viet-qg-finetuned/checkpoint-2500/model.safetensors filter=lfs diff=lfs merge=lfs -text
17
+ HVU_QA/t5-viet-qg-finetuned/checkpoint-2500/optimizer.pt filter=lfs diff=lfs merge=lfs -text
18
+ HVU_QA/t5-viet-qg-finetuned/checkpoint-2500/spiece.model filter=lfs diff=lfs merge=lfs -text
19
+ HVU_QA/t5-viet-qg-finetuned/checkpoint-3000/model.safetensors filter=lfs diff=lfs merge=lfs -text
20
+ HVU_QA/t5-viet-qg-finetuned/checkpoint-3000/optimizer.pt filter=lfs diff=lfs merge=lfs -text
21
+ HVU_QA/t5-viet-qg-finetuned/checkpoint-3000/spiece.model filter=lfs diff=lfs merge=lfs -text
22
+ HVU_QA/t5-viet-qg-finetuned/checkpoint-3500/model.safetensors filter=lfs diff=lfs merge=lfs -text
23
+ HVU_QA/t5-viet-qg-finetuned/checkpoint-3500/optimizer.pt filter=lfs diff=lfs merge=lfs -text
24
+ HVU_QA/t5-viet-qg-finetuned/checkpoint-3500/spiece.model filter=lfs diff=lfs merge=lfs -text
HVU_QA/t5-viet-qg-finetuned/checkpoint-2000/added_tokens.json ADDED
@@ -0,0 +1,98 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "<extra_id_0>": 36095,
3
+ "<extra_id_10>": 36085,
4
+ "<extra_id_11>": 36084,
5
+ "<extra_id_12>": 36083,
6
+ "<extra_id_13>": 36082,
7
+ "<extra_id_14>": 36081,
8
+ "<extra_id_15>": 36080,
9
+ "<extra_id_16>": 36079,
10
+ "<extra_id_17>": 36078,
11
+ "<extra_id_18>": 36077,
12
+ "<extra_id_19>": 36076,
13
+ "<extra_id_1>": 36094,
14
+ "<extra_id_20>": 36075,
15
+ "<extra_id_21>": 36074,
16
+ "<extra_id_22>": 36073,
17
+ "<extra_id_23>": 36072,
18
+ "<extra_id_24>": 36071,
19
+ "<extra_id_25>": 36070,
20
+ "<extra_id_26>": 36069,
21
+ "<extra_id_27>": 36068,
22
+ "<extra_id_28>": 36067,
23
+ "<extra_id_29>": 36066,
24
+ "<extra_id_2>": 36093,
25
+ "<extra_id_30>": 36065,
26
+ "<extra_id_31>": 36064,
27
+ "<extra_id_32>": 36063,
28
+ "<extra_id_33>": 36062,
29
+ "<extra_id_34>": 36061,
30
+ "<extra_id_35>": 36060,
31
+ "<extra_id_36>": 36059,
32
+ "<extra_id_37>": 36058,
33
+ "<extra_id_38>": 36057,
34
+ "<extra_id_39>": 36056,
35
+ "<extra_id_3>": 36092,
36
+ "<extra_id_40>": 36055,
37
+ "<extra_id_41>": 36054,
38
+ "<extra_id_42>": 36053,
39
+ "<extra_id_43>": 36052,
40
+ "<extra_id_44>": 36051,
41
+ "<extra_id_45>": 36050,
42
+ "<extra_id_46>": 36049,
43
+ "<extra_id_47>": 36048,
44
+ "<extra_id_48>": 36047,
45
+ "<extra_id_49>": 36046,
46
+ "<extra_id_4>": 36091,
47
+ "<extra_id_50>": 36045,
48
+ "<extra_id_51>": 36044,
49
+ "<extra_id_52>": 36043,
50
+ "<extra_id_53>": 36042,
51
+ "<extra_id_54>": 36041,
52
+ "<extra_id_55>": 36040,
53
+ "<extra_id_56>": 36039,
54
+ "<extra_id_57>": 36038,
55
+ "<extra_id_58>": 36037,
56
+ "<extra_id_59>": 36036,
57
+ "<extra_id_5>": 36090,
58
+ "<extra_id_60>": 36035,
59
+ "<extra_id_61>": 36034,
60
+ "<extra_id_62>": 36033,
61
+ "<extra_id_63>": 36032,
62
+ "<extra_id_64>": 36031,
63
+ "<extra_id_65>": 36030,
64
+ "<extra_id_66>": 36029,
65
+ "<extra_id_67>": 36028,
66
+ "<extra_id_68>": 36027,
67
+ "<extra_id_69>": 36026,
68
+ "<extra_id_6>": 36089,
69
+ "<extra_id_70>": 36025,
70
+ "<extra_id_71>": 36024,
71
+ "<extra_id_72>": 36023,
72
+ "<extra_id_73>": 36022,
73
+ "<extra_id_74>": 36021,
74
+ "<extra_id_75>": 36020,
75
+ "<extra_id_76>": 36019,
76
+ "<extra_id_77>": 36018,
77
+ "<extra_id_78>": 36017,
78
+ "<extra_id_79>": 36016,
79
+ "<extra_id_7>": 36088,
80
+ "<extra_id_80>": 36015,
81
+ "<extra_id_81>": 36014,
82
+ "<extra_id_82>": 36013,
83
+ "<extra_id_83>": 36012,
84
+ "<extra_id_84>": 36011,
85
+ "<extra_id_85>": 36010,
86
+ "<extra_id_86>": 36009,
87
+ "<extra_id_87>": 36008,
88
+ "<extra_id_88>": 36007,
89
+ "<extra_id_89>": 36006,
90
+ "<extra_id_8>": 36087,
91
+ "<extra_id_90>": 36005,
92
+ "<extra_id_91>": 36004,
93
+ "<extra_id_92>": 36003,
94
+ "<extra_id_93>": 36002,
95
+ "<extra_id_94>": 36001,
96
+ "<extra_id_95>": 36000,
97
+ "<extra_id_9>": 36086
98
+ }
HVU_QA/t5-viet-qg-finetuned/checkpoint-2000/config.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "T5ForConditionalGeneration"
4
+ ],
5
+ "classifier_dropout": 0.0,
6
+ "d_ff": 3072,
7
+ "d_kv": 64,
8
+ "d_model": 768,
9
+ "decoder_start_token_id": 0,
10
+ "dense_act_fn": "relu",
11
+ "dropout_rate": 0.1,
12
+ "eos_token_id": 1,
13
+ "feed_forward_proj": "relu",
14
+ "initializer_factor": 1.0,
15
+ "is_encoder_decoder": true,
16
+ "is_gated_act": false,
17
+ "layer_norm_epsilon": 1e-06,
18
+ "model_type": "t5",
19
+ "n_positions": 512,
20
+ "num_decoder_layers": 12,
21
+ "num_heads": 12,
22
+ "num_layers": 12,
23
+ "output_past": true,
24
+ "pad_token_id": 0,
25
+ "relative_attention_max_distance": 128,
26
+ "relative_attention_num_buckets": 32,
27
+ "torch_dtype": "float32",
28
+ "transformers_version": "4.53.2",
29
+ "use_cache": true,
30
+ "vocab_size": 36096
31
+ }
HVU_QA/t5-viet-qg-finetuned/checkpoint-2000/generation_config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "decoder_start_token_id": 0,
4
+ "eos_token_id": 1,
5
+ "pad_token_id": 0,
6
+ "transformers_version": "4.53.2"
7
+ }
HVU_QA/t5-viet-qg-finetuned/checkpoint-2000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:814d141277d4902c872e99cbb6db6bd36071b4eaa5dcac9f14835eeebeec0790
3
+ size 62652416
HVU_QA/t5-viet-qg-finetuned/checkpoint-2000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4ed9374a4ceb529c13848dd09ef350fc4fab3c2493c12fdf4bef559e9fc2ce67
3
+ size 62652416
HVU_QA/t5-viet-qg-finetuned/checkpoint-2000/rng_state.pth ADDED
Binary file (14.6 kB). View file
 
HVU_QA/t5-viet-qg-finetuned/checkpoint-2000/scheduler.pt ADDED
Binary file (1.47 kB). View file
 
HVU_QA/t5-viet-qg-finetuned/checkpoint-2000/special_tokens_map.json ADDED
@@ -0,0 +1,121 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<extra_id_0>",
4
+ "<extra_id_1>",
5
+ "<extra_id_2>",
6
+ "<extra_id_3>",
7
+ "<extra_id_4>",
8
+ "<extra_id_5>",
9
+ "<extra_id_6>",
10
+ "<extra_id_7>",
11
+ "<extra_id_8>",
12
+ "<extra_id_9>",
13
+ "<extra_id_10>",
14
+ "<extra_id_11>",
15
+ "<extra_id_12>",
16
+ "<extra_id_13>",
17
+ "<extra_id_14>",
18
+ "<extra_id_15>",
19
+ "<extra_id_16>",
20
+ "<extra_id_17>",
21
+ "<extra_id_18>",
22
+ "<extra_id_19>",
23
+ "<extra_id_20>",
24
+ "<extra_id_21>",
25
+ "<extra_id_22>",
26
+ "<extra_id_23>",
27
+ "<extra_id_24>",
28
+ "<extra_id_25>",
29
+ "<extra_id_26>",
30
+ "<extra_id_27>",
31
+ "<extra_id_28>",
32
+ "<extra_id_29>",
33
+ "<extra_id_30>",
34
+ "<extra_id_31>",
35
+ "<extra_id_32>",
36
+ "<extra_id_33>",
37
+ "<extra_id_34>",
38
+ "<extra_id_35>",
39
+ "<extra_id_36>",
40
+ "<extra_id_37>",
41
+ "<extra_id_38>",
42
+ "<extra_id_39>",
43
+ "<extra_id_40>",
44
+ "<extra_id_41>",
45
+ "<extra_id_42>",
46
+ "<extra_id_43>",
47
+ "<extra_id_44>",
48
+ "<extra_id_45>",
49
+ "<extra_id_46>",
50
+ "<extra_id_47>",
51
+ "<extra_id_48>",
52
+ "<extra_id_49>",
53
+ "<extra_id_50>",
54
+ "<extra_id_51>",
55
+ "<extra_id_52>",
56
+ "<extra_id_53>",
57
+ "<extra_id_54>",
58
+ "<extra_id_55>",
59
+ "<extra_id_56>",
60
+ "<extra_id_57>",
61
+ "<extra_id_58>",
62
+ "<extra_id_59>",
63
+ "<extra_id_60>",
64
+ "<extra_id_61>",
65
+ "<extra_id_62>",
66
+ "<extra_id_63>",
67
+ "<extra_id_64>",
68
+ "<extra_id_65>",
69
+ "<extra_id_66>",
70
+ "<extra_id_67>",
71
+ "<extra_id_68>",
72
+ "<extra_id_69>",
73
+ "<extra_id_70>",
74
+ "<extra_id_71>",
75
+ "<extra_id_72>",
76
+ "<extra_id_73>",
77
+ "<extra_id_74>",
78
+ "<extra_id_75>",
79
+ "<extra_id_76>",
80
+ "<extra_id_77>",
81
+ "<extra_id_78>",
82
+ "<extra_id_79>",
83
+ "<extra_id_80>",
84
+ "<extra_id_81>",
85
+ "<extra_id_82>",
86
+ "<extra_id_83>",
87
+ "<extra_id_84>",
88
+ "<extra_id_85>",
89
+ "<extra_id_86>",
90
+ "<extra_id_87>",
91
+ "<extra_id_88>",
92
+ "<extra_id_89>",
93
+ "<extra_id_90>",
94
+ "<extra_id_91>",
95
+ "<extra_id_92>",
96
+ "<extra_id_93>",
97
+ "<extra_id_94>",
98
+ "<extra_id_95>"
99
+ ],
100
+ "eos_token": {
101
+ "content": "</s>",
102
+ "lstrip": false,
103
+ "normalized": false,
104
+ "rstrip": false,
105
+ "single_word": false
106
+ },
107
+ "pad_token": {
108
+ "content": "<pad>",
109
+ "lstrip": false,
110
+ "normalized": false,
111
+ "rstrip": false,
112
+ "single_word": false
113
+ },
114
+ "unk_token": {
115
+ "content": "<unk>",
116
+ "lstrip": false,
117
+ "normalized": false,
118
+ "rstrip": false,
119
+ "single_word": false
120
+ }
121
+ }
HVU_QA/t5-viet-qg-finetuned/checkpoint-2000/spiece.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6
3
+ size 820370
HVU_QA/t5-viet-qg-finetuned/checkpoint-2000/tokenizer_config.json ADDED
@@ -0,0 +1,905 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": true,
3
+ "added_tokens_decoder": {
4
+ "0": {
5
+ "content": "<pad>",
6
+ "lstrip": false,
7
+ "normalized": false,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "1": {
13
+ "content": "</s>",
14
+ "lstrip": false,
15
+ "normalized": false,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
+ },
20
+ "2": {
21
+ "content": "<unk>",
22
+ "lstrip": false,
23
+ "normalized": false,
24
+ "rstrip": false,
25
+ "single_word": false,
26
+ "special": true
27
+ },
28
+ "36000": {
29
+ "content": "<extra_id_95>",
30
+ "lstrip": true,
31
+ "normalized": false,
32
+ "rstrip": true,
33
+ "single_word": false,
34
+ "special": true
35
+ },
36
+ "36001": {
37
+ "content": "<extra_id_94>",
38
+ "lstrip": true,
39
+ "normalized": false,
40
+ "rstrip": true,
41
+ "single_word": false,
42
+ "special": true
43
+ },
44
+ "36002": {
45
+ "content": "<extra_id_93>",
46
+ "lstrip": true,
47
+ "normalized": false,
48
+ "rstrip": true,
49
+ "single_word": false,
50
+ "special": true
51
+ },
52
+ "36003": {
53
+ "content": "<extra_id_92>",
54
+ "lstrip": true,
55
+ "normalized": false,
56
+ "rstrip": true,
57
+ "single_word": false,
58
+ "special": true
59
+ },
60
+ "36004": {
61
+ "content": "<extra_id_91>",
62
+ "lstrip": true,
63
+ "normalized": false,
64
+ "rstrip": true,
65
+ "single_word": false,
66
+ "special": true
67
+ },
68
+ "36005": {
69
+ "content": "<extra_id_90>",
70
+ "lstrip": true,
71
+ "normalized": false,
72
+ "rstrip": true,
73
+ "single_word": false,
74
+ "special": true
75
+ },
76
+ "36006": {
77
+ "content": "<extra_id_89>",
78
+ "lstrip": true,
79
+ "normalized": false,
80
+ "rstrip": true,
81
+ "single_word": false,
82
+ "special": true
83
+ },
84
+ "36007": {
85
+ "content": "<extra_id_88>",
86
+ "lstrip": true,
87
+ "normalized": false,
88
+ "rstrip": true,
89
+ "single_word": false,
90
+ "special": true
91
+ },
92
+ "36008": {
93
+ "content": "<extra_id_87>",
94
+ "lstrip": true,
95
+ "normalized": false,
96
+ "rstrip": true,
97
+ "single_word": false,
98
+ "special": true
99
+ },
100
+ "36009": {
101
+ "content": "<extra_id_86>",
102
+ "lstrip": true,
103
+ "normalized": false,
104
+ "rstrip": true,
105
+ "single_word": false,
106
+ "special": true
107
+ },
108
+ "36010": {
109
+ "content": "<extra_id_85>",
110
+ "lstrip": true,
111
+ "normalized": false,
112
+ "rstrip": true,
113
+ "single_word": false,
114
+ "special": true
115
+ },
116
+ "36011": {
117
+ "content": "<extra_id_84>",
118
+ "lstrip": true,
119
+ "normalized": false,
120
+ "rstrip": true,
121
+ "single_word": false,
122
+ "special": true
123
+ },
124
+ "36012": {
125
+ "content": "<extra_id_83>",
126
+ "lstrip": true,
127
+ "normalized": false,
128
+ "rstrip": true,
129
+ "single_word": false,
130
+ "special": true
131
+ },
132
+ "36013": {
133
+ "content": "<extra_id_82>",
134
+ "lstrip": true,
135
+ "normalized": false,
136
+ "rstrip": true,
137
+ "single_word": false,
138
+ "special": true
139
+ },
140
+ "36014": {
141
+ "content": "<extra_id_81>",
142
+ "lstrip": true,
143
+ "normalized": false,
144
+ "rstrip": true,
145
+ "single_word": false,
146
+ "special": true
147
+ },
148
+ "36015": {
149
+ "content": "<extra_id_80>",
150
+ "lstrip": true,
151
+ "normalized": false,
152
+ "rstrip": true,
153
+ "single_word": false,
154
+ "special": true
155
+ },
156
+ "36016": {
157
+ "content": "<extra_id_79>",
158
+ "lstrip": true,
159
+ "normalized": false,
160
+ "rstrip": true,
161
+ "single_word": false,
162
+ "special": true
163
+ },
164
+ "36017": {
165
+ "content": "<extra_id_78>",
166
+ "lstrip": true,
167
+ "normalized": false,
168
+ "rstrip": true,
169
+ "single_word": false,
170
+ "special": true
171
+ },
172
+ "36018": {
173
+ "content": "<extra_id_77>",
174
+ "lstrip": true,
175
+ "normalized": false,
176
+ "rstrip": true,
177
+ "single_word": false,
178
+ "special": true
179
+ },
180
+ "36019": {
181
+ "content": "<extra_id_76>",
182
+ "lstrip": true,
183
+ "normalized": false,
184
+ "rstrip": true,
185
+ "single_word": false,
186
+ "special": true
187
+ },
188
+ "36020": {
189
+ "content": "<extra_id_75>",
190
+ "lstrip": true,
191
+ "normalized": false,
192
+ "rstrip": true,
193
+ "single_word": false,
194
+ "special": true
195
+ },
196
+ "36021": {
197
+ "content": "<extra_id_74>",
198
+ "lstrip": true,
199
+ "normalized": false,
200
+ "rstrip": true,
201
+ "single_word": false,
202
+ "special": true
203
+ },
204
+ "36022": {
205
+ "content": "<extra_id_73>",
206
+ "lstrip": true,
207
+ "normalized": false,
208
+ "rstrip": true,
209
+ "single_word": false,
210
+ "special": true
211
+ },
212
+ "36023": {
213
+ "content": "<extra_id_72>",
214
+ "lstrip": true,
215
+ "normalized": false,
216
+ "rstrip": true,
217
+ "single_word": false,
218
+ "special": true
219
+ },
220
+ "36024": {
221
+ "content": "<extra_id_71>",
222
+ "lstrip": true,
223
+ "normalized": false,
224
+ "rstrip": true,
225
+ "single_word": false,
226
+ "special": true
227
+ },
228
+ "36025": {
229
+ "content": "<extra_id_70>",
230
+ "lstrip": true,
231
+ "normalized": false,
232
+ "rstrip": true,
233
+ "single_word": false,
234
+ "special": true
235
+ },
236
+ "36026": {
237
+ "content": "<extra_id_69>",
238
+ "lstrip": true,
239
+ "normalized": false,
240
+ "rstrip": true,
241
+ "single_word": false,
242
+ "special": true
243
+ },
244
+ "36027": {
245
+ "content": "<extra_id_68>",
246
+ "lstrip": true,
247
+ "normalized": false,
248
+ "rstrip": true,
249
+ "single_word": false,
250
+ "special": true
251
+ },
252
+ "36028": {
253
+ "content": "<extra_id_67>",
254
+ "lstrip": true,
255
+ "normalized": false,
256
+ "rstrip": true,
257
+ "single_word": false,
258
+ "special": true
259
+ },
260
+ "36029": {
261
+ "content": "<extra_id_66>",
262
+ "lstrip": true,
263
+ "normalized": false,
264
+ "rstrip": true,
265
+ "single_word": false,
266
+ "special": true
267
+ },
268
+ "36030": {
269
+ "content": "<extra_id_65>",
270
+ "lstrip": true,
271
+ "normalized": false,
272
+ "rstrip": true,
273
+ "single_word": false,
274
+ "special": true
275
+ },
276
+ "36031": {
277
+ "content": "<extra_id_64>",
278
+ "lstrip": true,
279
+ "normalized": false,
280
+ "rstrip": true,
281
+ "single_word": false,
282
+ "special": true
283
+ },
284
+ "36032": {
285
+ "content": "<extra_id_63>",
286
+ "lstrip": true,
287
+ "normalized": false,
288
+ "rstrip": true,
289
+ "single_word": false,
290
+ "special": true
291
+ },
292
+ "36033": {
293
+ "content": "<extra_id_62>",
294
+ "lstrip": true,
295
+ "normalized": false,
296
+ "rstrip": true,
297
+ "single_word": false,
298
+ "special": true
299
+ },
300
+ "36034": {
301
+ "content": "<extra_id_61>",
302
+ "lstrip": true,
303
+ "normalized": false,
304
+ "rstrip": true,
305
+ "single_word": false,
306
+ "special": true
307
+ },
308
+ "36035": {
309
+ "content": "<extra_id_60>",
310
+ "lstrip": true,
311
+ "normalized": false,
312
+ "rstrip": true,
313
+ "single_word": false,
314
+ "special": true
315
+ },
316
+ "36036": {
317
+ "content": "<extra_id_59>",
318
+ "lstrip": true,
319
+ "normalized": false,
320
+ "rstrip": true,
321
+ "single_word": false,
322
+ "special": true
323
+ },
324
+ "36037": {
325
+ "content": "<extra_id_58>",
326
+ "lstrip": true,
327
+ "normalized": false,
328
+ "rstrip": true,
329
+ "single_word": false,
330
+ "special": true
331
+ },
332
+ "36038": {
333
+ "content": "<extra_id_57>",
334
+ "lstrip": true,
335
+ "normalized": false,
336
+ "rstrip": true,
337
+ "single_word": false,
338
+ "special": true
339
+ },
340
+ "36039": {
341
+ "content": "<extra_id_56>",
342
+ "lstrip": true,
343
+ "normalized": false,
344
+ "rstrip": true,
345
+ "single_word": false,
346
+ "special": true
347
+ },
348
+ "36040": {
349
+ "content": "<extra_id_55>",
350
+ "lstrip": true,
351
+ "normalized": false,
352
+ "rstrip": true,
353
+ "single_word": false,
354
+ "special": true
355
+ },
356
+ "36041": {
357
+ "content": "<extra_id_54>",
358
+ "lstrip": true,
359
+ "normalized": false,
360
+ "rstrip": true,
361
+ "single_word": false,
362
+ "special": true
363
+ },
364
+ "36042": {
365
+ "content": "<extra_id_53>",
366
+ "lstrip": true,
367
+ "normalized": false,
368
+ "rstrip": true,
369
+ "single_word": false,
370
+ "special": true
371
+ },
372
+ "36043": {
373
+ "content": "<extra_id_52>",
374
+ "lstrip": true,
375
+ "normalized": false,
376
+ "rstrip": true,
377
+ "single_word": false,
378
+ "special": true
379
+ },
380
+ "36044": {
381
+ "content": "<extra_id_51>",
382
+ "lstrip": true,
383
+ "normalized": false,
384
+ "rstrip": true,
385
+ "single_word": false,
386
+ "special": true
387
+ },
388
+ "36045": {
389
+ "content": "<extra_id_50>",
390
+ "lstrip": true,
391
+ "normalized": false,
392
+ "rstrip": true,
393
+ "single_word": false,
394
+ "special": true
395
+ },
396
+ "36046": {
397
+ "content": "<extra_id_49>",
398
+ "lstrip": true,
399
+ "normalized": false,
400
+ "rstrip": true,
401
+ "single_word": false,
402
+ "special": true
403
+ },
404
+ "36047": {
405
+ "content": "<extra_id_48>",
406
+ "lstrip": true,
407
+ "normalized": false,
408
+ "rstrip": true,
409
+ "single_word": false,
410
+ "special": true
411
+ },
412
+ "36048": {
413
+ "content": "<extra_id_47>",
414
+ "lstrip": true,
415
+ "normalized": false,
416
+ "rstrip": true,
417
+ "single_word": false,
418
+ "special": true
419
+ },
420
+ "36049": {
421
+ "content": "<extra_id_46>",
422
+ "lstrip": true,
423
+ "normalized": false,
424
+ "rstrip": true,
425
+ "single_word": false,
426
+ "special": true
427
+ },
428
+ "36050": {
429
+ "content": "<extra_id_45>",
430
+ "lstrip": true,
431
+ "normalized": false,
432
+ "rstrip": true,
433
+ "single_word": false,
434
+ "special": true
435
+ },
436
+ "36051": {
437
+ "content": "<extra_id_44>",
438
+ "lstrip": true,
439
+ "normalized": false,
440
+ "rstrip": true,
441
+ "single_word": false,
442
+ "special": true
443
+ },
444
+ "36052": {
445
+ "content": "<extra_id_43>",
446
+ "lstrip": true,
447
+ "normalized": false,
448
+ "rstrip": true,
449
+ "single_word": false,
450
+ "special": true
451
+ },
452
+ "36053": {
453
+ "content": "<extra_id_42>",
454
+ "lstrip": true,
455
+ "normalized": false,
456
+ "rstrip": true,
457
+ "single_word": false,
458
+ "special": true
459
+ },
460
+ "36054": {
461
+ "content": "<extra_id_41>",
462
+ "lstrip": true,
463
+ "normalized": false,
464
+ "rstrip": true,
465
+ "single_word": false,
466
+ "special": true
467
+ },
468
+ "36055": {
469
+ "content": "<extra_id_40>",
470
+ "lstrip": true,
471
+ "normalized": false,
472
+ "rstrip": true,
473
+ "single_word": false,
474
+ "special": true
475
+ },
476
+ "36056": {
477
+ "content": "<extra_id_39>",
478
+ "lstrip": true,
479
+ "normalized": false,
480
+ "rstrip": true,
481
+ "single_word": false,
482
+ "special": true
483
+ },
484
+ "36057": {
485
+ "content": "<extra_id_38>",
486
+ "lstrip": true,
487
+ "normalized": false,
488
+ "rstrip": true,
489
+ "single_word": false,
490
+ "special": true
491
+ },
492
+ "36058": {
493
+ "content": "<extra_id_37>",
494
+ "lstrip": true,
495
+ "normalized": false,
496
+ "rstrip": true,
497
+ "single_word": false,
498
+ "special": true
499
+ },
500
+ "36059": {
501
+ "content": "<extra_id_36>",
502
+ "lstrip": true,
503
+ "normalized": false,
504
+ "rstrip": true,
505
+ "single_word": false,
506
+ "special": true
507
+ },
508
+ "36060": {
509
+ "content": "<extra_id_35>",
510
+ "lstrip": true,
511
+ "normalized": false,
512
+ "rstrip": true,
513
+ "single_word": false,
514
+ "special": true
515
+ },
516
+ "36061": {
517
+ "content": "<extra_id_34>",
518
+ "lstrip": true,
519
+ "normalized": false,
520
+ "rstrip": true,
521
+ "single_word": false,
522
+ "special": true
523
+ },
524
+ "36062": {
525
+ "content": "<extra_id_33>",
526
+ "lstrip": true,
527
+ "normalized": false,
528
+ "rstrip": true,
529
+ "single_word": false,
530
+ "special": true
531
+ },
532
+ "36063": {
533
+ "content": "<extra_id_32>",
534
+ "lstrip": true,
535
+ "normalized": false,
536
+ "rstrip": true,
537
+ "single_word": false,
538
+ "special": true
539
+ },
540
+ "36064": {
541
+ "content": "<extra_id_31>",
542
+ "lstrip": true,
543
+ "normalized": false,
544
+ "rstrip": true,
545
+ "single_word": false,
546
+ "special": true
547
+ },
548
+ "36065": {
549
+ "content": "<extra_id_30>",
550
+ "lstrip": true,
551
+ "normalized": false,
552
+ "rstrip": true,
553
+ "single_word": false,
554
+ "special": true
555
+ },
556
+ "36066": {
557
+ "content": "<extra_id_29>",
558
+ "lstrip": true,
559
+ "normalized": false,
560
+ "rstrip": true,
561
+ "single_word": false,
562
+ "special": true
563
+ },
564
+ "36067": {
565
+ "content": "<extra_id_28>",
566
+ "lstrip": true,
567
+ "normalized": false,
568
+ "rstrip": true,
569
+ "single_word": false,
570
+ "special": true
571
+ },
572
+ "36068": {
573
+ "content": "<extra_id_27>",
574
+ "lstrip": true,
575
+ "normalized": false,
576
+ "rstrip": true,
577
+ "single_word": false,
578
+ "special": true
579
+ },
580
+ "36069": {
581
+ "content": "<extra_id_26>",
582
+ "lstrip": true,
583
+ "normalized": false,
584
+ "rstrip": true,
585
+ "single_word": false,
586
+ "special": true
587
+ },
588
+ "36070": {
589
+ "content": "<extra_id_25>",
590
+ "lstrip": true,
591
+ "normalized": false,
592
+ "rstrip": true,
593
+ "single_word": false,
594
+ "special": true
595
+ },
596
+ "36071": {
597
+ "content": "<extra_id_24>",
598
+ "lstrip": true,
599
+ "normalized": false,
600
+ "rstrip": true,
601
+ "single_word": false,
602
+ "special": true
603
+ },
604
+ "36072": {
605
+ "content": "<extra_id_23>",
606
+ "lstrip": true,
607
+ "normalized": false,
608
+ "rstrip": true,
609
+ "single_word": false,
610
+ "special": true
611
+ },
612
+ "36073": {
613
+ "content": "<extra_id_22>",
614
+ "lstrip": true,
615
+ "normalized": false,
616
+ "rstrip": true,
617
+ "single_word": false,
618
+ "special": true
619
+ },
620
+ "36074": {
621
+ "content": "<extra_id_21>",
622
+ "lstrip": true,
623
+ "normalized": false,
624
+ "rstrip": true,
625
+ "single_word": false,
626
+ "special": true
627
+ },
628
+ "36075": {
629
+ "content": "<extra_id_20>",
630
+ "lstrip": true,
631
+ "normalized": false,
632
+ "rstrip": true,
633
+ "single_word": false,
634
+ "special": true
635
+ },
636
+ "36076": {
637
+ "content": "<extra_id_19>",
638
+ "lstrip": true,
639
+ "normalized": false,
640
+ "rstrip": true,
641
+ "single_word": false,
642
+ "special": true
643
+ },
644
+ "36077": {
645
+ "content": "<extra_id_18>",
646
+ "lstrip": true,
647
+ "normalized": false,
648
+ "rstrip": true,
649
+ "single_word": false,
650
+ "special": true
651
+ },
652
+ "36078": {
653
+ "content": "<extra_id_17>",
654
+ "lstrip": true,
655
+ "normalized": false,
656
+ "rstrip": true,
657
+ "single_word": false,
658
+ "special": true
659
+ },
660
+ "36079": {
661
+ "content": "<extra_id_16>",
662
+ "lstrip": true,
663
+ "normalized": false,
664
+ "rstrip": true,
665
+ "single_word": false,
666
+ "special": true
667
+ },
668
+ "36080": {
669
+ "content": "<extra_id_15>",
670
+ "lstrip": true,
671
+ "normalized": false,
672
+ "rstrip": true,
673
+ "single_word": false,
674
+ "special": true
675
+ },
676
+ "36081": {
677
+ "content": "<extra_id_14>",
678
+ "lstrip": true,
679
+ "normalized": false,
680
+ "rstrip": true,
681
+ "single_word": false,
682
+ "special": true
683
+ },
684
+ "36082": {
685
+ "content": "<extra_id_13>",
686
+ "lstrip": true,
687
+ "normalized": false,
688
+ "rstrip": true,
689
+ "single_word": false,
690
+ "special": true
691
+ },
692
+ "36083": {
693
+ "content": "<extra_id_12>",
694
+ "lstrip": true,
695
+ "normalized": false,
696
+ "rstrip": true,
697
+ "single_word": false,
698
+ "special": true
699
+ },
700
+ "36084": {
701
+ "content": "<extra_id_11>",
702
+ "lstrip": true,
703
+ "normalized": false,
704
+ "rstrip": true,
705
+ "single_word": false,
706
+ "special": true
707
+ },
708
+ "36085": {
709
+ "content": "<extra_id_10>",
710
+ "lstrip": true,
711
+ "normalized": false,
712
+ "rstrip": true,
713
+ "single_word": false,
714
+ "special": true
715
+ },
716
+ "36086": {
717
+ "content": "<extra_id_9>",
718
+ "lstrip": true,
719
+ "normalized": false,
720
+ "rstrip": true,
721
+ "single_word": false,
722
+ "special": true
723
+ },
724
+ "36087": {
725
+ "content": "<extra_id_8>",
726
+ "lstrip": true,
727
+ "normalized": false,
728
+ "rstrip": true,
729
+ "single_word": false,
730
+ "special": true
731
+ },
732
+ "36088": {
733
+ "content": "<extra_id_7>",
734
+ "lstrip": true,
735
+ "normalized": false,
736
+ "rstrip": true,
737
+ "single_word": false,
738
+ "special": true
739
+ },
740
+ "36089": {
741
+ "content": "<extra_id_6>",
742
+ "lstrip": true,
743
+ "normalized": false,
744
+ "rstrip": true,
745
+ "single_word": false,
746
+ "special": true
747
+ },
748
+ "36090": {
749
+ "content": "<extra_id_5>",
750
+ "lstrip": true,
751
+ "normalized": false,
752
+ "rstrip": true,
753
+ "single_word": false,
754
+ "special": true
755
+ },
756
+ "36091": {
757
+ "content": "<extra_id_4>",
758
+ "lstrip": true,
759
+ "normalized": false,
760
+ "rstrip": true,
761
+ "single_word": false,
762
+ "special": true
763
+ },
764
+ "36092": {
765
+ "content": "<extra_id_3>",
766
+ "lstrip": true,
767
+ "normalized": false,
768
+ "rstrip": true,
769
+ "single_word": false,
770
+ "special": true
771
+ },
772
+ "36093": {
773
+ "content": "<extra_id_2>",
774
+ "lstrip": true,
775
+ "normalized": false,
776
+ "rstrip": true,
777
+ "single_word": false,
778
+ "special": true
779
+ },
780
+ "36094": {
781
+ "content": "<extra_id_1>",
782
+ "lstrip": true,
783
+ "normalized": false,
784
+ "rstrip": true,
785
+ "single_word": false,
786
+ "special": true
787
+ },
788
+ "36095": {
789
+ "content": "<extra_id_0>",
790
+ "lstrip": true,
791
+ "normalized": false,
792
+ "rstrip": true,
793
+ "single_word": false,
794
+ "special": true
795
+ }
796
+ },
797
+ "additional_special_tokens": [
798
+ "<extra_id_0>",
799
+ "<extra_id_1>",
800
+ "<extra_id_2>",
801
+ "<extra_id_3>",
802
+ "<extra_id_4>",
803
+ "<extra_id_5>",
804
+ "<extra_id_6>",
805
+ "<extra_id_7>",
806
+ "<extra_id_8>",
807
+ "<extra_id_9>",
808
+ "<extra_id_10>",
809
+ "<extra_id_11>",
810
+ "<extra_id_12>",
811
+ "<extra_id_13>",
812
+ "<extra_id_14>",
813
+ "<extra_id_15>",
814
+ "<extra_id_16>",
815
+ "<extra_id_17>",
816
+ "<extra_id_18>",
817
+ "<extra_id_19>",
818
+ "<extra_id_20>",
819
+ "<extra_id_21>",
820
+ "<extra_id_22>",
821
+ "<extra_id_23>",
822
+ "<extra_id_24>",
823
+ "<extra_id_25>",
824
+ "<extra_id_26>",
825
+ "<extra_id_27>",
826
+ "<extra_id_28>",
827
+ "<extra_id_29>",
828
+ "<extra_id_30>",
829
+ "<extra_id_31>",
830
+ "<extra_id_32>",
831
+ "<extra_id_33>",
832
+ "<extra_id_34>",
833
+ "<extra_id_35>",
834
+ "<extra_id_36>",
835
+ "<extra_id_37>",
836
+ "<extra_id_38>",
837
+ "<extra_id_39>",
838
+ "<extra_id_40>",
839
+ "<extra_id_41>",
840
+ "<extra_id_42>",
841
+ "<extra_id_43>",
842
+ "<extra_id_44>",
843
+ "<extra_id_45>",
844
+ "<extra_id_46>",
845
+ "<extra_id_47>",
846
+ "<extra_id_48>",
847
+ "<extra_id_49>",
848
+ "<extra_id_50>",
849
+ "<extra_id_51>",
850
+ "<extra_id_52>",
851
+ "<extra_id_53>",
852
+ "<extra_id_54>",
853
+ "<extra_id_55>",
854
+ "<extra_id_56>",
855
+ "<extra_id_57>",
856
+ "<extra_id_58>",
857
+ "<extra_id_59>",
858
+ "<extra_id_60>",
859
+ "<extra_id_61>",
860
+ "<extra_id_62>",
861
+ "<extra_id_63>",
862
+ "<extra_id_64>",
863
+ "<extra_id_65>",
864
+ "<extra_id_66>",
865
+ "<extra_id_67>",
866
+ "<extra_id_68>",
867
+ "<extra_id_69>",
868
+ "<extra_id_70>",
869
+ "<extra_id_71>",
870
+ "<extra_id_72>",
871
+ "<extra_id_73>",
872
+ "<extra_id_74>",
873
+ "<extra_id_75>",
874
+ "<extra_id_76>",
875
+ "<extra_id_77>",
876
+ "<extra_id_78>",
877
+ "<extra_id_79>",
878
+ "<extra_id_80>",
879
+ "<extra_id_81>",
880
+ "<extra_id_82>",
881
+ "<extra_id_83>",
882
+ "<extra_id_84>",
883
+ "<extra_id_85>",
884
+ "<extra_id_86>",
885
+ "<extra_id_87>",
886
+ "<extra_id_88>",
887
+ "<extra_id_89>",
888
+ "<extra_id_90>",
889
+ "<extra_id_91>",
890
+ "<extra_id_92>",
891
+ "<extra_id_93>",
892
+ "<extra_id_94>",
893
+ "<extra_id_95>"
894
+ ],
895
+ "clean_up_tokenization_spaces": false,
896
+ "eos_token": "</s>",
897
+ "extra_ids": 96,
898
+ "extra_special_tokens": {},
899
+ "legacy": true,
900
+ "model_max_length": 1000000000000000019884624838656,
901
+ "pad_token": "<pad>",
902
+ "sp_model_kwargs": {},
903
+ "tokenizer_class": "T5Tokenizer",
904
+ "unk_token": "<unk>"
905
+ }
HVU_QA/t5-viet-qg-finetuned/checkpoint-2000/trainer_state.json ADDED
@@ -0,0 +1,1434 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 0.031197354464341423,
6
+ "eval_steps": 500,
7
+ "global_step": 2000,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.00015598677232170712,
14
+ "grad_norm": 3.216968059539795,
15
+ "learning_rate": 0.0001999906407936607,
16
+ "loss": 4.7744,
17
+ "step": 10
18
+ },
19
+ {
20
+ "epoch": 0.00031197354464341425,
21
+ "grad_norm": 9.366360664367676,
22
+ "learning_rate": 0.00019998024167550594,
23
+ "loss": 0.9639,
24
+ "step": 20
25
+ },
26
+ {
27
+ "epoch": 0.00046796031696512137,
28
+ "grad_norm": 4.382473945617676,
29
+ "learning_rate": 0.00019996984255735113,
30
+ "loss": 0.6338,
31
+ "step": 30
32
+ },
33
+ {
34
+ "epoch": 0.0006239470892868285,
35
+ "grad_norm": 14.415914535522461,
36
+ "learning_rate": 0.00019995944343919637,
37
+ "loss": 0.4962,
38
+ "step": 40
39
+ },
40
+ {
41
+ "epoch": 0.0007799338616085356,
42
+ "grad_norm": 7.440260410308838,
43
+ "learning_rate": 0.00019994904432104157,
44
+ "loss": 0.4619,
45
+ "step": 50
46
+ },
47
+ {
48
+ "epoch": 0.0009359206339302427,
49
+ "grad_norm": 27.18022918701172,
50
+ "learning_rate": 0.00019993864520288681,
51
+ "loss": 0.5335,
52
+ "step": 60
53
+ },
54
+ {
55
+ "epoch": 0.0010919074062519499,
56
+ "grad_norm": 7.954453468322754,
57
+ "learning_rate": 0.000199928246084732,
58
+ "loss": 0.8085,
59
+ "step": 70
60
+ },
61
+ {
62
+ "epoch": 0.001247894178573657,
63
+ "grad_norm": 4.762608528137207,
64
+ "learning_rate": 0.00019991784696657725,
65
+ "loss": 0.6333,
66
+ "step": 80
67
+ },
68
+ {
69
+ "epoch": 0.0014038809508953641,
70
+ "grad_norm": 9.577717781066895,
71
+ "learning_rate": 0.00019990744784842245,
72
+ "loss": 0.4558,
73
+ "step": 90
74
+ },
75
+ {
76
+ "epoch": 0.0015598677232170712,
77
+ "grad_norm": 7.586806297302246,
78
+ "learning_rate": 0.0001998970487302677,
79
+ "loss": 0.3227,
80
+ "step": 100
81
+ },
82
+ {
83
+ "epoch": 0.0017158544955387784,
84
+ "grad_norm": 4.262249946594238,
85
+ "learning_rate": 0.00019988664961211288,
86
+ "loss": 0.3261,
87
+ "step": 110
88
+ },
89
+ {
90
+ "epoch": 0.0018718412678604855,
91
+ "grad_norm": 6.639514923095703,
92
+ "learning_rate": 0.00019987625049395813,
93
+ "loss": 0.6788,
94
+ "step": 120
95
+ },
96
+ {
97
+ "epoch": 0.0020278280401821924,
98
+ "grad_norm": 3.7448127269744873,
99
+ "learning_rate": 0.00019986585137580332,
100
+ "loss": 0.4224,
101
+ "step": 130
102
+ },
103
+ {
104
+ "epoch": 0.0021838148125038997,
105
+ "grad_norm": 13.205582618713379,
106
+ "learning_rate": 0.00019985545225764857,
107
+ "loss": 0.4074,
108
+ "step": 140
109
+ },
110
+ {
111
+ "epoch": 0.0023398015848256066,
112
+ "grad_norm": 5.55623722076416,
113
+ "learning_rate": 0.00019984505313949376,
114
+ "loss": 0.3408,
115
+ "step": 150
116
+ },
117
+ {
118
+ "epoch": 0.002495788357147314,
119
+ "grad_norm": 10.253294944763184,
120
+ "learning_rate": 0.000199834654021339,
121
+ "loss": 0.4513,
122
+ "step": 160
123
+ },
124
+ {
125
+ "epoch": 0.002651775129469021,
126
+ "grad_norm": 12.778948783874512,
127
+ "learning_rate": 0.0001998242549031842,
128
+ "loss": 0.4287,
129
+ "step": 170
130
+ },
131
+ {
132
+ "epoch": 0.0028077619017907282,
133
+ "grad_norm": 8.961956977844238,
134
+ "learning_rate": 0.00019981385578502945,
135
+ "loss": 0.6746,
136
+ "step": 180
137
+ },
138
+ {
139
+ "epoch": 0.002963748674112435,
140
+ "grad_norm": 6.611206531524658,
141
+ "learning_rate": 0.00019980345666687467,
142
+ "loss": 0.4106,
143
+ "step": 190
144
+ },
145
+ {
146
+ "epoch": 0.0031197354464341425,
147
+ "grad_norm": 4.179671764373779,
148
+ "learning_rate": 0.0001997930575487199,
149
+ "loss": 0.4234,
150
+ "step": 200
151
+ },
152
+ {
153
+ "epoch": 0.0032757222187558494,
154
+ "grad_norm": 7.957104206085205,
155
+ "learning_rate": 0.0001997826584305651,
156
+ "loss": 0.4062,
157
+ "step": 210
158
+ },
159
+ {
160
+ "epoch": 0.0034317089910775567,
161
+ "grad_norm": 10.04617977142334,
162
+ "learning_rate": 0.00019977225931241033,
163
+ "loss": 0.4792,
164
+ "step": 220
165
+ },
166
+ {
167
+ "epoch": 0.0035876957633992636,
168
+ "grad_norm": 5.119971752166748,
169
+ "learning_rate": 0.00019976186019425555,
170
+ "loss": 0.3385,
171
+ "step": 230
172
+ },
173
+ {
174
+ "epoch": 0.003743682535720971,
175
+ "grad_norm": 6.039366722106934,
176
+ "learning_rate": 0.00019975146107610077,
177
+ "loss": 0.4406,
178
+ "step": 240
179
+ },
180
+ {
181
+ "epoch": 0.003899669308042678,
182
+ "grad_norm": 9.345914840698242,
183
+ "learning_rate": 0.000199741061957946,
184
+ "loss": 0.5163,
185
+ "step": 250
186
+ },
187
+ {
188
+ "epoch": 0.004055656080364385,
189
+ "grad_norm": 3.707460641860962,
190
+ "learning_rate": 0.0001997306628397912,
191
+ "loss": 0.2971,
192
+ "step": 260
193
+ },
194
+ {
195
+ "epoch": 0.0042116428526860926,
196
+ "grad_norm": 11.500152587890625,
197
+ "learning_rate": 0.00019972026372163643,
198
+ "loss": 0.3659,
199
+ "step": 270
200
+ },
201
+ {
202
+ "epoch": 0.0043676296250077995,
203
+ "grad_norm": 0.7893226742744446,
204
+ "learning_rate": 0.00019970986460348165,
205
+ "loss": 0.5052,
206
+ "step": 280
207
+ },
208
+ {
209
+ "epoch": 0.004523616397329506,
210
+ "grad_norm": 3.2410888671875,
211
+ "learning_rate": 0.00019969946548532687,
212
+ "loss": 0.4038,
213
+ "step": 290
214
+ },
215
+ {
216
+ "epoch": 0.004679603169651213,
217
+ "grad_norm": 4.468445777893066,
218
+ "learning_rate": 0.00019968906636717208,
219
+ "loss": 0.6076,
220
+ "step": 300
221
+ },
222
+ {
223
+ "epoch": 0.004835589941972921,
224
+ "grad_norm": 8.308616638183594,
225
+ "learning_rate": 0.0001996786672490173,
226
+ "loss": 0.4465,
227
+ "step": 310
228
+ },
229
+ {
230
+ "epoch": 0.004991576714294628,
231
+ "grad_norm": 3.736629009246826,
232
+ "learning_rate": 0.0001996682681308625,
233
+ "loss": 0.4133,
234
+ "step": 320
235
+ },
236
+ {
237
+ "epoch": 0.005147563486616335,
238
+ "grad_norm": 2.157092809677124,
239
+ "learning_rate": 0.00019965786901270774,
240
+ "loss": 0.3663,
241
+ "step": 330
242
+ },
243
+ {
244
+ "epoch": 0.005303550258938042,
245
+ "grad_norm": 5.357859134674072,
246
+ "learning_rate": 0.00019964746989455294,
247
+ "loss": 0.5599,
248
+ "step": 340
249
+ },
250
+ {
251
+ "epoch": 0.0054595370312597495,
252
+ "grad_norm": 5.352982044219971,
253
+ "learning_rate": 0.00019963707077639818,
254
+ "loss": 0.6796,
255
+ "step": 350
256
+ },
257
+ {
258
+ "epoch": 0.0056155238035814565,
259
+ "grad_norm": 3.7645204067230225,
260
+ "learning_rate": 0.00019962667165824338,
261
+ "loss": 0.5774,
262
+ "step": 360
263
+ },
264
+ {
265
+ "epoch": 0.005771510575903163,
266
+ "grad_norm": 6.222738265991211,
267
+ "learning_rate": 0.00019961627254008862,
268
+ "loss": 0.641,
269
+ "step": 370
270
+ },
271
+ {
272
+ "epoch": 0.00592749734822487,
273
+ "grad_norm": 6.460178852081299,
274
+ "learning_rate": 0.00019960587342193381,
275
+ "loss": 0.3702,
276
+ "step": 380
277
+ },
278
+ {
279
+ "epoch": 0.006083484120546578,
280
+ "grad_norm": 3.1196858882904053,
281
+ "learning_rate": 0.00019959547430377906,
282
+ "loss": 0.5036,
283
+ "step": 390
284
+ },
285
+ {
286
+ "epoch": 0.006239470892868285,
287
+ "grad_norm": 7.03672981262207,
288
+ "learning_rate": 0.00019958507518562425,
289
+ "loss": 0.41,
290
+ "step": 400
291
+ },
292
+ {
293
+ "epoch": 0.006395457665189992,
294
+ "grad_norm": 15.546775817871094,
295
+ "learning_rate": 0.0001995746760674695,
296
+ "loss": 0.5623,
297
+ "step": 410
298
+ },
299
+ {
300
+ "epoch": 0.006551444437511699,
301
+ "grad_norm": 4.868813514709473,
302
+ "learning_rate": 0.0001995642769493147,
303
+ "loss": 0.4,
304
+ "step": 420
305
+ },
306
+ {
307
+ "epoch": 0.0067074312098334065,
308
+ "grad_norm": 3.5767409801483154,
309
+ "learning_rate": 0.00019955387783115994,
310
+ "loss": 0.5193,
311
+ "step": 430
312
+ },
313
+ {
314
+ "epoch": 0.0068634179821551134,
315
+ "grad_norm": 10.07199764251709,
316
+ "learning_rate": 0.00019954347871300513,
317
+ "loss": 0.5692,
318
+ "step": 440
319
+ },
320
+ {
321
+ "epoch": 0.00701940475447682,
322
+ "grad_norm": 2.809025764465332,
323
+ "learning_rate": 0.00019953307959485038,
324
+ "loss": 0.4361,
325
+ "step": 450
326
+ },
327
+ {
328
+ "epoch": 0.007175391526798527,
329
+ "grad_norm": 14.244531631469727,
330
+ "learning_rate": 0.00019952268047669557,
331
+ "loss": 0.3646,
332
+ "step": 460
333
+ },
334
+ {
335
+ "epoch": 0.007331378299120235,
336
+ "grad_norm": 5.827071666717529,
337
+ "learning_rate": 0.00019951228135854082,
338
+ "loss": 0.5302,
339
+ "step": 470
340
+ },
341
+ {
342
+ "epoch": 0.007487365071441942,
343
+ "grad_norm": 9.067925453186035,
344
+ "learning_rate": 0.000199501882240386,
345
+ "loss": 0.5045,
346
+ "step": 480
347
+ },
348
+ {
349
+ "epoch": 0.007643351843763649,
350
+ "grad_norm": 8.674382209777832,
351
+ "learning_rate": 0.00019949148312223126,
352
+ "loss": 0.5778,
353
+ "step": 490
354
+ },
355
+ {
356
+ "epoch": 0.007799338616085356,
357
+ "grad_norm": 8.94734001159668,
358
+ "learning_rate": 0.00019948108400407645,
359
+ "loss": 0.4141,
360
+ "step": 500
361
+ },
362
+ {
363
+ "epoch": 0.007955325388407063,
364
+ "grad_norm": 10.810057640075684,
365
+ "learning_rate": 0.0001994706848859217,
366
+ "loss": 0.4069,
367
+ "step": 510
368
+ },
369
+ {
370
+ "epoch": 0.00811131216072877,
371
+ "grad_norm": 9.531766891479492,
372
+ "learning_rate": 0.0001994602857677669,
373
+ "loss": 0.5935,
374
+ "step": 520
375
+ },
376
+ {
377
+ "epoch": 0.008267298933050476,
378
+ "grad_norm": 13.168464660644531,
379
+ "learning_rate": 0.00019944988664961214,
380
+ "loss": 0.4725,
381
+ "step": 530
382
+ },
383
+ {
384
+ "epoch": 0.008423285705372185,
385
+ "grad_norm": 4.697941780090332,
386
+ "learning_rate": 0.00019943948753145733,
387
+ "loss": 0.4516,
388
+ "step": 540
389
+ },
390
+ {
391
+ "epoch": 0.008579272477693892,
392
+ "grad_norm": 2.92124342918396,
393
+ "learning_rate": 0.00019942908841330258,
394
+ "loss": 0.4562,
395
+ "step": 550
396
+ },
397
+ {
398
+ "epoch": 0.008735259250015599,
399
+ "grad_norm": 9.23039722442627,
400
+ "learning_rate": 0.00019941868929514777,
401
+ "loss": 0.4051,
402
+ "step": 560
403
+ },
404
+ {
405
+ "epoch": 0.008891246022337306,
406
+ "grad_norm": 10.528753280639648,
407
+ "learning_rate": 0.00019940829017699302,
408
+ "loss": 0.3475,
409
+ "step": 570
410
+ },
411
+ {
412
+ "epoch": 0.009047232794659013,
413
+ "grad_norm": 6.963693618774414,
414
+ "learning_rate": 0.0001993978910588382,
415
+ "loss": 0.5338,
416
+ "step": 580
417
+ },
418
+ {
419
+ "epoch": 0.00920321956698072,
420
+ "grad_norm": 6.09390926361084,
421
+ "learning_rate": 0.00019938749194068345,
422
+ "loss": 0.6501,
423
+ "step": 590
424
+ },
425
+ {
426
+ "epoch": 0.009359206339302427,
427
+ "grad_norm": 4.319295406341553,
428
+ "learning_rate": 0.00019937709282252865,
429
+ "loss": 0.4163,
430
+ "step": 600
431
+ },
432
+ {
433
+ "epoch": 0.009515193111624133,
434
+ "grad_norm": 7.514628887176514,
435
+ "learning_rate": 0.0001993666937043739,
436
+ "loss": 0.5193,
437
+ "step": 610
438
+ },
439
+ {
440
+ "epoch": 0.009671179883945842,
441
+ "grad_norm": 5.011287212371826,
442
+ "learning_rate": 0.00019935629458621909,
443
+ "loss": 0.5743,
444
+ "step": 620
445
+ },
446
+ {
447
+ "epoch": 0.009827166656267549,
448
+ "grad_norm": 6.090816974639893,
449
+ "learning_rate": 0.00019934589546806433,
450
+ "loss": 0.3902,
451
+ "step": 630
452
+ },
453
+ {
454
+ "epoch": 0.009983153428589256,
455
+ "grad_norm": 10.161015510559082,
456
+ "learning_rate": 0.00019933549634990953,
457
+ "loss": 0.4463,
458
+ "step": 640
459
+ },
460
+ {
461
+ "epoch": 0.010139140200910963,
462
+ "grad_norm": 6.14881706237793,
463
+ "learning_rate": 0.00019932509723175477,
464
+ "loss": 0.5009,
465
+ "step": 650
466
+ },
467
+ {
468
+ "epoch": 0.01029512697323267,
469
+ "grad_norm": 3.2960498332977295,
470
+ "learning_rate": 0.00019931469811359996,
471
+ "loss": 0.4597,
472
+ "step": 660
473
+ },
474
+ {
475
+ "epoch": 0.010451113745554377,
476
+ "grad_norm": 7.595890998840332,
477
+ "learning_rate": 0.0001993042989954452,
478
+ "loss": 0.5948,
479
+ "step": 670
480
+ },
481
+ {
482
+ "epoch": 0.010607100517876084,
483
+ "grad_norm": 7.178264617919922,
484
+ "learning_rate": 0.0001992938998772904,
485
+ "loss": 0.7185,
486
+ "step": 680
487
+ },
488
+ {
489
+ "epoch": 0.01076308729019779,
490
+ "grad_norm": 4.092066764831543,
491
+ "learning_rate": 0.00019928350075913565,
492
+ "loss": 0.3544,
493
+ "step": 690
494
+ },
495
+ {
496
+ "epoch": 0.010919074062519499,
497
+ "grad_norm": 13.203527450561523,
498
+ "learning_rate": 0.00019927310164098084,
499
+ "loss": 0.5719,
500
+ "step": 700
501
+ },
502
+ {
503
+ "epoch": 0.011075060834841206,
504
+ "grad_norm": 5.39072847366333,
505
+ "learning_rate": 0.0001992627025228261,
506
+ "loss": 0.5527,
507
+ "step": 710
508
+ },
509
+ {
510
+ "epoch": 0.011231047607162913,
511
+ "grad_norm": 5.769185543060303,
512
+ "learning_rate": 0.00019925230340467128,
513
+ "loss": 0.4555,
514
+ "step": 720
515
+ },
516
+ {
517
+ "epoch": 0.01138703437948462,
518
+ "grad_norm": 3.0080807209014893,
519
+ "learning_rate": 0.00019924190428651653,
520
+ "loss": 0.277,
521
+ "step": 730
522
+ },
523
+ {
524
+ "epoch": 0.011543021151806327,
525
+ "grad_norm": 1.0591176748275757,
526
+ "learning_rate": 0.00019923150516836172,
527
+ "loss": 0.513,
528
+ "step": 740
529
+ },
530
+ {
531
+ "epoch": 0.011699007924128034,
532
+ "grad_norm": 8.458100318908691,
533
+ "learning_rate": 0.00019922110605020697,
534
+ "loss": 0.5038,
535
+ "step": 750
536
+ },
537
+ {
538
+ "epoch": 0.01185499469644974,
539
+ "grad_norm": 3.121551990509033,
540
+ "learning_rate": 0.00019921070693205216,
541
+ "loss": 0.2639,
542
+ "step": 760
543
+ },
544
+ {
545
+ "epoch": 0.012010981468771447,
546
+ "grad_norm": 6.6408610343933105,
547
+ "learning_rate": 0.00019920030781389738,
548
+ "loss": 0.6632,
549
+ "step": 770
550
+ },
551
+ {
552
+ "epoch": 0.012166968241093156,
553
+ "grad_norm": 7.180611610412598,
554
+ "learning_rate": 0.0001991899086957426,
555
+ "loss": 0.5173,
556
+ "step": 780
557
+ },
558
+ {
559
+ "epoch": 0.012322955013414863,
560
+ "grad_norm": 4.404491901397705,
561
+ "learning_rate": 0.00019917950957758782,
562
+ "loss": 0.3628,
563
+ "step": 790
564
+ },
565
+ {
566
+ "epoch": 0.01247894178573657,
567
+ "grad_norm": 5.945431232452393,
568
+ "learning_rate": 0.00019916911045943304,
569
+ "loss": 0.3574,
570
+ "step": 800
571
+ },
572
+ {
573
+ "epoch": 0.012634928558058277,
574
+ "grad_norm": 5.105412483215332,
575
+ "learning_rate": 0.00019915871134127826,
576
+ "loss": 0.7038,
577
+ "step": 810
578
+ },
579
+ {
580
+ "epoch": 0.012790915330379984,
581
+ "grad_norm": 3.7237210273742676,
582
+ "learning_rate": 0.00019914831222312348,
583
+ "loss": 0.4426,
584
+ "step": 820
585
+ },
586
+ {
587
+ "epoch": 0.01294690210270169,
588
+ "grad_norm": 10.387605667114258,
589
+ "learning_rate": 0.0001991379131049687,
590
+ "loss": 0.3152,
591
+ "step": 830
592
+ },
593
+ {
594
+ "epoch": 0.013102888875023398,
595
+ "grad_norm": 2.875211477279663,
596
+ "learning_rate": 0.00019912751398681392,
597
+ "loss": 0.424,
598
+ "step": 840
599
+ },
600
+ {
601
+ "epoch": 0.013258875647345104,
602
+ "grad_norm": 1.0501248836517334,
603
+ "learning_rate": 0.00019911711486865914,
604
+ "loss": 0.5571,
605
+ "step": 850
606
+ },
607
+ {
608
+ "epoch": 0.013414862419666813,
609
+ "grad_norm": 6.748145580291748,
610
+ "learning_rate": 0.00019910671575050436,
611
+ "loss": 0.5226,
612
+ "step": 860
613
+ },
614
+ {
615
+ "epoch": 0.01357084919198852,
616
+ "grad_norm": 6.29374885559082,
617
+ "learning_rate": 0.00019909631663234958,
618
+ "loss": 0.3036,
619
+ "step": 870
620
+ },
621
+ {
622
+ "epoch": 0.013726835964310227,
623
+ "grad_norm": 5.8310699462890625,
624
+ "learning_rate": 0.0001990859175141948,
625
+ "loss": 0.4704,
626
+ "step": 880
627
+ },
628
+ {
629
+ "epoch": 0.013882822736631934,
630
+ "grad_norm": 1.4619086980819702,
631
+ "learning_rate": 0.00019907551839604002,
632
+ "loss": 0.3639,
633
+ "step": 890
634
+ },
635
+ {
636
+ "epoch": 0.01403880950895364,
637
+ "grad_norm": 8.177449226379395,
638
+ "learning_rate": 0.00019906511927788524,
639
+ "loss": 0.3853,
640
+ "step": 900
641
+ },
642
+ {
643
+ "epoch": 0.014194796281275348,
644
+ "grad_norm": 2.045549154281616,
645
+ "learning_rate": 0.00019905472015973046,
646
+ "loss": 0.3728,
647
+ "step": 910
648
+ },
649
+ {
650
+ "epoch": 0.014350783053597055,
651
+ "grad_norm": 6.4138875007629395,
652
+ "learning_rate": 0.00019904432104157568,
653
+ "loss": 0.4615,
654
+ "step": 920
655
+ },
656
+ {
657
+ "epoch": 0.014506769825918761,
658
+ "grad_norm": 2.442739486694336,
659
+ "learning_rate": 0.0001990339219234209,
660
+ "loss": 0.3592,
661
+ "step": 930
662
+ },
663
+ {
664
+ "epoch": 0.01466275659824047,
665
+ "grad_norm": 3.0741333961486816,
666
+ "learning_rate": 0.00019902352280526611,
667
+ "loss": 0.4118,
668
+ "step": 940
669
+ },
670
+ {
671
+ "epoch": 0.014818743370562177,
672
+ "grad_norm": 5.163224220275879,
673
+ "learning_rate": 0.00019901312368711133,
674
+ "loss": 0.7433,
675
+ "step": 950
676
+ },
677
+ {
678
+ "epoch": 0.014974730142883884,
679
+ "grad_norm": 5.6979193687438965,
680
+ "learning_rate": 0.00019900272456895655,
681
+ "loss": 0.3094,
682
+ "step": 960
683
+ },
684
+ {
685
+ "epoch": 0.01513071691520559,
686
+ "grad_norm": 3.2027347087860107,
687
+ "learning_rate": 0.00019899232545080177,
688
+ "loss": 0.3749,
689
+ "step": 970
690
+ },
691
+ {
692
+ "epoch": 0.015286703687527298,
693
+ "grad_norm": 9.097466468811035,
694
+ "learning_rate": 0.000198981926332647,
695
+ "loss": 0.5344,
696
+ "step": 980
697
+ },
698
+ {
699
+ "epoch": 0.015442690459849005,
700
+ "grad_norm": 3.730250835418701,
701
+ "learning_rate": 0.0001989715272144922,
702
+ "loss": 0.4085,
703
+ "step": 990
704
+ },
705
+ {
706
+ "epoch": 0.015598677232170712,
707
+ "grad_norm": 4.03507661819458,
708
+ "learning_rate": 0.00019896112809633743,
709
+ "loss": 0.7935,
710
+ "step": 1000
711
+ },
712
+ {
713
+ "epoch": 0.01575466400449242,
714
+ "grad_norm": 6.676806449890137,
715
+ "learning_rate": 0.00019895072897818265,
716
+ "loss": 0.4276,
717
+ "step": 1010
718
+ },
719
+ {
720
+ "epoch": 0.015910650776814125,
721
+ "grad_norm": 2.294931173324585,
722
+ "learning_rate": 0.00019894032986002787,
723
+ "loss": 0.2544,
724
+ "step": 1020
725
+ },
726
+ {
727
+ "epoch": 0.016066637549135832,
728
+ "grad_norm": 15.660760879516602,
729
+ "learning_rate": 0.0001989299307418731,
730
+ "loss": 0.2441,
731
+ "step": 1030
732
+ },
733
+ {
734
+ "epoch": 0.01622262432145754,
735
+ "grad_norm": 3.4223055839538574,
736
+ "learning_rate": 0.00019891953162371834,
737
+ "loss": 0.4861,
738
+ "step": 1040
739
+ },
740
+ {
741
+ "epoch": 0.016378611093779246,
742
+ "grad_norm": 8.601675987243652,
743
+ "learning_rate": 0.00019890913250556353,
744
+ "loss": 0.7212,
745
+ "step": 1050
746
+ },
747
+ {
748
+ "epoch": 0.016534597866100953,
749
+ "grad_norm": 3.814286708831787,
750
+ "learning_rate": 0.00019889873338740878,
751
+ "loss": 0.4785,
752
+ "step": 1060
753
+ },
754
+ {
755
+ "epoch": 0.016690584638422663,
756
+ "grad_norm": 10.6279296875,
757
+ "learning_rate": 0.00019888833426925397,
758
+ "loss": 0.278,
759
+ "step": 1070
760
+ },
761
+ {
762
+ "epoch": 0.01684657141074437,
763
+ "grad_norm": 2.3790910243988037,
764
+ "learning_rate": 0.00019887793515109922,
765
+ "loss": 0.3589,
766
+ "step": 1080
767
+ },
768
+ {
769
+ "epoch": 0.017002558183066077,
770
+ "grad_norm": 8.939804077148438,
771
+ "learning_rate": 0.0001988675360329444,
772
+ "loss": 0.5794,
773
+ "step": 1090
774
+ },
775
+ {
776
+ "epoch": 0.017158544955387784,
777
+ "grad_norm": 7.628971576690674,
778
+ "learning_rate": 0.00019885713691478966,
779
+ "loss": 0.5317,
780
+ "step": 1100
781
+ },
782
+ {
783
+ "epoch": 0.01731453172770949,
784
+ "grad_norm": 1.8617459535598755,
785
+ "learning_rate": 0.00019884673779663485,
786
+ "loss": 0.433,
787
+ "step": 1110
788
+ },
789
+ {
790
+ "epoch": 0.017470518500031198,
791
+ "grad_norm": 6.928651809692383,
792
+ "learning_rate": 0.0001988363386784801,
793
+ "loss": 0.216,
794
+ "step": 1120
795
+ },
796
+ {
797
+ "epoch": 0.017626505272352905,
798
+ "grad_norm": 3.6206703186035156,
799
+ "learning_rate": 0.0001988259395603253,
800
+ "loss": 0.4963,
801
+ "step": 1130
802
+ },
803
+ {
804
+ "epoch": 0.01778249204467461,
805
+ "grad_norm": 4.615293025970459,
806
+ "learning_rate": 0.00019881554044217053,
807
+ "loss": 0.3579,
808
+ "step": 1140
809
+ },
810
+ {
811
+ "epoch": 0.01793847881699632,
812
+ "grad_norm": 3.3206489086151123,
813
+ "learning_rate": 0.00019880514132401573,
814
+ "loss": 0.3717,
815
+ "step": 1150
816
+ },
817
+ {
818
+ "epoch": 0.018094465589318025,
819
+ "grad_norm": 7.5789408683776855,
820
+ "learning_rate": 0.00019879474220586097,
821
+ "loss": 0.2646,
822
+ "step": 1160
823
+ },
824
+ {
825
+ "epoch": 0.018250452361639732,
826
+ "grad_norm": 4.467035293579102,
827
+ "learning_rate": 0.00019878434308770617,
828
+ "loss": 0.1804,
829
+ "step": 1170
830
+ },
831
+ {
832
+ "epoch": 0.01840643913396144,
833
+ "grad_norm": 9.318495750427246,
834
+ "learning_rate": 0.0001987739439695514,
835
+ "loss": 0.3257,
836
+ "step": 1180
837
+ },
838
+ {
839
+ "epoch": 0.018562425906283146,
840
+ "grad_norm": 3.1046931743621826,
841
+ "learning_rate": 0.0001987635448513966,
842
+ "loss": 0.4267,
843
+ "step": 1190
844
+ },
845
+ {
846
+ "epoch": 0.018718412678604853,
847
+ "grad_norm": 4.199711799621582,
848
+ "learning_rate": 0.00019875314573324185,
849
+ "loss": 0.6676,
850
+ "step": 1200
851
+ },
852
+ {
853
+ "epoch": 0.01887439945092656,
854
+ "grad_norm": 12.733593940734863,
855
+ "learning_rate": 0.00019874274661508705,
856
+ "loss": 0.5991,
857
+ "step": 1210
858
+ },
859
+ {
860
+ "epoch": 0.019030386223248267,
861
+ "grad_norm": 5.1266984939575195,
862
+ "learning_rate": 0.0001987323474969323,
863
+ "loss": 0.2945,
864
+ "step": 1220
865
+ },
866
+ {
867
+ "epoch": 0.019186372995569977,
868
+ "grad_norm": 6.397432804107666,
869
+ "learning_rate": 0.00019872194837877748,
870
+ "loss": 0.4774,
871
+ "step": 1230
872
+ },
873
+ {
874
+ "epoch": 0.019342359767891684,
875
+ "grad_norm": 3.761129379272461,
876
+ "learning_rate": 0.0001987115492606227,
877
+ "loss": 0.4345,
878
+ "step": 1240
879
+ },
880
+ {
881
+ "epoch": 0.01949834654021339,
882
+ "grad_norm": 2.7355475425720215,
883
+ "learning_rate": 0.00019870115014246792,
884
+ "loss": 0.497,
885
+ "step": 1250
886
+ },
887
+ {
888
+ "epoch": 0.019654333312535098,
889
+ "grad_norm": 2.7893755435943604,
890
+ "learning_rate": 0.00019869075102431314,
891
+ "loss": 0.2169,
892
+ "step": 1260
893
+ },
894
+ {
895
+ "epoch": 0.019810320084856805,
896
+ "grad_norm": 10.722111701965332,
897
+ "learning_rate": 0.00019868035190615836,
898
+ "loss": 0.3867,
899
+ "step": 1270
900
+ },
901
+ {
902
+ "epoch": 0.019966306857178512,
903
+ "grad_norm": 11.647889137268066,
904
+ "learning_rate": 0.00019866995278800358,
905
+ "loss": 0.4138,
906
+ "step": 1280
907
+ },
908
+ {
909
+ "epoch": 0.02012229362950022,
910
+ "grad_norm": 1.4313205480575562,
911
+ "learning_rate": 0.0001986595536698488,
912
+ "loss": 0.3581,
913
+ "step": 1290
914
+ },
915
+ {
916
+ "epoch": 0.020278280401821926,
917
+ "grad_norm": 5.193609714508057,
918
+ "learning_rate": 0.00019864915455169402,
919
+ "loss": 0.5259,
920
+ "step": 1300
921
+ },
922
+ {
923
+ "epoch": 0.020434267174143633,
924
+ "grad_norm": 3.1939520835876465,
925
+ "learning_rate": 0.00019863875543353924,
926
+ "loss": 0.5391,
927
+ "step": 1310
928
+ },
929
+ {
930
+ "epoch": 0.02059025394646534,
931
+ "grad_norm": 7.716986179351807,
932
+ "learning_rate": 0.00019862835631538446,
933
+ "loss": 0.4773,
934
+ "step": 1320
935
+ },
936
+ {
937
+ "epoch": 0.020746240718787046,
938
+ "grad_norm": 3.191506862640381,
939
+ "learning_rate": 0.00019861795719722968,
940
+ "loss": 0.6051,
941
+ "step": 1330
942
+ },
943
+ {
944
+ "epoch": 0.020902227491108753,
945
+ "grad_norm": 5.67583703994751,
946
+ "learning_rate": 0.0001986075580790749,
947
+ "loss": 0.2716,
948
+ "step": 1340
949
+ },
950
+ {
951
+ "epoch": 0.02105821426343046,
952
+ "grad_norm": 3.1657073497772217,
953
+ "learning_rate": 0.00019859715896092012,
954
+ "loss": 0.4838,
955
+ "step": 1350
956
+ },
957
+ {
958
+ "epoch": 0.021214201035752167,
959
+ "grad_norm": 5.144644260406494,
960
+ "learning_rate": 0.00019858675984276534,
961
+ "loss": 0.3185,
962
+ "step": 1360
963
+ },
964
+ {
965
+ "epoch": 0.021370187808073874,
966
+ "grad_norm": 0.3853607177734375,
967
+ "learning_rate": 0.00019857636072461056,
968
+ "loss": 0.21,
969
+ "step": 1370
970
+ },
971
+ {
972
+ "epoch": 0.02152617458039558,
973
+ "grad_norm": 15.329313278198242,
974
+ "learning_rate": 0.00019856596160645578,
975
+ "loss": 0.4216,
976
+ "step": 1380
977
+ },
978
+ {
979
+ "epoch": 0.02168216135271729,
980
+ "grad_norm": 1.9162156581878662,
981
+ "learning_rate": 0.000198555562488301,
982
+ "loss": 0.3766,
983
+ "step": 1390
984
+ },
985
+ {
986
+ "epoch": 0.021838148125038998,
987
+ "grad_norm": 2.5800933837890625,
988
+ "learning_rate": 0.00019854516337014622,
989
+ "loss": 0.4478,
990
+ "step": 1400
991
+ },
992
+ {
993
+ "epoch": 0.021994134897360705,
994
+ "grad_norm": 12.898444175720215,
995
+ "learning_rate": 0.00019853476425199144,
996
+ "loss": 0.7923,
997
+ "step": 1410
998
+ },
999
+ {
1000
+ "epoch": 0.022150121669682412,
1001
+ "grad_norm": 11.576621055603027,
1002
+ "learning_rate": 0.00019852436513383666,
1003
+ "loss": 0.3582,
1004
+ "step": 1420
1005
+ },
1006
+ {
1007
+ "epoch": 0.02230610844200412,
1008
+ "grad_norm": 10.80225658416748,
1009
+ "learning_rate": 0.00019851396601568188,
1010
+ "loss": 0.4698,
1011
+ "step": 1430
1012
+ },
1013
+ {
1014
+ "epoch": 0.022462095214325826,
1015
+ "grad_norm": 5.019324779510498,
1016
+ "learning_rate": 0.0001985035668975271,
1017
+ "loss": 0.4463,
1018
+ "step": 1440
1019
+ },
1020
+ {
1021
+ "epoch": 0.022618081986647533,
1022
+ "grad_norm": 5.366154670715332,
1023
+ "learning_rate": 0.00019849316777937232,
1024
+ "loss": 0.4288,
1025
+ "step": 1450
1026
+ },
1027
+ {
1028
+ "epoch": 0.02277406875896924,
1029
+ "grad_norm": 3.945873498916626,
1030
+ "learning_rate": 0.00019848276866121754,
1031
+ "loss": 0.5625,
1032
+ "step": 1460
1033
+ },
1034
+ {
1035
+ "epoch": 0.022930055531290947,
1036
+ "grad_norm": 1.8938435316085815,
1037
+ "learning_rate": 0.00019847236954306276,
1038
+ "loss": 0.4298,
1039
+ "step": 1470
1040
+ },
1041
+ {
1042
+ "epoch": 0.023086042303612653,
1043
+ "grad_norm": 7.040036678314209,
1044
+ "learning_rate": 0.00019846197042490798,
1045
+ "loss": 0.5806,
1046
+ "step": 1480
1047
+ },
1048
+ {
1049
+ "epoch": 0.02324202907593436,
1050
+ "grad_norm": 12.009136199951172,
1051
+ "learning_rate": 0.0001984515713067532,
1052
+ "loss": 0.3603,
1053
+ "step": 1490
1054
+ },
1055
+ {
1056
+ "epoch": 0.023398015848256067,
1057
+ "grad_norm": 3.157111883163452,
1058
+ "learning_rate": 0.00019844117218859841,
1059
+ "loss": 0.3811,
1060
+ "step": 1500
1061
+ },
1062
+ {
1063
+ "epoch": 0.023554002620577774,
1064
+ "grad_norm": 1.0227491855621338,
1065
+ "learning_rate": 0.00019843077307044363,
1066
+ "loss": 0.3459,
1067
+ "step": 1510
1068
+ },
1069
+ {
1070
+ "epoch": 0.02370998939289948,
1071
+ "grad_norm": 2.699263334274292,
1072
+ "learning_rate": 0.00019842037395228885,
1073
+ "loss": 0.4047,
1074
+ "step": 1520
1075
+ },
1076
+ {
1077
+ "epoch": 0.023865976165221188,
1078
+ "grad_norm": 3.879258871078491,
1079
+ "learning_rate": 0.00019840997483413407,
1080
+ "loss": 0.2083,
1081
+ "step": 1530
1082
+ },
1083
+ {
1084
+ "epoch": 0.024021962937542895,
1085
+ "grad_norm": 1.0468460321426392,
1086
+ "learning_rate": 0.0001983995757159793,
1087
+ "loss": 0.3146,
1088
+ "step": 1540
1089
+ },
1090
+ {
1091
+ "epoch": 0.024177949709864602,
1092
+ "grad_norm": 3.5738143920898438,
1093
+ "learning_rate": 0.0001983891765978245,
1094
+ "loss": 0.3181,
1095
+ "step": 1550
1096
+ },
1097
+ {
1098
+ "epoch": 0.024333936482186312,
1099
+ "grad_norm": 6.3198113441467285,
1100
+ "learning_rate": 0.00019837877747966973,
1101
+ "loss": 0.3196,
1102
+ "step": 1560
1103
+ },
1104
+ {
1105
+ "epoch": 0.02448992325450802,
1106
+ "grad_norm": 1.6974947452545166,
1107
+ "learning_rate": 0.00019836837836151495,
1108
+ "loss": 0.3832,
1109
+ "step": 1570
1110
+ },
1111
+ {
1112
+ "epoch": 0.024645910026829726,
1113
+ "grad_norm": 6.220627307891846,
1114
+ "learning_rate": 0.00019835797924336017,
1115
+ "loss": 0.3445,
1116
+ "step": 1580
1117
+ },
1118
+ {
1119
+ "epoch": 0.024801896799151433,
1120
+ "grad_norm": 3.9202497005462646,
1121
+ "learning_rate": 0.0001983475801252054,
1122
+ "loss": 0.4304,
1123
+ "step": 1590
1124
+ },
1125
+ {
1126
+ "epoch": 0.02495788357147314,
1127
+ "grad_norm": 2.2945311069488525,
1128
+ "learning_rate": 0.0001983371810070506,
1129
+ "loss": 0.4487,
1130
+ "step": 1600
1131
+ },
1132
+ {
1133
+ "epoch": 0.025113870343794847,
1134
+ "grad_norm": 9.262955665588379,
1135
+ "learning_rate": 0.00019832678188889583,
1136
+ "loss": 0.3691,
1137
+ "step": 1610
1138
+ },
1139
+ {
1140
+ "epoch": 0.025269857116116554,
1141
+ "grad_norm": 2.6905221939086914,
1142
+ "learning_rate": 0.00019831638277074105,
1143
+ "loss": 0.5117,
1144
+ "step": 1620
1145
+ },
1146
+ {
1147
+ "epoch": 0.02542584388843826,
1148
+ "grad_norm": 4.1330060958862305,
1149
+ "learning_rate": 0.00019830598365258627,
1150
+ "loss": 0.5721,
1151
+ "step": 1630
1152
+ },
1153
+ {
1154
+ "epoch": 0.025581830660759967,
1155
+ "grad_norm": 1.0681806802749634,
1156
+ "learning_rate": 0.0001982955845344315,
1157
+ "loss": 0.4794,
1158
+ "step": 1640
1159
+ },
1160
+ {
1161
+ "epoch": 0.025737817433081674,
1162
+ "grad_norm": 2.6171841621398926,
1163
+ "learning_rate": 0.0001982851854162767,
1164
+ "loss": 0.4596,
1165
+ "step": 1650
1166
+ },
1167
+ {
1168
+ "epoch": 0.02589380420540338,
1169
+ "grad_norm": 3.3451762199401855,
1170
+ "learning_rate": 0.00019827478629812193,
1171
+ "loss": 0.4745,
1172
+ "step": 1660
1173
+ },
1174
+ {
1175
+ "epoch": 0.026049790977725088,
1176
+ "grad_norm": 4.152349472045898,
1177
+ "learning_rate": 0.00019826438717996715,
1178
+ "loss": 0.4452,
1179
+ "step": 1670
1180
+ },
1181
+ {
1182
+ "epoch": 0.026205777750046795,
1183
+ "grad_norm": 2.7109954357147217,
1184
+ "learning_rate": 0.00019825398806181237,
1185
+ "loss": 0.3966,
1186
+ "step": 1680
1187
+ },
1188
+ {
1189
+ "epoch": 0.026361764522368502,
1190
+ "grad_norm": 5.731235980987549,
1191
+ "learning_rate": 0.0001982435889436576,
1192
+ "loss": 0.4423,
1193
+ "step": 1690
1194
+ },
1195
+ {
1196
+ "epoch": 0.02651775129469021,
1197
+ "grad_norm": 2.370852470397949,
1198
+ "learning_rate": 0.0001982331898255028,
1199
+ "loss": 0.5619,
1200
+ "step": 1700
1201
+ },
1202
+ {
1203
+ "epoch": 0.026673738067011916,
1204
+ "grad_norm": 2.5500876903533936,
1205
+ "learning_rate": 0.00019822279070734803,
1206
+ "loss": 0.2658,
1207
+ "step": 1710
1208
+ },
1209
+ {
1210
+ "epoch": 0.026829724839333626,
1211
+ "grad_norm": 0.20157204568386078,
1212
+ "learning_rate": 0.00019821239158919325,
1213
+ "loss": 0.2939,
1214
+ "step": 1720
1215
+ },
1216
+ {
1217
+ "epoch": 0.026985711611655333,
1218
+ "grad_norm": 2.6305084228515625,
1219
+ "learning_rate": 0.00019820199247103847,
1220
+ "loss": 0.5339,
1221
+ "step": 1730
1222
+ },
1223
+ {
1224
+ "epoch": 0.02714169838397704,
1225
+ "grad_norm": 6.241075038909912,
1226
+ "learning_rate": 0.00019819159335288369,
1227
+ "loss": 0.3625,
1228
+ "step": 1740
1229
+ },
1230
+ {
1231
+ "epoch": 0.027297685156298747,
1232
+ "grad_norm": 5.584851264953613,
1233
+ "learning_rate": 0.0001981811942347289,
1234
+ "loss": 0.6158,
1235
+ "step": 1750
1236
+ },
1237
+ {
1238
+ "epoch": 0.027453671928620454,
1239
+ "grad_norm": 6.756530284881592,
1240
+ "learning_rate": 0.00019817079511657413,
1241
+ "loss": 0.3513,
1242
+ "step": 1760
1243
+ },
1244
+ {
1245
+ "epoch": 0.02760965870094216,
1246
+ "grad_norm": 5.153111457824707,
1247
+ "learning_rate": 0.00019816039599841935,
1248
+ "loss": 0.2885,
1249
+ "step": 1770
1250
+ },
1251
+ {
1252
+ "epoch": 0.027765645473263868,
1253
+ "grad_norm": 6.222873210906982,
1254
+ "learning_rate": 0.00019814999688026456,
1255
+ "loss": 0.3291,
1256
+ "step": 1780
1257
+ },
1258
+ {
1259
+ "epoch": 0.027921632245585574,
1260
+ "grad_norm": 4.744391918182373,
1261
+ "learning_rate": 0.00019813959776210978,
1262
+ "loss": 0.5053,
1263
+ "step": 1790
1264
+ },
1265
+ {
1266
+ "epoch": 0.02807761901790728,
1267
+ "grad_norm": 3.7181131839752197,
1268
+ "learning_rate": 0.000198129198643955,
1269
+ "loss": 0.3983,
1270
+ "step": 1800
1271
+ },
1272
+ {
1273
+ "epoch": 0.02823360579022899,
1274
+ "grad_norm": 1.8923834562301636,
1275
+ "learning_rate": 0.00019811879952580022,
1276
+ "loss": 0.4465,
1277
+ "step": 1810
1278
+ },
1279
+ {
1280
+ "epoch": 0.028389592562550695,
1281
+ "grad_norm": 6.049956321716309,
1282
+ "learning_rate": 0.00019810840040764544,
1283
+ "loss": 0.5028,
1284
+ "step": 1820
1285
+ },
1286
+ {
1287
+ "epoch": 0.028545579334872402,
1288
+ "grad_norm": 8.034428596496582,
1289
+ "learning_rate": 0.00019809800128949066,
1290
+ "loss": 0.4799,
1291
+ "step": 1830
1292
+ },
1293
+ {
1294
+ "epoch": 0.02870156610719411,
1295
+ "grad_norm": 3.916890859603882,
1296
+ "learning_rate": 0.00019808760217133588,
1297
+ "loss": 0.3109,
1298
+ "step": 1840
1299
+ },
1300
+ {
1301
+ "epoch": 0.028857552879515816,
1302
+ "grad_norm": 3.351285457611084,
1303
+ "learning_rate": 0.0001980772030531811,
1304
+ "loss": 0.5981,
1305
+ "step": 1850
1306
+ },
1307
+ {
1308
+ "epoch": 0.029013539651837523,
1309
+ "grad_norm": 4.443714141845703,
1310
+ "learning_rate": 0.00019806680393502632,
1311
+ "loss": 0.3243,
1312
+ "step": 1860
1313
+ },
1314
+ {
1315
+ "epoch": 0.02916952642415923,
1316
+ "grad_norm": 0.6998550295829773,
1317
+ "learning_rate": 0.00019805640481687154,
1318
+ "loss": 0.2057,
1319
+ "step": 1870
1320
+ },
1321
+ {
1322
+ "epoch": 0.02932551319648094,
1323
+ "grad_norm": 12.252562522888184,
1324
+ "learning_rate": 0.00019804600569871676,
1325
+ "loss": 0.3806,
1326
+ "step": 1880
1327
+ },
1328
+ {
1329
+ "epoch": 0.029481499968802647,
1330
+ "grad_norm": 9.03868293762207,
1331
+ "learning_rate": 0.00019803560658056198,
1332
+ "loss": 0.3351,
1333
+ "step": 1890
1334
+ },
1335
+ {
1336
+ "epoch": 0.029637486741124354,
1337
+ "grad_norm": 2.3400044441223145,
1338
+ "learning_rate": 0.0001980252074624072,
1339
+ "loss": 0.3103,
1340
+ "step": 1900
1341
+ },
1342
+ {
1343
+ "epoch": 0.02979347351344606,
1344
+ "grad_norm": 2.399489164352417,
1345
+ "learning_rate": 0.00019801480834425242,
1346
+ "loss": 0.4792,
1347
+ "step": 1910
1348
+ },
1349
+ {
1350
+ "epoch": 0.029949460285767768,
1351
+ "grad_norm": 1.5152381658554077,
1352
+ "learning_rate": 0.00019800440922609764,
1353
+ "loss": 0.2873,
1354
+ "step": 1920
1355
+ },
1356
+ {
1357
+ "epoch": 0.030105447058089475,
1358
+ "grad_norm": 2.7013959884643555,
1359
+ "learning_rate": 0.00019799401010794286,
1360
+ "loss": 0.381,
1361
+ "step": 1930
1362
+ },
1363
+ {
1364
+ "epoch": 0.03026143383041118,
1365
+ "grad_norm": 0.7135261297225952,
1366
+ "learning_rate": 0.00019798361098978808,
1367
+ "loss": 0.4001,
1368
+ "step": 1940
1369
+ },
1370
+ {
1371
+ "epoch": 0.03041742060273289,
1372
+ "grad_norm": 15.61109447479248,
1373
+ "learning_rate": 0.0001979732118716333,
1374
+ "loss": 0.474,
1375
+ "step": 1950
1376
+ },
1377
+ {
1378
+ "epoch": 0.030573407375054595,
1379
+ "grad_norm": 3.236668348312378,
1380
+ "learning_rate": 0.00019796281275347852,
1381
+ "loss": 0.2149,
1382
+ "step": 1960
1383
+ },
1384
+ {
1385
+ "epoch": 0.030729394147376302,
1386
+ "grad_norm": 4.722929954528809,
1387
+ "learning_rate": 0.00019795241363532374,
1388
+ "loss": 0.5474,
1389
+ "step": 1970
1390
+ },
1391
+ {
1392
+ "epoch": 0.03088538091969801,
1393
+ "grad_norm": 6.002335548400879,
1394
+ "learning_rate": 0.00019794201451716896,
1395
+ "loss": 0.5196,
1396
+ "step": 1980
1397
+ },
1398
+ {
1399
+ "epoch": 0.031041367692019716,
1400
+ "grad_norm": 5.210022926330566,
1401
+ "learning_rate": 0.00019793161539901418,
1402
+ "loss": 0.4741,
1403
+ "step": 1990
1404
+ },
1405
+ {
1406
+ "epoch": 0.031197354464341423,
1407
+ "grad_norm": 7.984034538269043,
1408
+ "learning_rate": 0.0001979212162808594,
1409
+ "loss": 0.5399,
1410
+ "step": 2000
1411
+ }
1412
+ ],
1413
+ "logging_steps": 10,
1414
+ "max_steps": 192324,
1415
+ "num_input_tokens_seen": 0,
1416
+ "num_train_epochs": 3,
1417
+ "save_steps": 500,
1418
+ "stateful_callbacks": {
1419
+ "TrainerControl": {
1420
+ "args": {
1421
+ "should_epoch_stop": false,
1422
+ "should_evaluate": false,
1423
+ "should_log": false,
1424
+ "should_save": true,
1425
+ "should_training_stop": false
1426
+ },
1427
+ "attributes": {}
1428
+ }
1429
+ },
1430
+ "total_flos": 1217915781120000.0,
1431
+ "train_batch_size": 1,
1432
+ "trial_name": null,
1433
+ "trial_params": null
1434
+ }
HVU_QA/t5-viet-qg-finetuned/checkpoint-2000/training_args.bin ADDED
Binary file (5.71 kB). View file
 
HVU_QA/t5-viet-qg-finetuned/checkpoint-2500/added_tokens.json ADDED
@@ -0,0 +1,98 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "<extra_id_0>": 36095,
3
+ "<extra_id_10>": 36085,
4
+ "<extra_id_11>": 36084,
5
+ "<extra_id_12>": 36083,
6
+ "<extra_id_13>": 36082,
7
+ "<extra_id_14>": 36081,
8
+ "<extra_id_15>": 36080,
9
+ "<extra_id_16>": 36079,
10
+ "<extra_id_17>": 36078,
11
+ "<extra_id_18>": 36077,
12
+ "<extra_id_19>": 36076,
13
+ "<extra_id_1>": 36094,
14
+ "<extra_id_20>": 36075,
15
+ "<extra_id_21>": 36074,
16
+ "<extra_id_22>": 36073,
17
+ "<extra_id_23>": 36072,
18
+ "<extra_id_24>": 36071,
19
+ "<extra_id_25>": 36070,
20
+ "<extra_id_26>": 36069,
21
+ "<extra_id_27>": 36068,
22
+ "<extra_id_28>": 36067,
23
+ "<extra_id_29>": 36066,
24
+ "<extra_id_2>": 36093,
25
+ "<extra_id_30>": 36065,
26
+ "<extra_id_31>": 36064,
27
+ "<extra_id_32>": 36063,
28
+ "<extra_id_33>": 36062,
29
+ "<extra_id_34>": 36061,
30
+ "<extra_id_35>": 36060,
31
+ "<extra_id_36>": 36059,
32
+ "<extra_id_37>": 36058,
33
+ "<extra_id_38>": 36057,
34
+ "<extra_id_39>": 36056,
35
+ "<extra_id_3>": 36092,
36
+ "<extra_id_40>": 36055,
37
+ "<extra_id_41>": 36054,
38
+ "<extra_id_42>": 36053,
39
+ "<extra_id_43>": 36052,
40
+ "<extra_id_44>": 36051,
41
+ "<extra_id_45>": 36050,
42
+ "<extra_id_46>": 36049,
43
+ "<extra_id_47>": 36048,
44
+ "<extra_id_48>": 36047,
45
+ "<extra_id_49>": 36046,
46
+ "<extra_id_4>": 36091,
47
+ "<extra_id_50>": 36045,
48
+ "<extra_id_51>": 36044,
49
+ "<extra_id_52>": 36043,
50
+ "<extra_id_53>": 36042,
51
+ "<extra_id_54>": 36041,
52
+ "<extra_id_55>": 36040,
53
+ "<extra_id_56>": 36039,
54
+ "<extra_id_57>": 36038,
55
+ "<extra_id_58>": 36037,
56
+ "<extra_id_59>": 36036,
57
+ "<extra_id_5>": 36090,
58
+ "<extra_id_60>": 36035,
59
+ "<extra_id_61>": 36034,
60
+ "<extra_id_62>": 36033,
61
+ "<extra_id_63>": 36032,
62
+ "<extra_id_64>": 36031,
63
+ "<extra_id_65>": 36030,
64
+ "<extra_id_66>": 36029,
65
+ "<extra_id_67>": 36028,
66
+ "<extra_id_68>": 36027,
67
+ "<extra_id_69>": 36026,
68
+ "<extra_id_6>": 36089,
69
+ "<extra_id_70>": 36025,
70
+ "<extra_id_71>": 36024,
71
+ "<extra_id_72>": 36023,
72
+ "<extra_id_73>": 36022,
73
+ "<extra_id_74>": 36021,
74
+ "<extra_id_75>": 36020,
75
+ "<extra_id_76>": 36019,
76
+ "<extra_id_77>": 36018,
77
+ "<extra_id_78>": 36017,
78
+ "<extra_id_79>": 36016,
79
+ "<extra_id_7>": 36088,
80
+ "<extra_id_80>": 36015,
81
+ "<extra_id_81>": 36014,
82
+ "<extra_id_82>": 36013,
83
+ "<extra_id_83>": 36012,
84
+ "<extra_id_84>": 36011,
85
+ "<extra_id_85>": 36010,
86
+ "<extra_id_86>": 36009,
87
+ "<extra_id_87>": 36008,
88
+ "<extra_id_88>": 36007,
89
+ "<extra_id_89>": 36006,
90
+ "<extra_id_8>": 36087,
91
+ "<extra_id_90>": 36005,
92
+ "<extra_id_91>": 36004,
93
+ "<extra_id_92>": 36003,
94
+ "<extra_id_93>": 36002,
95
+ "<extra_id_94>": 36001,
96
+ "<extra_id_95>": 36000,
97
+ "<extra_id_9>": 36086
98
+ }
HVU_QA/t5-viet-qg-finetuned/checkpoint-2500/config.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "T5ForConditionalGeneration"
4
+ ],
5
+ "classifier_dropout": 0.0,
6
+ "d_ff": 3072,
7
+ "d_kv": 64,
8
+ "d_model": 768,
9
+ "decoder_start_token_id": 0,
10
+ "dense_act_fn": "relu",
11
+ "dropout_rate": 0.1,
12
+ "eos_token_id": 1,
13
+ "feed_forward_proj": "relu",
14
+ "initializer_factor": 1.0,
15
+ "is_encoder_decoder": true,
16
+ "is_gated_act": false,
17
+ "layer_norm_epsilon": 1e-06,
18
+ "model_type": "t5",
19
+ "n_positions": 512,
20
+ "num_decoder_layers": 12,
21
+ "num_heads": 12,
22
+ "num_layers": 12,
23
+ "output_past": true,
24
+ "pad_token_id": 0,
25
+ "relative_attention_max_distance": 128,
26
+ "relative_attention_num_buckets": 32,
27
+ "torch_dtype": "float32",
28
+ "transformers_version": "4.53.2",
29
+ "use_cache": true,
30
+ "vocab_size": 36096
31
+ }
HVU_QA/t5-viet-qg-finetuned/checkpoint-2500/generation_config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "decoder_start_token_id": 0,
4
+ "eos_token_id": 1,
5
+ "pad_token_id": 0,
6
+ "transformers_version": "4.53.2"
7
+ }
HVU_QA/t5-viet-qg-finetuned/checkpoint-2500/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1d263ea5f8f3a6ff45757ab499322cdd30a33b351bd40930eec7c6de19421bf1
3
+ size 62652416
HVU_QA/t5-viet-qg-finetuned/checkpoint-2500/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:05c95757cde0e203ab6b291e2f9940666f33ce2fd0b48da06a899026775f3953
3
+ size 64487424
HVU_QA/t5-viet-qg-finetuned/checkpoint-2500/rng_state.pth ADDED
Binary file (14.6 kB). View file
 
HVU_QA/t5-viet-qg-finetuned/checkpoint-2500/scheduler.pt ADDED
Binary file (1.47 kB). View file
 
HVU_QA/t5-viet-qg-finetuned/checkpoint-2500/special_tokens_map.json ADDED
@@ -0,0 +1,121 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<extra_id_0>",
4
+ "<extra_id_1>",
5
+ "<extra_id_2>",
6
+ "<extra_id_3>",
7
+ "<extra_id_4>",
8
+ "<extra_id_5>",
9
+ "<extra_id_6>",
10
+ "<extra_id_7>",
11
+ "<extra_id_8>",
12
+ "<extra_id_9>",
13
+ "<extra_id_10>",
14
+ "<extra_id_11>",
15
+ "<extra_id_12>",
16
+ "<extra_id_13>",
17
+ "<extra_id_14>",
18
+ "<extra_id_15>",
19
+ "<extra_id_16>",
20
+ "<extra_id_17>",
21
+ "<extra_id_18>",
22
+ "<extra_id_19>",
23
+ "<extra_id_20>",
24
+ "<extra_id_21>",
25
+ "<extra_id_22>",
26
+ "<extra_id_23>",
27
+ "<extra_id_24>",
28
+ "<extra_id_25>",
29
+ "<extra_id_26>",
30
+ "<extra_id_27>",
31
+ "<extra_id_28>",
32
+ "<extra_id_29>",
33
+ "<extra_id_30>",
34
+ "<extra_id_31>",
35
+ "<extra_id_32>",
36
+ "<extra_id_33>",
37
+ "<extra_id_34>",
38
+ "<extra_id_35>",
39
+ "<extra_id_36>",
40
+ "<extra_id_37>",
41
+ "<extra_id_38>",
42
+ "<extra_id_39>",
43
+ "<extra_id_40>",
44
+ "<extra_id_41>",
45
+ "<extra_id_42>",
46
+ "<extra_id_43>",
47
+ "<extra_id_44>",
48
+ "<extra_id_45>",
49
+ "<extra_id_46>",
50
+ "<extra_id_47>",
51
+ "<extra_id_48>",
52
+ "<extra_id_49>",
53
+ "<extra_id_50>",
54
+ "<extra_id_51>",
55
+ "<extra_id_52>",
56
+ "<extra_id_53>",
57
+ "<extra_id_54>",
58
+ "<extra_id_55>",
59
+ "<extra_id_56>",
60
+ "<extra_id_57>",
61
+ "<extra_id_58>",
62
+ "<extra_id_59>",
63
+ "<extra_id_60>",
64
+ "<extra_id_61>",
65
+ "<extra_id_62>",
66
+ "<extra_id_63>",
67
+ "<extra_id_64>",
68
+ "<extra_id_65>",
69
+ "<extra_id_66>",
70
+ "<extra_id_67>",
71
+ "<extra_id_68>",
72
+ "<extra_id_69>",
73
+ "<extra_id_70>",
74
+ "<extra_id_71>",
75
+ "<extra_id_72>",
76
+ "<extra_id_73>",
77
+ "<extra_id_74>",
78
+ "<extra_id_75>",
79
+ "<extra_id_76>",
80
+ "<extra_id_77>",
81
+ "<extra_id_78>",
82
+ "<extra_id_79>",
83
+ "<extra_id_80>",
84
+ "<extra_id_81>",
85
+ "<extra_id_82>",
86
+ "<extra_id_83>",
87
+ "<extra_id_84>",
88
+ "<extra_id_85>",
89
+ "<extra_id_86>",
90
+ "<extra_id_87>",
91
+ "<extra_id_88>",
92
+ "<extra_id_89>",
93
+ "<extra_id_90>",
94
+ "<extra_id_91>",
95
+ "<extra_id_92>",
96
+ "<extra_id_93>",
97
+ "<extra_id_94>",
98
+ "<extra_id_95>"
99
+ ],
100
+ "eos_token": {
101
+ "content": "</s>",
102
+ "lstrip": false,
103
+ "normalized": false,
104
+ "rstrip": false,
105
+ "single_word": false
106
+ },
107
+ "pad_token": {
108
+ "content": "<pad>",
109
+ "lstrip": false,
110
+ "normalized": false,
111
+ "rstrip": false,
112
+ "single_word": false
113
+ },
114
+ "unk_token": {
115
+ "content": "<unk>",
116
+ "lstrip": false,
117
+ "normalized": false,
118
+ "rstrip": false,
119
+ "single_word": false
120
+ }
121
+ }
HVU_QA/t5-viet-qg-finetuned/checkpoint-2500/spiece.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6
3
+ size 820370
HVU_QA/t5-viet-qg-finetuned/checkpoint-2500/tokenizer_config.json ADDED
@@ -0,0 +1,905 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": true,
3
+ "added_tokens_decoder": {
4
+ "0": {
5
+ "content": "<pad>",
6
+ "lstrip": false,
7
+ "normalized": false,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "1": {
13
+ "content": "</s>",
14
+ "lstrip": false,
15
+ "normalized": false,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
+ },
20
+ "2": {
21
+ "content": "<unk>",
22
+ "lstrip": false,
23
+ "normalized": false,
24
+ "rstrip": false,
25
+ "single_word": false,
26
+ "special": true
27
+ },
28
+ "36000": {
29
+ "content": "<extra_id_95>",
30
+ "lstrip": true,
31
+ "normalized": false,
32
+ "rstrip": true,
33
+ "single_word": false,
34
+ "special": true
35
+ },
36
+ "36001": {
37
+ "content": "<extra_id_94>",
38
+ "lstrip": true,
39
+ "normalized": false,
40
+ "rstrip": true,
41
+ "single_word": false,
42
+ "special": true
43
+ },
44
+ "36002": {
45
+ "content": "<extra_id_93>",
46
+ "lstrip": true,
47
+ "normalized": false,
48
+ "rstrip": true,
49
+ "single_word": false,
50
+ "special": true
51
+ },
52
+ "36003": {
53
+ "content": "<extra_id_92>",
54
+ "lstrip": true,
55
+ "normalized": false,
56
+ "rstrip": true,
57
+ "single_word": false,
58
+ "special": true
59
+ },
60
+ "36004": {
61
+ "content": "<extra_id_91>",
62
+ "lstrip": true,
63
+ "normalized": false,
64
+ "rstrip": true,
65
+ "single_word": false,
66
+ "special": true
67
+ },
68
+ "36005": {
69
+ "content": "<extra_id_90>",
70
+ "lstrip": true,
71
+ "normalized": false,
72
+ "rstrip": true,
73
+ "single_word": false,
74
+ "special": true
75
+ },
76
+ "36006": {
77
+ "content": "<extra_id_89>",
78
+ "lstrip": true,
79
+ "normalized": false,
80
+ "rstrip": true,
81
+ "single_word": false,
82
+ "special": true
83
+ },
84
+ "36007": {
85
+ "content": "<extra_id_88>",
86
+ "lstrip": true,
87
+ "normalized": false,
88
+ "rstrip": true,
89
+ "single_word": false,
90
+ "special": true
91
+ },
92
+ "36008": {
93
+ "content": "<extra_id_87>",
94
+ "lstrip": true,
95
+ "normalized": false,
96
+ "rstrip": true,
97
+ "single_word": false,
98
+ "special": true
99
+ },
100
+ "36009": {
101
+ "content": "<extra_id_86>",
102
+ "lstrip": true,
103
+ "normalized": false,
104
+ "rstrip": true,
105
+ "single_word": false,
106
+ "special": true
107
+ },
108
+ "36010": {
109
+ "content": "<extra_id_85>",
110
+ "lstrip": true,
111
+ "normalized": false,
112
+ "rstrip": true,
113
+ "single_word": false,
114
+ "special": true
115
+ },
116
+ "36011": {
117
+ "content": "<extra_id_84>",
118
+ "lstrip": true,
119
+ "normalized": false,
120
+ "rstrip": true,
121
+ "single_word": false,
122
+ "special": true
123
+ },
124
+ "36012": {
125
+ "content": "<extra_id_83>",
126
+ "lstrip": true,
127
+ "normalized": false,
128
+ "rstrip": true,
129
+ "single_word": false,
130
+ "special": true
131
+ },
132
+ "36013": {
133
+ "content": "<extra_id_82>",
134
+ "lstrip": true,
135
+ "normalized": false,
136
+ "rstrip": true,
137
+ "single_word": false,
138
+ "special": true
139
+ },
140
+ "36014": {
141
+ "content": "<extra_id_81>",
142
+ "lstrip": true,
143
+ "normalized": false,
144
+ "rstrip": true,
145
+ "single_word": false,
146
+ "special": true
147
+ },
148
+ "36015": {
149
+ "content": "<extra_id_80>",
150
+ "lstrip": true,
151
+ "normalized": false,
152
+ "rstrip": true,
153
+ "single_word": false,
154
+ "special": true
155
+ },
156
+ "36016": {
157
+ "content": "<extra_id_79>",
158
+ "lstrip": true,
159
+ "normalized": false,
160
+ "rstrip": true,
161
+ "single_word": false,
162
+ "special": true
163
+ },
164
+ "36017": {
165
+ "content": "<extra_id_78>",
166
+ "lstrip": true,
167
+ "normalized": false,
168
+ "rstrip": true,
169
+ "single_word": false,
170
+ "special": true
171
+ },
172
+ "36018": {
173
+ "content": "<extra_id_77>",
174
+ "lstrip": true,
175
+ "normalized": false,
176
+ "rstrip": true,
177
+ "single_word": false,
178
+ "special": true
179
+ },
180
+ "36019": {
181
+ "content": "<extra_id_76>",
182
+ "lstrip": true,
183
+ "normalized": false,
184
+ "rstrip": true,
185
+ "single_word": false,
186
+ "special": true
187
+ },
188
+ "36020": {
189
+ "content": "<extra_id_75>",
190
+ "lstrip": true,
191
+ "normalized": false,
192
+ "rstrip": true,
193
+ "single_word": false,
194
+ "special": true
195
+ },
196
+ "36021": {
197
+ "content": "<extra_id_74>",
198
+ "lstrip": true,
199
+ "normalized": false,
200
+ "rstrip": true,
201
+ "single_word": false,
202
+ "special": true
203
+ },
204
+ "36022": {
205
+ "content": "<extra_id_73>",
206
+ "lstrip": true,
207
+ "normalized": false,
208
+ "rstrip": true,
209
+ "single_word": false,
210
+ "special": true
211
+ },
212
+ "36023": {
213
+ "content": "<extra_id_72>",
214
+ "lstrip": true,
215
+ "normalized": false,
216
+ "rstrip": true,
217
+ "single_word": false,
218
+ "special": true
219
+ },
220
+ "36024": {
221
+ "content": "<extra_id_71>",
222
+ "lstrip": true,
223
+ "normalized": false,
224
+ "rstrip": true,
225
+ "single_word": false,
226
+ "special": true
227
+ },
228
+ "36025": {
229
+ "content": "<extra_id_70>",
230
+ "lstrip": true,
231
+ "normalized": false,
232
+ "rstrip": true,
233
+ "single_word": false,
234
+ "special": true
235
+ },
236
+ "36026": {
237
+ "content": "<extra_id_69>",
238
+ "lstrip": true,
239
+ "normalized": false,
240
+ "rstrip": true,
241
+ "single_word": false,
242
+ "special": true
243
+ },
244
+ "36027": {
245
+ "content": "<extra_id_68>",
246
+ "lstrip": true,
247
+ "normalized": false,
248
+ "rstrip": true,
249
+ "single_word": false,
250
+ "special": true
251
+ },
252
+ "36028": {
253
+ "content": "<extra_id_67>",
254
+ "lstrip": true,
255
+ "normalized": false,
256
+ "rstrip": true,
257
+ "single_word": false,
258
+ "special": true
259
+ },
260
+ "36029": {
261
+ "content": "<extra_id_66>",
262
+ "lstrip": true,
263
+ "normalized": false,
264
+ "rstrip": true,
265
+ "single_word": false,
266
+ "special": true
267
+ },
268
+ "36030": {
269
+ "content": "<extra_id_65>",
270
+ "lstrip": true,
271
+ "normalized": false,
272
+ "rstrip": true,
273
+ "single_word": false,
274
+ "special": true
275
+ },
276
+ "36031": {
277
+ "content": "<extra_id_64>",
278
+ "lstrip": true,
279
+ "normalized": false,
280
+ "rstrip": true,
281
+ "single_word": false,
282
+ "special": true
283
+ },
284
+ "36032": {
285
+ "content": "<extra_id_63>",
286
+ "lstrip": true,
287
+ "normalized": false,
288
+ "rstrip": true,
289
+ "single_word": false,
290
+ "special": true
291
+ },
292
+ "36033": {
293
+ "content": "<extra_id_62>",
294
+ "lstrip": true,
295
+ "normalized": false,
296
+ "rstrip": true,
297
+ "single_word": false,
298
+ "special": true
299
+ },
300
+ "36034": {
301
+ "content": "<extra_id_61>",
302
+ "lstrip": true,
303
+ "normalized": false,
304
+ "rstrip": true,
305
+ "single_word": false,
306
+ "special": true
307
+ },
308
+ "36035": {
309
+ "content": "<extra_id_60>",
310
+ "lstrip": true,
311
+ "normalized": false,
312
+ "rstrip": true,
313
+ "single_word": false,
314
+ "special": true
315
+ },
316
+ "36036": {
317
+ "content": "<extra_id_59>",
318
+ "lstrip": true,
319
+ "normalized": false,
320
+ "rstrip": true,
321
+ "single_word": false,
322
+ "special": true
323
+ },
324
+ "36037": {
325
+ "content": "<extra_id_58>",
326
+ "lstrip": true,
327
+ "normalized": false,
328
+ "rstrip": true,
329
+ "single_word": false,
330
+ "special": true
331
+ },
332
+ "36038": {
333
+ "content": "<extra_id_57>",
334
+ "lstrip": true,
335
+ "normalized": false,
336
+ "rstrip": true,
337
+ "single_word": false,
338
+ "special": true
339
+ },
340
+ "36039": {
341
+ "content": "<extra_id_56>",
342
+ "lstrip": true,
343
+ "normalized": false,
344
+ "rstrip": true,
345
+ "single_word": false,
346
+ "special": true
347
+ },
348
+ "36040": {
349
+ "content": "<extra_id_55>",
350
+ "lstrip": true,
351
+ "normalized": false,
352
+ "rstrip": true,
353
+ "single_word": false,
354
+ "special": true
355
+ },
356
+ "36041": {
357
+ "content": "<extra_id_54>",
358
+ "lstrip": true,
359
+ "normalized": false,
360
+ "rstrip": true,
361
+ "single_word": false,
362
+ "special": true
363
+ },
364
+ "36042": {
365
+ "content": "<extra_id_53>",
366
+ "lstrip": true,
367
+ "normalized": false,
368
+ "rstrip": true,
369
+ "single_word": false,
370
+ "special": true
371
+ },
372
+ "36043": {
373
+ "content": "<extra_id_52>",
374
+ "lstrip": true,
375
+ "normalized": false,
376
+ "rstrip": true,
377
+ "single_word": false,
378
+ "special": true
379
+ },
380
+ "36044": {
381
+ "content": "<extra_id_51>",
382
+ "lstrip": true,
383
+ "normalized": false,
384
+ "rstrip": true,
385
+ "single_word": false,
386
+ "special": true
387
+ },
388
+ "36045": {
389
+ "content": "<extra_id_50>",
390
+ "lstrip": true,
391
+ "normalized": false,
392
+ "rstrip": true,
393
+ "single_word": false,
394
+ "special": true
395
+ },
396
+ "36046": {
397
+ "content": "<extra_id_49>",
398
+ "lstrip": true,
399
+ "normalized": false,
400
+ "rstrip": true,
401
+ "single_word": false,
402
+ "special": true
403
+ },
404
+ "36047": {
405
+ "content": "<extra_id_48>",
406
+ "lstrip": true,
407
+ "normalized": false,
408
+ "rstrip": true,
409
+ "single_word": false,
410
+ "special": true
411
+ },
412
+ "36048": {
413
+ "content": "<extra_id_47>",
414
+ "lstrip": true,
415
+ "normalized": false,
416
+ "rstrip": true,
417
+ "single_word": false,
418
+ "special": true
419
+ },
420
+ "36049": {
421
+ "content": "<extra_id_46>",
422
+ "lstrip": true,
423
+ "normalized": false,
424
+ "rstrip": true,
425
+ "single_word": false,
426
+ "special": true
427
+ },
428
+ "36050": {
429
+ "content": "<extra_id_45>",
430
+ "lstrip": true,
431
+ "normalized": false,
432
+ "rstrip": true,
433
+ "single_word": false,
434
+ "special": true
435
+ },
436
+ "36051": {
437
+ "content": "<extra_id_44>",
438
+ "lstrip": true,
439
+ "normalized": false,
440
+ "rstrip": true,
441
+ "single_word": false,
442
+ "special": true
443
+ },
444
+ "36052": {
445
+ "content": "<extra_id_43>",
446
+ "lstrip": true,
447
+ "normalized": false,
448
+ "rstrip": true,
449
+ "single_word": false,
450
+ "special": true
451
+ },
452
+ "36053": {
453
+ "content": "<extra_id_42>",
454
+ "lstrip": true,
455
+ "normalized": false,
456
+ "rstrip": true,
457
+ "single_word": false,
458
+ "special": true
459
+ },
460
+ "36054": {
461
+ "content": "<extra_id_41>",
462
+ "lstrip": true,
463
+ "normalized": false,
464
+ "rstrip": true,
465
+ "single_word": false,
466
+ "special": true
467
+ },
468
+ "36055": {
469
+ "content": "<extra_id_40>",
470
+ "lstrip": true,
471
+ "normalized": false,
472
+ "rstrip": true,
473
+ "single_word": false,
474
+ "special": true
475
+ },
476
+ "36056": {
477
+ "content": "<extra_id_39>",
478
+ "lstrip": true,
479
+ "normalized": false,
480
+ "rstrip": true,
481
+ "single_word": false,
482
+ "special": true
483
+ },
484
+ "36057": {
485
+ "content": "<extra_id_38>",
486
+ "lstrip": true,
487
+ "normalized": false,
488
+ "rstrip": true,
489
+ "single_word": false,
490
+ "special": true
491
+ },
492
+ "36058": {
493
+ "content": "<extra_id_37>",
494
+ "lstrip": true,
495
+ "normalized": false,
496
+ "rstrip": true,
497
+ "single_word": false,
498
+ "special": true
499
+ },
500
+ "36059": {
501
+ "content": "<extra_id_36>",
502
+ "lstrip": true,
503
+ "normalized": false,
504
+ "rstrip": true,
505
+ "single_word": false,
506
+ "special": true
507
+ },
508
+ "36060": {
509
+ "content": "<extra_id_35>",
510
+ "lstrip": true,
511
+ "normalized": false,
512
+ "rstrip": true,
513
+ "single_word": false,
514
+ "special": true
515
+ },
516
+ "36061": {
517
+ "content": "<extra_id_34>",
518
+ "lstrip": true,
519
+ "normalized": false,
520
+ "rstrip": true,
521
+ "single_word": false,
522
+ "special": true
523
+ },
524
+ "36062": {
525
+ "content": "<extra_id_33>",
526
+ "lstrip": true,
527
+ "normalized": false,
528
+ "rstrip": true,
529
+ "single_word": false,
530
+ "special": true
531
+ },
532
+ "36063": {
533
+ "content": "<extra_id_32>",
534
+ "lstrip": true,
535
+ "normalized": false,
536
+ "rstrip": true,
537
+ "single_word": false,
538
+ "special": true
539
+ },
540
+ "36064": {
541
+ "content": "<extra_id_31>",
542
+ "lstrip": true,
543
+ "normalized": false,
544
+ "rstrip": true,
545
+ "single_word": false,
546
+ "special": true
547
+ },
548
+ "36065": {
549
+ "content": "<extra_id_30>",
550
+ "lstrip": true,
551
+ "normalized": false,
552
+ "rstrip": true,
553
+ "single_word": false,
554
+ "special": true
555
+ },
556
+ "36066": {
557
+ "content": "<extra_id_29>",
558
+ "lstrip": true,
559
+ "normalized": false,
560
+ "rstrip": true,
561
+ "single_word": false,
562
+ "special": true
563
+ },
564
+ "36067": {
565
+ "content": "<extra_id_28>",
566
+ "lstrip": true,
567
+ "normalized": false,
568
+ "rstrip": true,
569
+ "single_word": false,
570
+ "special": true
571
+ },
572
+ "36068": {
573
+ "content": "<extra_id_27>",
574
+ "lstrip": true,
575
+ "normalized": false,
576
+ "rstrip": true,
577
+ "single_word": false,
578
+ "special": true
579
+ },
580
+ "36069": {
581
+ "content": "<extra_id_26>",
582
+ "lstrip": true,
583
+ "normalized": false,
584
+ "rstrip": true,
585
+ "single_word": false,
586
+ "special": true
587
+ },
588
+ "36070": {
589
+ "content": "<extra_id_25>",
590
+ "lstrip": true,
591
+ "normalized": false,
592
+ "rstrip": true,
593
+ "single_word": false,
594
+ "special": true
595
+ },
596
+ "36071": {
597
+ "content": "<extra_id_24>",
598
+ "lstrip": true,
599
+ "normalized": false,
600
+ "rstrip": true,
601
+ "single_word": false,
602
+ "special": true
603
+ },
604
+ "36072": {
605
+ "content": "<extra_id_23>",
606
+ "lstrip": true,
607
+ "normalized": false,
608
+ "rstrip": true,
609
+ "single_word": false,
610
+ "special": true
611
+ },
612
+ "36073": {
613
+ "content": "<extra_id_22>",
614
+ "lstrip": true,
615
+ "normalized": false,
616
+ "rstrip": true,
617
+ "single_word": false,
618
+ "special": true
619
+ },
620
+ "36074": {
621
+ "content": "<extra_id_21>",
622
+ "lstrip": true,
623
+ "normalized": false,
624
+ "rstrip": true,
625
+ "single_word": false,
626
+ "special": true
627
+ },
628
+ "36075": {
629
+ "content": "<extra_id_20>",
630
+ "lstrip": true,
631
+ "normalized": false,
632
+ "rstrip": true,
633
+ "single_word": false,
634
+ "special": true
635
+ },
636
+ "36076": {
637
+ "content": "<extra_id_19>",
638
+ "lstrip": true,
639
+ "normalized": false,
640
+ "rstrip": true,
641
+ "single_word": false,
642
+ "special": true
643
+ },
644
+ "36077": {
645
+ "content": "<extra_id_18>",
646
+ "lstrip": true,
647
+ "normalized": false,
648
+ "rstrip": true,
649
+ "single_word": false,
650
+ "special": true
651
+ },
652
+ "36078": {
653
+ "content": "<extra_id_17>",
654
+ "lstrip": true,
655
+ "normalized": false,
656
+ "rstrip": true,
657
+ "single_word": false,
658
+ "special": true
659
+ },
660
+ "36079": {
661
+ "content": "<extra_id_16>",
662
+ "lstrip": true,
663
+ "normalized": false,
664
+ "rstrip": true,
665
+ "single_word": false,
666
+ "special": true
667
+ },
668
+ "36080": {
669
+ "content": "<extra_id_15>",
670
+ "lstrip": true,
671
+ "normalized": false,
672
+ "rstrip": true,
673
+ "single_word": false,
674
+ "special": true
675
+ },
676
+ "36081": {
677
+ "content": "<extra_id_14>",
678
+ "lstrip": true,
679
+ "normalized": false,
680
+ "rstrip": true,
681
+ "single_word": false,
682
+ "special": true
683
+ },
684
+ "36082": {
685
+ "content": "<extra_id_13>",
686
+ "lstrip": true,
687
+ "normalized": false,
688
+ "rstrip": true,
689
+ "single_word": false,
690
+ "special": true
691
+ },
692
+ "36083": {
693
+ "content": "<extra_id_12>",
694
+ "lstrip": true,
695
+ "normalized": false,
696
+ "rstrip": true,
697
+ "single_word": false,
698
+ "special": true
699
+ },
700
+ "36084": {
701
+ "content": "<extra_id_11>",
702
+ "lstrip": true,
703
+ "normalized": false,
704
+ "rstrip": true,
705
+ "single_word": false,
706
+ "special": true
707
+ },
708
+ "36085": {
709
+ "content": "<extra_id_10>",
710
+ "lstrip": true,
711
+ "normalized": false,
712
+ "rstrip": true,
713
+ "single_word": false,
714
+ "special": true
715
+ },
716
+ "36086": {
717
+ "content": "<extra_id_9>",
718
+ "lstrip": true,
719
+ "normalized": false,
720
+ "rstrip": true,
721
+ "single_word": false,
722
+ "special": true
723
+ },
724
+ "36087": {
725
+ "content": "<extra_id_8>",
726
+ "lstrip": true,
727
+ "normalized": false,
728
+ "rstrip": true,
729
+ "single_word": false,
730
+ "special": true
731
+ },
732
+ "36088": {
733
+ "content": "<extra_id_7>",
734
+ "lstrip": true,
735
+ "normalized": false,
736
+ "rstrip": true,
737
+ "single_word": false,
738
+ "special": true
739
+ },
740
+ "36089": {
741
+ "content": "<extra_id_6>",
742
+ "lstrip": true,
743
+ "normalized": false,
744
+ "rstrip": true,
745
+ "single_word": false,
746
+ "special": true
747
+ },
748
+ "36090": {
749
+ "content": "<extra_id_5>",
750
+ "lstrip": true,
751
+ "normalized": false,
752
+ "rstrip": true,
753
+ "single_word": false,
754
+ "special": true
755
+ },
756
+ "36091": {
757
+ "content": "<extra_id_4>",
758
+ "lstrip": true,
759
+ "normalized": false,
760
+ "rstrip": true,
761
+ "single_word": false,
762
+ "special": true
763
+ },
764
+ "36092": {
765
+ "content": "<extra_id_3>",
766
+ "lstrip": true,
767
+ "normalized": false,
768
+ "rstrip": true,
769
+ "single_word": false,
770
+ "special": true
771
+ },
772
+ "36093": {
773
+ "content": "<extra_id_2>",
774
+ "lstrip": true,
775
+ "normalized": false,
776
+ "rstrip": true,
777
+ "single_word": false,
778
+ "special": true
779
+ },
780
+ "36094": {
781
+ "content": "<extra_id_1>",
782
+ "lstrip": true,
783
+ "normalized": false,
784
+ "rstrip": true,
785
+ "single_word": false,
786
+ "special": true
787
+ },
788
+ "36095": {
789
+ "content": "<extra_id_0>",
790
+ "lstrip": true,
791
+ "normalized": false,
792
+ "rstrip": true,
793
+ "single_word": false,
794
+ "special": true
795
+ }
796
+ },
797
+ "additional_special_tokens": [
798
+ "<extra_id_0>",
799
+ "<extra_id_1>",
800
+ "<extra_id_2>",
801
+ "<extra_id_3>",
802
+ "<extra_id_4>",
803
+ "<extra_id_5>",
804
+ "<extra_id_6>",
805
+ "<extra_id_7>",
806
+ "<extra_id_8>",
807
+ "<extra_id_9>",
808
+ "<extra_id_10>",
809
+ "<extra_id_11>",
810
+ "<extra_id_12>",
811
+ "<extra_id_13>",
812
+ "<extra_id_14>",
813
+ "<extra_id_15>",
814
+ "<extra_id_16>",
815
+ "<extra_id_17>",
816
+ "<extra_id_18>",
817
+ "<extra_id_19>",
818
+ "<extra_id_20>",
819
+ "<extra_id_21>",
820
+ "<extra_id_22>",
821
+ "<extra_id_23>",
822
+ "<extra_id_24>",
823
+ "<extra_id_25>",
824
+ "<extra_id_26>",
825
+ "<extra_id_27>",
826
+ "<extra_id_28>",
827
+ "<extra_id_29>",
828
+ "<extra_id_30>",
829
+ "<extra_id_31>",
830
+ "<extra_id_32>",
831
+ "<extra_id_33>",
832
+ "<extra_id_34>",
833
+ "<extra_id_35>",
834
+ "<extra_id_36>",
835
+ "<extra_id_37>",
836
+ "<extra_id_38>",
837
+ "<extra_id_39>",
838
+ "<extra_id_40>",
839
+ "<extra_id_41>",
840
+ "<extra_id_42>",
841
+ "<extra_id_43>",
842
+ "<extra_id_44>",
843
+ "<extra_id_45>",
844
+ "<extra_id_46>",
845
+ "<extra_id_47>",
846
+ "<extra_id_48>",
847
+ "<extra_id_49>",
848
+ "<extra_id_50>",
849
+ "<extra_id_51>",
850
+ "<extra_id_52>",
851
+ "<extra_id_53>",
852
+ "<extra_id_54>",
853
+ "<extra_id_55>",
854
+ "<extra_id_56>",
855
+ "<extra_id_57>",
856
+ "<extra_id_58>",
857
+ "<extra_id_59>",
858
+ "<extra_id_60>",
859
+ "<extra_id_61>",
860
+ "<extra_id_62>",
861
+ "<extra_id_63>",
862
+ "<extra_id_64>",
863
+ "<extra_id_65>",
864
+ "<extra_id_66>",
865
+ "<extra_id_67>",
866
+ "<extra_id_68>",
867
+ "<extra_id_69>",
868
+ "<extra_id_70>",
869
+ "<extra_id_71>",
870
+ "<extra_id_72>",
871
+ "<extra_id_73>",
872
+ "<extra_id_74>",
873
+ "<extra_id_75>",
874
+ "<extra_id_76>",
875
+ "<extra_id_77>",
876
+ "<extra_id_78>",
877
+ "<extra_id_79>",
878
+ "<extra_id_80>",
879
+ "<extra_id_81>",
880
+ "<extra_id_82>",
881
+ "<extra_id_83>",
882
+ "<extra_id_84>",
883
+ "<extra_id_85>",
884
+ "<extra_id_86>",
885
+ "<extra_id_87>",
886
+ "<extra_id_88>",
887
+ "<extra_id_89>",
888
+ "<extra_id_90>",
889
+ "<extra_id_91>",
890
+ "<extra_id_92>",
891
+ "<extra_id_93>",
892
+ "<extra_id_94>",
893
+ "<extra_id_95>"
894
+ ],
895
+ "clean_up_tokenization_spaces": false,
896
+ "eos_token": "</s>",
897
+ "extra_ids": 96,
898
+ "extra_special_tokens": {},
899
+ "legacy": true,
900
+ "model_max_length": 1000000000000000019884624838656,
901
+ "pad_token": "<pad>",
902
+ "sp_model_kwargs": {},
903
+ "tokenizer_class": "T5Tokenizer",
904
+ "unk_token": "<unk>"
905
+ }
HVU_QA/t5-viet-qg-finetuned/checkpoint-2500/trainer_state.json ADDED
@@ -0,0 +1,1784 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 0.03899669308042678,
6
+ "eval_steps": 500,
7
+ "global_step": 2500,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.00015598677232170712,
14
+ "grad_norm": 3.216968059539795,
15
+ "learning_rate": 0.0001999906407936607,
16
+ "loss": 4.7744,
17
+ "step": 10
18
+ },
19
+ {
20
+ "epoch": 0.00031197354464341425,
21
+ "grad_norm": 9.366360664367676,
22
+ "learning_rate": 0.00019998024167550594,
23
+ "loss": 0.9639,
24
+ "step": 20
25
+ },
26
+ {
27
+ "epoch": 0.00046796031696512137,
28
+ "grad_norm": 4.382473945617676,
29
+ "learning_rate": 0.00019996984255735113,
30
+ "loss": 0.6338,
31
+ "step": 30
32
+ },
33
+ {
34
+ "epoch": 0.0006239470892868285,
35
+ "grad_norm": 14.415914535522461,
36
+ "learning_rate": 0.00019995944343919637,
37
+ "loss": 0.4962,
38
+ "step": 40
39
+ },
40
+ {
41
+ "epoch": 0.0007799338616085356,
42
+ "grad_norm": 7.440260410308838,
43
+ "learning_rate": 0.00019994904432104157,
44
+ "loss": 0.4619,
45
+ "step": 50
46
+ },
47
+ {
48
+ "epoch": 0.0009359206339302427,
49
+ "grad_norm": 27.18022918701172,
50
+ "learning_rate": 0.00019993864520288681,
51
+ "loss": 0.5335,
52
+ "step": 60
53
+ },
54
+ {
55
+ "epoch": 0.0010919074062519499,
56
+ "grad_norm": 7.954453468322754,
57
+ "learning_rate": 0.000199928246084732,
58
+ "loss": 0.8085,
59
+ "step": 70
60
+ },
61
+ {
62
+ "epoch": 0.001247894178573657,
63
+ "grad_norm": 4.762608528137207,
64
+ "learning_rate": 0.00019991784696657725,
65
+ "loss": 0.6333,
66
+ "step": 80
67
+ },
68
+ {
69
+ "epoch": 0.0014038809508953641,
70
+ "grad_norm": 9.577717781066895,
71
+ "learning_rate": 0.00019990744784842245,
72
+ "loss": 0.4558,
73
+ "step": 90
74
+ },
75
+ {
76
+ "epoch": 0.0015598677232170712,
77
+ "grad_norm": 7.586806297302246,
78
+ "learning_rate": 0.0001998970487302677,
79
+ "loss": 0.3227,
80
+ "step": 100
81
+ },
82
+ {
83
+ "epoch": 0.0017158544955387784,
84
+ "grad_norm": 4.262249946594238,
85
+ "learning_rate": 0.00019988664961211288,
86
+ "loss": 0.3261,
87
+ "step": 110
88
+ },
89
+ {
90
+ "epoch": 0.0018718412678604855,
91
+ "grad_norm": 6.639514923095703,
92
+ "learning_rate": 0.00019987625049395813,
93
+ "loss": 0.6788,
94
+ "step": 120
95
+ },
96
+ {
97
+ "epoch": 0.0020278280401821924,
98
+ "grad_norm": 3.7448127269744873,
99
+ "learning_rate": 0.00019986585137580332,
100
+ "loss": 0.4224,
101
+ "step": 130
102
+ },
103
+ {
104
+ "epoch": 0.0021838148125038997,
105
+ "grad_norm": 13.205582618713379,
106
+ "learning_rate": 0.00019985545225764857,
107
+ "loss": 0.4074,
108
+ "step": 140
109
+ },
110
+ {
111
+ "epoch": 0.0023398015848256066,
112
+ "grad_norm": 5.55623722076416,
113
+ "learning_rate": 0.00019984505313949376,
114
+ "loss": 0.3408,
115
+ "step": 150
116
+ },
117
+ {
118
+ "epoch": 0.002495788357147314,
119
+ "grad_norm": 10.253294944763184,
120
+ "learning_rate": 0.000199834654021339,
121
+ "loss": 0.4513,
122
+ "step": 160
123
+ },
124
+ {
125
+ "epoch": 0.002651775129469021,
126
+ "grad_norm": 12.778948783874512,
127
+ "learning_rate": 0.0001998242549031842,
128
+ "loss": 0.4287,
129
+ "step": 170
130
+ },
131
+ {
132
+ "epoch": 0.0028077619017907282,
133
+ "grad_norm": 8.961956977844238,
134
+ "learning_rate": 0.00019981385578502945,
135
+ "loss": 0.6746,
136
+ "step": 180
137
+ },
138
+ {
139
+ "epoch": 0.002963748674112435,
140
+ "grad_norm": 6.611206531524658,
141
+ "learning_rate": 0.00019980345666687467,
142
+ "loss": 0.4106,
143
+ "step": 190
144
+ },
145
+ {
146
+ "epoch": 0.0031197354464341425,
147
+ "grad_norm": 4.179671764373779,
148
+ "learning_rate": 0.0001997930575487199,
149
+ "loss": 0.4234,
150
+ "step": 200
151
+ },
152
+ {
153
+ "epoch": 0.0032757222187558494,
154
+ "grad_norm": 7.957104206085205,
155
+ "learning_rate": 0.0001997826584305651,
156
+ "loss": 0.4062,
157
+ "step": 210
158
+ },
159
+ {
160
+ "epoch": 0.0034317089910775567,
161
+ "grad_norm": 10.04617977142334,
162
+ "learning_rate": 0.00019977225931241033,
163
+ "loss": 0.4792,
164
+ "step": 220
165
+ },
166
+ {
167
+ "epoch": 0.0035876957633992636,
168
+ "grad_norm": 5.119971752166748,
169
+ "learning_rate": 0.00019976186019425555,
170
+ "loss": 0.3385,
171
+ "step": 230
172
+ },
173
+ {
174
+ "epoch": 0.003743682535720971,
175
+ "grad_norm": 6.039366722106934,
176
+ "learning_rate": 0.00019975146107610077,
177
+ "loss": 0.4406,
178
+ "step": 240
179
+ },
180
+ {
181
+ "epoch": 0.003899669308042678,
182
+ "grad_norm": 9.345914840698242,
183
+ "learning_rate": 0.000199741061957946,
184
+ "loss": 0.5163,
185
+ "step": 250
186
+ },
187
+ {
188
+ "epoch": 0.004055656080364385,
189
+ "grad_norm": 3.707460641860962,
190
+ "learning_rate": 0.0001997306628397912,
191
+ "loss": 0.2971,
192
+ "step": 260
193
+ },
194
+ {
195
+ "epoch": 0.0042116428526860926,
196
+ "grad_norm": 11.500152587890625,
197
+ "learning_rate": 0.00019972026372163643,
198
+ "loss": 0.3659,
199
+ "step": 270
200
+ },
201
+ {
202
+ "epoch": 0.0043676296250077995,
203
+ "grad_norm": 0.7893226742744446,
204
+ "learning_rate": 0.00019970986460348165,
205
+ "loss": 0.5052,
206
+ "step": 280
207
+ },
208
+ {
209
+ "epoch": 0.004523616397329506,
210
+ "grad_norm": 3.2410888671875,
211
+ "learning_rate": 0.00019969946548532687,
212
+ "loss": 0.4038,
213
+ "step": 290
214
+ },
215
+ {
216
+ "epoch": 0.004679603169651213,
217
+ "grad_norm": 4.468445777893066,
218
+ "learning_rate": 0.00019968906636717208,
219
+ "loss": 0.6076,
220
+ "step": 300
221
+ },
222
+ {
223
+ "epoch": 0.004835589941972921,
224
+ "grad_norm": 8.308616638183594,
225
+ "learning_rate": 0.0001996786672490173,
226
+ "loss": 0.4465,
227
+ "step": 310
228
+ },
229
+ {
230
+ "epoch": 0.004991576714294628,
231
+ "grad_norm": 3.736629009246826,
232
+ "learning_rate": 0.0001996682681308625,
233
+ "loss": 0.4133,
234
+ "step": 320
235
+ },
236
+ {
237
+ "epoch": 0.005147563486616335,
238
+ "grad_norm": 2.157092809677124,
239
+ "learning_rate": 0.00019965786901270774,
240
+ "loss": 0.3663,
241
+ "step": 330
242
+ },
243
+ {
244
+ "epoch": 0.005303550258938042,
245
+ "grad_norm": 5.357859134674072,
246
+ "learning_rate": 0.00019964746989455294,
247
+ "loss": 0.5599,
248
+ "step": 340
249
+ },
250
+ {
251
+ "epoch": 0.0054595370312597495,
252
+ "grad_norm": 5.352982044219971,
253
+ "learning_rate": 0.00019963707077639818,
254
+ "loss": 0.6796,
255
+ "step": 350
256
+ },
257
+ {
258
+ "epoch": 0.0056155238035814565,
259
+ "grad_norm": 3.7645204067230225,
260
+ "learning_rate": 0.00019962667165824338,
261
+ "loss": 0.5774,
262
+ "step": 360
263
+ },
264
+ {
265
+ "epoch": 0.005771510575903163,
266
+ "grad_norm": 6.222738265991211,
267
+ "learning_rate": 0.00019961627254008862,
268
+ "loss": 0.641,
269
+ "step": 370
270
+ },
271
+ {
272
+ "epoch": 0.00592749734822487,
273
+ "grad_norm": 6.460178852081299,
274
+ "learning_rate": 0.00019960587342193381,
275
+ "loss": 0.3702,
276
+ "step": 380
277
+ },
278
+ {
279
+ "epoch": 0.006083484120546578,
280
+ "grad_norm": 3.1196858882904053,
281
+ "learning_rate": 0.00019959547430377906,
282
+ "loss": 0.5036,
283
+ "step": 390
284
+ },
285
+ {
286
+ "epoch": 0.006239470892868285,
287
+ "grad_norm": 7.03672981262207,
288
+ "learning_rate": 0.00019958507518562425,
289
+ "loss": 0.41,
290
+ "step": 400
291
+ },
292
+ {
293
+ "epoch": 0.006395457665189992,
294
+ "grad_norm": 15.546775817871094,
295
+ "learning_rate": 0.0001995746760674695,
296
+ "loss": 0.5623,
297
+ "step": 410
298
+ },
299
+ {
300
+ "epoch": 0.006551444437511699,
301
+ "grad_norm": 4.868813514709473,
302
+ "learning_rate": 0.0001995642769493147,
303
+ "loss": 0.4,
304
+ "step": 420
305
+ },
306
+ {
307
+ "epoch": 0.0067074312098334065,
308
+ "grad_norm": 3.5767409801483154,
309
+ "learning_rate": 0.00019955387783115994,
310
+ "loss": 0.5193,
311
+ "step": 430
312
+ },
313
+ {
314
+ "epoch": 0.0068634179821551134,
315
+ "grad_norm": 10.07199764251709,
316
+ "learning_rate": 0.00019954347871300513,
317
+ "loss": 0.5692,
318
+ "step": 440
319
+ },
320
+ {
321
+ "epoch": 0.00701940475447682,
322
+ "grad_norm": 2.809025764465332,
323
+ "learning_rate": 0.00019953307959485038,
324
+ "loss": 0.4361,
325
+ "step": 450
326
+ },
327
+ {
328
+ "epoch": 0.007175391526798527,
329
+ "grad_norm": 14.244531631469727,
330
+ "learning_rate": 0.00019952268047669557,
331
+ "loss": 0.3646,
332
+ "step": 460
333
+ },
334
+ {
335
+ "epoch": 0.007331378299120235,
336
+ "grad_norm": 5.827071666717529,
337
+ "learning_rate": 0.00019951228135854082,
338
+ "loss": 0.5302,
339
+ "step": 470
340
+ },
341
+ {
342
+ "epoch": 0.007487365071441942,
343
+ "grad_norm": 9.067925453186035,
344
+ "learning_rate": 0.000199501882240386,
345
+ "loss": 0.5045,
346
+ "step": 480
347
+ },
348
+ {
349
+ "epoch": 0.007643351843763649,
350
+ "grad_norm": 8.674382209777832,
351
+ "learning_rate": 0.00019949148312223126,
352
+ "loss": 0.5778,
353
+ "step": 490
354
+ },
355
+ {
356
+ "epoch": 0.007799338616085356,
357
+ "grad_norm": 8.94734001159668,
358
+ "learning_rate": 0.00019948108400407645,
359
+ "loss": 0.4141,
360
+ "step": 500
361
+ },
362
+ {
363
+ "epoch": 0.007955325388407063,
364
+ "grad_norm": 10.810057640075684,
365
+ "learning_rate": 0.0001994706848859217,
366
+ "loss": 0.4069,
367
+ "step": 510
368
+ },
369
+ {
370
+ "epoch": 0.00811131216072877,
371
+ "grad_norm": 9.531766891479492,
372
+ "learning_rate": 0.0001994602857677669,
373
+ "loss": 0.5935,
374
+ "step": 520
375
+ },
376
+ {
377
+ "epoch": 0.008267298933050476,
378
+ "grad_norm": 13.168464660644531,
379
+ "learning_rate": 0.00019944988664961214,
380
+ "loss": 0.4725,
381
+ "step": 530
382
+ },
383
+ {
384
+ "epoch": 0.008423285705372185,
385
+ "grad_norm": 4.697941780090332,
386
+ "learning_rate": 0.00019943948753145733,
387
+ "loss": 0.4516,
388
+ "step": 540
389
+ },
390
+ {
391
+ "epoch": 0.008579272477693892,
392
+ "grad_norm": 2.92124342918396,
393
+ "learning_rate": 0.00019942908841330258,
394
+ "loss": 0.4562,
395
+ "step": 550
396
+ },
397
+ {
398
+ "epoch": 0.008735259250015599,
399
+ "grad_norm": 9.23039722442627,
400
+ "learning_rate": 0.00019941868929514777,
401
+ "loss": 0.4051,
402
+ "step": 560
403
+ },
404
+ {
405
+ "epoch": 0.008891246022337306,
406
+ "grad_norm": 10.528753280639648,
407
+ "learning_rate": 0.00019940829017699302,
408
+ "loss": 0.3475,
409
+ "step": 570
410
+ },
411
+ {
412
+ "epoch": 0.009047232794659013,
413
+ "grad_norm": 6.963693618774414,
414
+ "learning_rate": 0.0001993978910588382,
415
+ "loss": 0.5338,
416
+ "step": 580
417
+ },
418
+ {
419
+ "epoch": 0.00920321956698072,
420
+ "grad_norm": 6.09390926361084,
421
+ "learning_rate": 0.00019938749194068345,
422
+ "loss": 0.6501,
423
+ "step": 590
424
+ },
425
+ {
426
+ "epoch": 0.009359206339302427,
427
+ "grad_norm": 4.319295406341553,
428
+ "learning_rate": 0.00019937709282252865,
429
+ "loss": 0.4163,
430
+ "step": 600
431
+ },
432
+ {
433
+ "epoch": 0.009515193111624133,
434
+ "grad_norm": 7.514628887176514,
435
+ "learning_rate": 0.0001993666937043739,
436
+ "loss": 0.5193,
437
+ "step": 610
438
+ },
439
+ {
440
+ "epoch": 0.009671179883945842,
441
+ "grad_norm": 5.011287212371826,
442
+ "learning_rate": 0.00019935629458621909,
443
+ "loss": 0.5743,
444
+ "step": 620
445
+ },
446
+ {
447
+ "epoch": 0.009827166656267549,
448
+ "grad_norm": 6.090816974639893,
449
+ "learning_rate": 0.00019934589546806433,
450
+ "loss": 0.3902,
451
+ "step": 630
452
+ },
453
+ {
454
+ "epoch": 0.009983153428589256,
455
+ "grad_norm": 10.161015510559082,
456
+ "learning_rate": 0.00019933549634990953,
457
+ "loss": 0.4463,
458
+ "step": 640
459
+ },
460
+ {
461
+ "epoch": 0.010139140200910963,
462
+ "grad_norm": 6.14881706237793,
463
+ "learning_rate": 0.00019932509723175477,
464
+ "loss": 0.5009,
465
+ "step": 650
466
+ },
467
+ {
468
+ "epoch": 0.01029512697323267,
469
+ "grad_norm": 3.2960498332977295,
470
+ "learning_rate": 0.00019931469811359996,
471
+ "loss": 0.4597,
472
+ "step": 660
473
+ },
474
+ {
475
+ "epoch": 0.010451113745554377,
476
+ "grad_norm": 7.595890998840332,
477
+ "learning_rate": 0.0001993042989954452,
478
+ "loss": 0.5948,
479
+ "step": 670
480
+ },
481
+ {
482
+ "epoch": 0.010607100517876084,
483
+ "grad_norm": 7.178264617919922,
484
+ "learning_rate": 0.0001992938998772904,
485
+ "loss": 0.7185,
486
+ "step": 680
487
+ },
488
+ {
489
+ "epoch": 0.01076308729019779,
490
+ "grad_norm": 4.092066764831543,
491
+ "learning_rate": 0.00019928350075913565,
492
+ "loss": 0.3544,
493
+ "step": 690
494
+ },
495
+ {
496
+ "epoch": 0.010919074062519499,
497
+ "grad_norm": 13.203527450561523,
498
+ "learning_rate": 0.00019927310164098084,
499
+ "loss": 0.5719,
500
+ "step": 700
501
+ },
502
+ {
503
+ "epoch": 0.011075060834841206,
504
+ "grad_norm": 5.39072847366333,
505
+ "learning_rate": 0.0001992627025228261,
506
+ "loss": 0.5527,
507
+ "step": 710
508
+ },
509
+ {
510
+ "epoch": 0.011231047607162913,
511
+ "grad_norm": 5.769185543060303,
512
+ "learning_rate": 0.00019925230340467128,
513
+ "loss": 0.4555,
514
+ "step": 720
515
+ },
516
+ {
517
+ "epoch": 0.01138703437948462,
518
+ "grad_norm": 3.0080807209014893,
519
+ "learning_rate": 0.00019924190428651653,
520
+ "loss": 0.277,
521
+ "step": 730
522
+ },
523
+ {
524
+ "epoch": 0.011543021151806327,
525
+ "grad_norm": 1.0591176748275757,
526
+ "learning_rate": 0.00019923150516836172,
527
+ "loss": 0.513,
528
+ "step": 740
529
+ },
530
+ {
531
+ "epoch": 0.011699007924128034,
532
+ "grad_norm": 8.458100318908691,
533
+ "learning_rate": 0.00019922110605020697,
534
+ "loss": 0.5038,
535
+ "step": 750
536
+ },
537
+ {
538
+ "epoch": 0.01185499469644974,
539
+ "grad_norm": 3.121551990509033,
540
+ "learning_rate": 0.00019921070693205216,
541
+ "loss": 0.2639,
542
+ "step": 760
543
+ },
544
+ {
545
+ "epoch": 0.012010981468771447,
546
+ "grad_norm": 6.6408610343933105,
547
+ "learning_rate": 0.00019920030781389738,
548
+ "loss": 0.6632,
549
+ "step": 770
550
+ },
551
+ {
552
+ "epoch": 0.012166968241093156,
553
+ "grad_norm": 7.180611610412598,
554
+ "learning_rate": 0.0001991899086957426,
555
+ "loss": 0.5173,
556
+ "step": 780
557
+ },
558
+ {
559
+ "epoch": 0.012322955013414863,
560
+ "grad_norm": 4.404491901397705,
561
+ "learning_rate": 0.00019917950957758782,
562
+ "loss": 0.3628,
563
+ "step": 790
564
+ },
565
+ {
566
+ "epoch": 0.01247894178573657,
567
+ "grad_norm": 5.945431232452393,
568
+ "learning_rate": 0.00019916911045943304,
569
+ "loss": 0.3574,
570
+ "step": 800
571
+ },
572
+ {
573
+ "epoch": 0.012634928558058277,
574
+ "grad_norm": 5.105412483215332,
575
+ "learning_rate": 0.00019915871134127826,
576
+ "loss": 0.7038,
577
+ "step": 810
578
+ },
579
+ {
580
+ "epoch": 0.012790915330379984,
581
+ "grad_norm": 3.7237210273742676,
582
+ "learning_rate": 0.00019914831222312348,
583
+ "loss": 0.4426,
584
+ "step": 820
585
+ },
586
+ {
587
+ "epoch": 0.01294690210270169,
588
+ "grad_norm": 10.387605667114258,
589
+ "learning_rate": 0.0001991379131049687,
590
+ "loss": 0.3152,
591
+ "step": 830
592
+ },
593
+ {
594
+ "epoch": 0.013102888875023398,
595
+ "grad_norm": 2.875211477279663,
596
+ "learning_rate": 0.00019912751398681392,
597
+ "loss": 0.424,
598
+ "step": 840
599
+ },
600
+ {
601
+ "epoch": 0.013258875647345104,
602
+ "grad_norm": 1.0501248836517334,
603
+ "learning_rate": 0.00019911711486865914,
604
+ "loss": 0.5571,
605
+ "step": 850
606
+ },
607
+ {
608
+ "epoch": 0.013414862419666813,
609
+ "grad_norm": 6.748145580291748,
610
+ "learning_rate": 0.00019910671575050436,
611
+ "loss": 0.5226,
612
+ "step": 860
613
+ },
614
+ {
615
+ "epoch": 0.01357084919198852,
616
+ "grad_norm": 6.29374885559082,
617
+ "learning_rate": 0.00019909631663234958,
618
+ "loss": 0.3036,
619
+ "step": 870
620
+ },
621
+ {
622
+ "epoch": 0.013726835964310227,
623
+ "grad_norm": 5.8310699462890625,
624
+ "learning_rate": 0.0001990859175141948,
625
+ "loss": 0.4704,
626
+ "step": 880
627
+ },
628
+ {
629
+ "epoch": 0.013882822736631934,
630
+ "grad_norm": 1.4619086980819702,
631
+ "learning_rate": 0.00019907551839604002,
632
+ "loss": 0.3639,
633
+ "step": 890
634
+ },
635
+ {
636
+ "epoch": 0.01403880950895364,
637
+ "grad_norm": 8.177449226379395,
638
+ "learning_rate": 0.00019906511927788524,
639
+ "loss": 0.3853,
640
+ "step": 900
641
+ },
642
+ {
643
+ "epoch": 0.014194796281275348,
644
+ "grad_norm": 2.045549154281616,
645
+ "learning_rate": 0.00019905472015973046,
646
+ "loss": 0.3728,
647
+ "step": 910
648
+ },
649
+ {
650
+ "epoch": 0.014350783053597055,
651
+ "grad_norm": 6.4138875007629395,
652
+ "learning_rate": 0.00019904432104157568,
653
+ "loss": 0.4615,
654
+ "step": 920
655
+ },
656
+ {
657
+ "epoch": 0.014506769825918761,
658
+ "grad_norm": 2.442739486694336,
659
+ "learning_rate": 0.0001990339219234209,
660
+ "loss": 0.3592,
661
+ "step": 930
662
+ },
663
+ {
664
+ "epoch": 0.01466275659824047,
665
+ "grad_norm": 3.0741333961486816,
666
+ "learning_rate": 0.00019902352280526611,
667
+ "loss": 0.4118,
668
+ "step": 940
669
+ },
670
+ {
671
+ "epoch": 0.014818743370562177,
672
+ "grad_norm": 5.163224220275879,
673
+ "learning_rate": 0.00019901312368711133,
674
+ "loss": 0.7433,
675
+ "step": 950
676
+ },
677
+ {
678
+ "epoch": 0.014974730142883884,
679
+ "grad_norm": 5.6979193687438965,
680
+ "learning_rate": 0.00019900272456895655,
681
+ "loss": 0.3094,
682
+ "step": 960
683
+ },
684
+ {
685
+ "epoch": 0.01513071691520559,
686
+ "grad_norm": 3.2027347087860107,
687
+ "learning_rate": 0.00019899232545080177,
688
+ "loss": 0.3749,
689
+ "step": 970
690
+ },
691
+ {
692
+ "epoch": 0.015286703687527298,
693
+ "grad_norm": 9.097466468811035,
694
+ "learning_rate": 0.000198981926332647,
695
+ "loss": 0.5344,
696
+ "step": 980
697
+ },
698
+ {
699
+ "epoch": 0.015442690459849005,
700
+ "grad_norm": 3.730250835418701,
701
+ "learning_rate": 0.0001989715272144922,
702
+ "loss": 0.4085,
703
+ "step": 990
704
+ },
705
+ {
706
+ "epoch": 0.015598677232170712,
707
+ "grad_norm": 4.03507661819458,
708
+ "learning_rate": 0.00019896112809633743,
709
+ "loss": 0.7935,
710
+ "step": 1000
711
+ },
712
+ {
713
+ "epoch": 0.01575466400449242,
714
+ "grad_norm": 6.676806449890137,
715
+ "learning_rate": 0.00019895072897818265,
716
+ "loss": 0.4276,
717
+ "step": 1010
718
+ },
719
+ {
720
+ "epoch": 0.015910650776814125,
721
+ "grad_norm": 2.294931173324585,
722
+ "learning_rate": 0.00019894032986002787,
723
+ "loss": 0.2544,
724
+ "step": 1020
725
+ },
726
+ {
727
+ "epoch": 0.016066637549135832,
728
+ "grad_norm": 15.660760879516602,
729
+ "learning_rate": 0.0001989299307418731,
730
+ "loss": 0.2441,
731
+ "step": 1030
732
+ },
733
+ {
734
+ "epoch": 0.01622262432145754,
735
+ "grad_norm": 3.4223055839538574,
736
+ "learning_rate": 0.00019891953162371834,
737
+ "loss": 0.4861,
738
+ "step": 1040
739
+ },
740
+ {
741
+ "epoch": 0.016378611093779246,
742
+ "grad_norm": 8.601675987243652,
743
+ "learning_rate": 0.00019890913250556353,
744
+ "loss": 0.7212,
745
+ "step": 1050
746
+ },
747
+ {
748
+ "epoch": 0.016534597866100953,
749
+ "grad_norm": 3.814286708831787,
750
+ "learning_rate": 0.00019889873338740878,
751
+ "loss": 0.4785,
752
+ "step": 1060
753
+ },
754
+ {
755
+ "epoch": 0.016690584638422663,
756
+ "grad_norm": 10.6279296875,
757
+ "learning_rate": 0.00019888833426925397,
758
+ "loss": 0.278,
759
+ "step": 1070
760
+ },
761
+ {
762
+ "epoch": 0.01684657141074437,
763
+ "grad_norm": 2.3790910243988037,
764
+ "learning_rate": 0.00019887793515109922,
765
+ "loss": 0.3589,
766
+ "step": 1080
767
+ },
768
+ {
769
+ "epoch": 0.017002558183066077,
770
+ "grad_norm": 8.939804077148438,
771
+ "learning_rate": 0.0001988675360329444,
772
+ "loss": 0.5794,
773
+ "step": 1090
774
+ },
775
+ {
776
+ "epoch": 0.017158544955387784,
777
+ "grad_norm": 7.628971576690674,
778
+ "learning_rate": 0.00019885713691478966,
779
+ "loss": 0.5317,
780
+ "step": 1100
781
+ },
782
+ {
783
+ "epoch": 0.01731453172770949,
784
+ "grad_norm": 1.8617459535598755,
785
+ "learning_rate": 0.00019884673779663485,
786
+ "loss": 0.433,
787
+ "step": 1110
788
+ },
789
+ {
790
+ "epoch": 0.017470518500031198,
791
+ "grad_norm": 6.928651809692383,
792
+ "learning_rate": 0.0001988363386784801,
793
+ "loss": 0.216,
794
+ "step": 1120
795
+ },
796
+ {
797
+ "epoch": 0.017626505272352905,
798
+ "grad_norm": 3.6206703186035156,
799
+ "learning_rate": 0.0001988259395603253,
800
+ "loss": 0.4963,
801
+ "step": 1130
802
+ },
803
+ {
804
+ "epoch": 0.01778249204467461,
805
+ "grad_norm": 4.615293025970459,
806
+ "learning_rate": 0.00019881554044217053,
807
+ "loss": 0.3579,
808
+ "step": 1140
809
+ },
810
+ {
811
+ "epoch": 0.01793847881699632,
812
+ "grad_norm": 3.3206489086151123,
813
+ "learning_rate": 0.00019880514132401573,
814
+ "loss": 0.3717,
815
+ "step": 1150
816
+ },
817
+ {
818
+ "epoch": 0.018094465589318025,
819
+ "grad_norm": 7.5789408683776855,
820
+ "learning_rate": 0.00019879474220586097,
821
+ "loss": 0.2646,
822
+ "step": 1160
823
+ },
824
+ {
825
+ "epoch": 0.018250452361639732,
826
+ "grad_norm": 4.467035293579102,
827
+ "learning_rate": 0.00019878434308770617,
828
+ "loss": 0.1804,
829
+ "step": 1170
830
+ },
831
+ {
832
+ "epoch": 0.01840643913396144,
833
+ "grad_norm": 9.318495750427246,
834
+ "learning_rate": 0.0001987739439695514,
835
+ "loss": 0.3257,
836
+ "step": 1180
837
+ },
838
+ {
839
+ "epoch": 0.018562425906283146,
840
+ "grad_norm": 3.1046931743621826,
841
+ "learning_rate": 0.0001987635448513966,
842
+ "loss": 0.4267,
843
+ "step": 1190
844
+ },
845
+ {
846
+ "epoch": 0.018718412678604853,
847
+ "grad_norm": 4.199711799621582,
848
+ "learning_rate": 0.00019875314573324185,
849
+ "loss": 0.6676,
850
+ "step": 1200
851
+ },
852
+ {
853
+ "epoch": 0.01887439945092656,
854
+ "grad_norm": 12.733593940734863,
855
+ "learning_rate": 0.00019874274661508705,
856
+ "loss": 0.5991,
857
+ "step": 1210
858
+ },
859
+ {
860
+ "epoch": 0.019030386223248267,
861
+ "grad_norm": 5.1266984939575195,
862
+ "learning_rate": 0.0001987323474969323,
863
+ "loss": 0.2945,
864
+ "step": 1220
865
+ },
866
+ {
867
+ "epoch": 0.019186372995569977,
868
+ "grad_norm": 6.397432804107666,
869
+ "learning_rate": 0.00019872194837877748,
870
+ "loss": 0.4774,
871
+ "step": 1230
872
+ },
873
+ {
874
+ "epoch": 0.019342359767891684,
875
+ "grad_norm": 3.761129379272461,
876
+ "learning_rate": 0.0001987115492606227,
877
+ "loss": 0.4345,
878
+ "step": 1240
879
+ },
880
+ {
881
+ "epoch": 0.01949834654021339,
882
+ "grad_norm": 2.7355475425720215,
883
+ "learning_rate": 0.00019870115014246792,
884
+ "loss": 0.497,
885
+ "step": 1250
886
+ },
887
+ {
888
+ "epoch": 0.019654333312535098,
889
+ "grad_norm": 2.7893755435943604,
890
+ "learning_rate": 0.00019869075102431314,
891
+ "loss": 0.2169,
892
+ "step": 1260
893
+ },
894
+ {
895
+ "epoch": 0.019810320084856805,
896
+ "grad_norm": 10.722111701965332,
897
+ "learning_rate": 0.00019868035190615836,
898
+ "loss": 0.3867,
899
+ "step": 1270
900
+ },
901
+ {
902
+ "epoch": 0.019966306857178512,
903
+ "grad_norm": 11.647889137268066,
904
+ "learning_rate": 0.00019866995278800358,
905
+ "loss": 0.4138,
906
+ "step": 1280
907
+ },
908
+ {
909
+ "epoch": 0.02012229362950022,
910
+ "grad_norm": 1.4313205480575562,
911
+ "learning_rate": 0.0001986595536698488,
912
+ "loss": 0.3581,
913
+ "step": 1290
914
+ },
915
+ {
916
+ "epoch": 0.020278280401821926,
917
+ "grad_norm": 5.193609714508057,
918
+ "learning_rate": 0.00019864915455169402,
919
+ "loss": 0.5259,
920
+ "step": 1300
921
+ },
922
+ {
923
+ "epoch": 0.020434267174143633,
924
+ "grad_norm": 3.1939520835876465,
925
+ "learning_rate": 0.00019863875543353924,
926
+ "loss": 0.5391,
927
+ "step": 1310
928
+ },
929
+ {
930
+ "epoch": 0.02059025394646534,
931
+ "grad_norm": 7.716986179351807,
932
+ "learning_rate": 0.00019862835631538446,
933
+ "loss": 0.4773,
934
+ "step": 1320
935
+ },
936
+ {
937
+ "epoch": 0.020746240718787046,
938
+ "grad_norm": 3.191506862640381,
939
+ "learning_rate": 0.00019861795719722968,
940
+ "loss": 0.6051,
941
+ "step": 1330
942
+ },
943
+ {
944
+ "epoch": 0.020902227491108753,
945
+ "grad_norm": 5.67583703994751,
946
+ "learning_rate": 0.0001986075580790749,
947
+ "loss": 0.2716,
948
+ "step": 1340
949
+ },
950
+ {
951
+ "epoch": 0.02105821426343046,
952
+ "grad_norm": 3.1657073497772217,
953
+ "learning_rate": 0.00019859715896092012,
954
+ "loss": 0.4838,
955
+ "step": 1350
956
+ },
957
+ {
958
+ "epoch": 0.021214201035752167,
959
+ "grad_norm": 5.144644260406494,
960
+ "learning_rate": 0.00019858675984276534,
961
+ "loss": 0.3185,
962
+ "step": 1360
963
+ },
964
+ {
965
+ "epoch": 0.021370187808073874,
966
+ "grad_norm": 0.3853607177734375,
967
+ "learning_rate": 0.00019857636072461056,
968
+ "loss": 0.21,
969
+ "step": 1370
970
+ },
971
+ {
972
+ "epoch": 0.02152617458039558,
973
+ "grad_norm": 15.329313278198242,
974
+ "learning_rate": 0.00019856596160645578,
975
+ "loss": 0.4216,
976
+ "step": 1380
977
+ },
978
+ {
979
+ "epoch": 0.02168216135271729,
980
+ "grad_norm": 1.9162156581878662,
981
+ "learning_rate": 0.000198555562488301,
982
+ "loss": 0.3766,
983
+ "step": 1390
984
+ },
985
+ {
986
+ "epoch": 0.021838148125038998,
987
+ "grad_norm": 2.5800933837890625,
988
+ "learning_rate": 0.00019854516337014622,
989
+ "loss": 0.4478,
990
+ "step": 1400
991
+ },
992
+ {
993
+ "epoch": 0.021994134897360705,
994
+ "grad_norm": 12.898444175720215,
995
+ "learning_rate": 0.00019853476425199144,
996
+ "loss": 0.7923,
997
+ "step": 1410
998
+ },
999
+ {
1000
+ "epoch": 0.022150121669682412,
1001
+ "grad_norm": 11.576621055603027,
1002
+ "learning_rate": 0.00019852436513383666,
1003
+ "loss": 0.3582,
1004
+ "step": 1420
1005
+ },
1006
+ {
1007
+ "epoch": 0.02230610844200412,
1008
+ "grad_norm": 10.80225658416748,
1009
+ "learning_rate": 0.00019851396601568188,
1010
+ "loss": 0.4698,
1011
+ "step": 1430
1012
+ },
1013
+ {
1014
+ "epoch": 0.022462095214325826,
1015
+ "grad_norm": 5.019324779510498,
1016
+ "learning_rate": 0.0001985035668975271,
1017
+ "loss": 0.4463,
1018
+ "step": 1440
1019
+ },
1020
+ {
1021
+ "epoch": 0.022618081986647533,
1022
+ "grad_norm": 5.366154670715332,
1023
+ "learning_rate": 0.00019849316777937232,
1024
+ "loss": 0.4288,
1025
+ "step": 1450
1026
+ },
1027
+ {
1028
+ "epoch": 0.02277406875896924,
1029
+ "grad_norm": 3.945873498916626,
1030
+ "learning_rate": 0.00019848276866121754,
1031
+ "loss": 0.5625,
1032
+ "step": 1460
1033
+ },
1034
+ {
1035
+ "epoch": 0.022930055531290947,
1036
+ "grad_norm": 1.8938435316085815,
1037
+ "learning_rate": 0.00019847236954306276,
1038
+ "loss": 0.4298,
1039
+ "step": 1470
1040
+ },
1041
+ {
1042
+ "epoch": 0.023086042303612653,
1043
+ "grad_norm": 7.040036678314209,
1044
+ "learning_rate": 0.00019846197042490798,
1045
+ "loss": 0.5806,
1046
+ "step": 1480
1047
+ },
1048
+ {
1049
+ "epoch": 0.02324202907593436,
1050
+ "grad_norm": 12.009136199951172,
1051
+ "learning_rate": 0.0001984515713067532,
1052
+ "loss": 0.3603,
1053
+ "step": 1490
1054
+ },
1055
+ {
1056
+ "epoch": 0.023398015848256067,
1057
+ "grad_norm": 3.157111883163452,
1058
+ "learning_rate": 0.00019844117218859841,
1059
+ "loss": 0.3811,
1060
+ "step": 1500
1061
+ },
1062
+ {
1063
+ "epoch": 0.023554002620577774,
1064
+ "grad_norm": 1.0227491855621338,
1065
+ "learning_rate": 0.00019843077307044363,
1066
+ "loss": 0.3459,
1067
+ "step": 1510
1068
+ },
1069
+ {
1070
+ "epoch": 0.02370998939289948,
1071
+ "grad_norm": 2.699263334274292,
1072
+ "learning_rate": 0.00019842037395228885,
1073
+ "loss": 0.4047,
1074
+ "step": 1520
1075
+ },
1076
+ {
1077
+ "epoch": 0.023865976165221188,
1078
+ "grad_norm": 3.879258871078491,
1079
+ "learning_rate": 0.00019840997483413407,
1080
+ "loss": 0.2083,
1081
+ "step": 1530
1082
+ },
1083
+ {
1084
+ "epoch": 0.024021962937542895,
1085
+ "grad_norm": 1.0468460321426392,
1086
+ "learning_rate": 0.0001983995757159793,
1087
+ "loss": 0.3146,
1088
+ "step": 1540
1089
+ },
1090
+ {
1091
+ "epoch": 0.024177949709864602,
1092
+ "grad_norm": 3.5738143920898438,
1093
+ "learning_rate": 0.0001983891765978245,
1094
+ "loss": 0.3181,
1095
+ "step": 1550
1096
+ },
1097
+ {
1098
+ "epoch": 0.024333936482186312,
1099
+ "grad_norm": 6.3198113441467285,
1100
+ "learning_rate": 0.00019837877747966973,
1101
+ "loss": 0.3196,
1102
+ "step": 1560
1103
+ },
1104
+ {
1105
+ "epoch": 0.02448992325450802,
1106
+ "grad_norm": 1.6974947452545166,
1107
+ "learning_rate": 0.00019836837836151495,
1108
+ "loss": 0.3832,
1109
+ "step": 1570
1110
+ },
1111
+ {
1112
+ "epoch": 0.024645910026829726,
1113
+ "grad_norm": 6.220627307891846,
1114
+ "learning_rate": 0.00019835797924336017,
1115
+ "loss": 0.3445,
1116
+ "step": 1580
1117
+ },
1118
+ {
1119
+ "epoch": 0.024801896799151433,
1120
+ "grad_norm": 3.9202497005462646,
1121
+ "learning_rate": 0.0001983475801252054,
1122
+ "loss": 0.4304,
1123
+ "step": 1590
1124
+ },
1125
+ {
1126
+ "epoch": 0.02495788357147314,
1127
+ "grad_norm": 2.2945311069488525,
1128
+ "learning_rate": 0.0001983371810070506,
1129
+ "loss": 0.4487,
1130
+ "step": 1600
1131
+ },
1132
+ {
1133
+ "epoch": 0.025113870343794847,
1134
+ "grad_norm": 9.262955665588379,
1135
+ "learning_rate": 0.00019832678188889583,
1136
+ "loss": 0.3691,
1137
+ "step": 1610
1138
+ },
1139
+ {
1140
+ "epoch": 0.025269857116116554,
1141
+ "grad_norm": 2.6905221939086914,
1142
+ "learning_rate": 0.00019831638277074105,
1143
+ "loss": 0.5117,
1144
+ "step": 1620
1145
+ },
1146
+ {
1147
+ "epoch": 0.02542584388843826,
1148
+ "grad_norm": 4.1330060958862305,
1149
+ "learning_rate": 0.00019830598365258627,
1150
+ "loss": 0.5721,
1151
+ "step": 1630
1152
+ },
1153
+ {
1154
+ "epoch": 0.025581830660759967,
1155
+ "grad_norm": 1.0681806802749634,
1156
+ "learning_rate": 0.0001982955845344315,
1157
+ "loss": 0.4794,
1158
+ "step": 1640
1159
+ },
1160
+ {
1161
+ "epoch": 0.025737817433081674,
1162
+ "grad_norm": 2.6171841621398926,
1163
+ "learning_rate": 0.0001982851854162767,
1164
+ "loss": 0.4596,
1165
+ "step": 1650
1166
+ },
1167
+ {
1168
+ "epoch": 0.02589380420540338,
1169
+ "grad_norm": 3.3451762199401855,
1170
+ "learning_rate": 0.00019827478629812193,
1171
+ "loss": 0.4745,
1172
+ "step": 1660
1173
+ },
1174
+ {
1175
+ "epoch": 0.026049790977725088,
1176
+ "grad_norm": 4.152349472045898,
1177
+ "learning_rate": 0.00019826438717996715,
1178
+ "loss": 0.4452,
1179
+ "step": 1670
1180
+ },
1181
+ {
1182
+ "epoch": 0.026205777750046795,
1183
+ "grad_norm": 2.7109954357147217,
1184
+ "learning_rate": 0.00019825398806181237,
1185
+ "loss": 0.3966,
1186
+ "step": 1680
1187
+ },
1188
+ {
1189
+ "epoch": 0.026361764522368502,
1190
+ "grad_norm": 5.731235980987549,
1191
+ "learning_rate": 0.0001982435889436576,
1192
+ "loss": 0.4423,
1193
+ "step": 1690
1194
+ },
1195
+ {
1196
+ "epoch": 0.02651775129469021,
1197
+ "grad_norm": 2.370852470397949,
1198
+ "learning_rate": 0.0001982331898255028,
1199
+ "loss": 0.5619,
1200
+ "step": 1700
1201
+ },
1202
+ {
1203
+ "epoch": 0.026673738067011916,
1204
+ "grad_norm": 2.5500876903533936,
1205
+ "learning_rate": 0.00019822279070734803,
1206
+ "loss": 0.2658,
1207
+ "step": 1710
1208
+ },
1209
+ {
1210
+ "epoch": 0.026829724839333626,
1211
+ "grad_norm": 0.20157204568386078,
1212
+ "learning_rate": 0.00019821239158919325,
1213
+ "loss": 0.2939,
1214
+ "step": 1720
1215
+ },
1216
+ {
1217
+ "epoch": 0.026985711611655333,
1218
+ "grad_norm": 2.6305084228515625,
1219
+ "learning_rate": 0.00019820199247103847,
1220
+ "loss": 0.5339,
1221
+ "step": 1730
1222
+ },
1223
+ {
1224
+ "epoch": 0.02714169838397704,
1225
+ "grad_norm": 6.241075038909912,
1226
+ "learning_rate": 0.00019819159335288369,
1227
+ "loss": 0.3625,
1228
+ "step": 1740
1229
+ },
1230
+ {
1231
+ "epoch": 0.027297685156298747,
1232
+ "grad_norm": 5.584851264953613,
1233
+ "learning_rate": 0.0001981811942347289,
1234
+ "loss": 0.6158,
1235
+ "step": 1750
1236
+ },
1237
+ {
1238
+ "epoch": 0.027453671928620454,
1239
+ "grad_norm": 6.756530284881592,
1240
+ "learning_rate": 0.00019817079511657413,
1241
+ "loss": 0.3513,
1242
+ "step": 1760
1243
+ },
1244
+ {
1245
+ "epoch": 0.02760965870094216,
1246
+ "grad_norm": 5.153111457824707,
1247
+ "learning_rate": 0.00019816039599841935,
1248
+ "loss": 0.2885,
1249
+ "step": 1770
1250
+ },
1251
+ {
1252
+ "epoch": 0.027765645473263868,
1253
+ "grad_norm": 6.222873210906982,
1254
+ "learning_rate": 0.00019814999688026456,
1255
+ "loss": 0.3291,
1256
+ "step": 1780
1257
+ },
1258
+ {
1259
+ "epoch": 0.027921632245585574,
1260
+ "grad_norm": 4.744391918182373,
1261
+ "learning_rate": 0.00019813959776210978,
1262
+ "loss": 0.5053,
1263
+ "step": 1790
1264
+ },
1265
+ {
1266
+ "epoch": 0.02807761901790728,
1267
+ "grad_norm": 3.7181131839752197,
1268
+ "learning_rate": 0.000198129198643955,
1269
+ "loss": 0.3983,
1270
+ "step": 1800
1271
+ },
1272
+ {
1273
+ "epoch": 0.02823360579022899,
1274
+ "grad_norm": 1.8923834562301636,
1275
+ "learning_rate": 0.00019811879952580022,
1276
+ "loss": 0.4465,
1277
+ "step": 1810
1278
+ },
1279
+ {
1280
+ "epoch": 0.028389592562550695,
1281
+ "grad_norm": 6.049956321716309,
1282
+ "learning_rate": 0.00019810840040764544,
1283
+ "loss": 0.5028,
1284
+ "step": 1820
1285
+ },
1286
+ {
1287
+ "epoch": 0.028545579334872402,
1288
+ "grad_norm": 8.034428596496582,
1289
+ "learning_rate": 0.00019809800128949066,
1290
+ "loss": 0.4799,
1291
+ "step": 1830
1292
+ },
1293
+ {
1294
+ "epoch": 0.02870156610719411,
1295
+ "grad_norm": 3.916890859603882,
1296
+ "learning_rate": 0.00019808760217133588,
1297
+ "loss": 0.3109,
1298
+ "step": 1840
1299
+ },
1300
+ {
1301
+ "epoch": 0.028857552879515816,
1302
+ "grad_norm": 3.351285457611084,
1303
+ "learning_rate": 0.0001980772030531811,
1304
+ "loss": 0.5981,
1305
+ "step": 1850
1306
+ },
1307
+ {
1308
+ "epoch": 0.029013539651837523,
1309
+ "grad_norm": 4.443714141845703,
1310
+ "learning_rate": 0.00019806680393502632,
1311
+ "loss": 0.3243,
1312
+ "step": 1860
1313
+ },
1314
+ {
1315
+ "epoch": 0.02916952642415923,
1316
+ "grad_norm": 0.6998550295829773,
1317
+ "learning_rate": 0.00019805640481687154,
1318
+ "loss": 0.2057,
1319
+ "step": 1870
1320
+ },
1321
+ {
1322
+ "epoch": 0.02932551319648094,
1323
+ "grad_norm": 12.252562522888184,
1324
+ "learning_rate": 0.00019804600569871676,
1325
+ "loss": 0.3806,
1326
+ "step": 1880
1327
+ },
1328
+ {
1329
+ "epoch": 0.029481499968802647,
1330
+ "grad_norm": 9.03868293762207,
1331
+ "learning_rate": 0.00019803560658056198,
1332
+ "loss": 0.3351,
1333
+ "step": 1890
1334
+ },
1335
+ {
1336
+ "epoch": 0.029637486741124354,
1337
+ "grad_norm": 2.3400044441223145,
1338
+ "learning_rate": 0.0001980252074624072,
1339
+ "loss": 0.3103,
1340
+ "step": 1900
1341
+ },
1342
+ {
1343
+ "epoch": 0.02979347351344606,
1344
+ "grad_norm": 2.399489164352417,
1345
+ "learning_rate": 0.00019801480834425242,
1346
+ "loss": 0.4792,
1347
+ "step": 1910
1348
+ },
1349
+ {
1350
+ "epoch": 0.029949460285767768,
1351
+ "grad_norm": 1.5152381658554077,
1352
+ "learning_rate": 0.00019800440922609764,
1353
+ "loss": 0.2873,
1354
+ "step": 1920
1355
+ },
1356
+ {
1357
+ "epoch": 0.030105447058089475,
1358
+ "grad_norm": 2.7013959884643555,
1359
+ "learning_rate": 0.00019799401010794286,
1360
+ "loss": 0.381,
1361
+ "step": 1930
1362
+ },
1363
+ {
1364
+ "epoch": 0.03026143383041118,
1365
+ "grad_norm": 0.7135261297225952,
1366
+ "learning_rate": 0.00019798361098978808,
1367
+ "loss": 0.4001,
1368
+ "step": 1940
1369
+ },
1370
+ {
1371
+ "epoch": 0.03041742060273289,
1372
+ "grad_norm": 15.61109447479248,
1373
+ "learning_rate": 0.0001979732118716333,
1374
+ "loss": 0.474,
1375
+ "step": 1950
1376
+ },
1377
+ {
1378
+ "epoch": 0.030573407375054595,
1379
+ "grad_norm": 3.236668348312378,
1380
+ "learning_rate": 0.00019796281275347852,
1381
+ "loss": 0.2149,
1382
+ "step": 1960
1383
+ },
1384
+ {
1385
+ "epoch": 0.030729394147376302,
1386
+ "grad_norm": 4.722929954528809,
1387
+ "learning_rate": 0.00019795241363532374,
1388
+ "loss": 0.5474,
1389
+ "step": 1970
1390
+ },
1391
+ {
1392
+ "epoch": 0.03088538091969801,
1393
+ "grad_norm": 6.002335548400879,
1394
+ "learning_rate": 0.00019794201451716896,
1395
+ "loss": 0.5196,
1396
+ "step": 1980
1397
+ },
1398
+ {
1399
+ "epoch": 0.031041367692019716,
1400
+ "grad_norm": 5.210022926330566,
1401
+ "learning_rate": 0.00019793161539901418,
1402
+ "loss": 0.4741,
1403
+ "step": 1990
1404
+ },
1405
+ {
1406
+ "epoch": 0.031197354464341423,
1407
+ "grad_norm": 7.984034538269043,
1408
+ "learning_rate": 0.0001979212162808594,
1409
+ "loss": 0.5399,
1410
+ "step": 2000
1411
+ },
1412
+ {
1413
+ "epoch": 0.03135334123666313,
1414
+ "grad_norm": 4.008991241455078,
1415
+ "learning_rate": 0.00019791081716270462,
1416
+ "loss": 0.4788,
1417
+ "step": 2010
1418
+ },
1419
+ {
1420
+ "epoch": 0.03150932800898484,
1421
+ "grad_norm": 4.395316123962402,
1422
+ "learning_rate": 0.00019790041804454984,
1423
+ "loss": 0.3456,
1424
+ "step": 2020
1425
+ },
1426
+ {
1427
+ "epoch": 0.03166531478130655,
1428
+ "grad_norm": 3.9784998893737793,
1429
+ "learning_rate": 0.00019789001892639506,
1430
+ "loss": 0.326,
1431
+ "step": 2030
1432
+ },
1433
+ {
1434
+ "epoch": 0.03182130155362825,
1435
+ "grad_norm": 5.040256023406982,
1436
+ "learning_rate": 0.00019787961980824028,
1437
+ "loss": 0.5264,
1438
+ "step": 2040
1439
+ },
1440
+ {
1441
+ "epoch": 0.03197728832594996,
1442
+ "grad_norm": 2.686203956604004,
1443
+ "learning_rate": 0.0001978692206900855,
1444
+ "loss": 0.5461,
1445
+ "step": 2050
1446
+ },
1447
+ {
1448
+ "epoch": 0.032133275098271664,
1449
+ "grad_norm": 1.8417892456054688,
1450
+ "learning_rate": 0.00019785882157193071,
1451
+ "loss": 0.4736,
1452
+ "step": 2060
1453
+ },
1454
+ {
1455
+ "epoch": 0.032289261870593375,
1456
+ "grad_norm": 3.623887300491333,
1457
+ "learning_rate": 0.00019784842245377593,
1458
+ "loss": 0.3742,
1459
+ "step": 2070
1460
+ },
1461
+ {
1462
+ "epoch": 0.03244524864291508,
1463
+ "grad_norm": 0.5184723734855652,
1464
+ "learning_rate": 0.00019783802333562115,
1465
+ "loss": 0.3941,
1466
+ "step": 2080
1467
+ },
1468
+ {
1469
+ "epoch": 0.03260123541523679,
1470
+ "grad_norm": 7.751046180725098,
1471
+ "learning_rate": 0.00019782762421746637,
1472
+ "loss": 0.4609,
1473
+ "step": 2090
1474
+ },
1475
+ {
1476
+ "epoch": 0.03275722218755849,
1477
+ "grad_norm": 2.547175168991089,
1478
+ "learning_rate": 0.0001978172250993116,
1479
+ "loss": 0.3089,
1480
+ "step": 2100
1481
+ },
1482
+ {
1483
+ "epoch": 0.0329132089598802,
1484
+ "grad_norm": 1.8453168869018555,
1485
+ "learning_rate": 0.0001978068259811568,
1486
+ "loss": 0.4371,
1487
+ "step": 2110
1488
+ },
1489
+ {
1490
+ "epoch": 0.033069195732201906,
1491
+ "grad_norm": 7.033441543579102,
1492
+ "learning_rate": 0.00019779642686300203,
1493
+ "loss": 0.3842,
1494
+ "step": 2120
1495
+ },
1496
+ {
1497
+ "epoch": 0.033225182504523616,
1498
+ "grad_norm": 2.2642929553985596,
1499
+ "learning_rate": 0.00019778602774484725,
1500
+ "loss": 0.2069,
1501
+ "step": 2130
1502
+ },
1503
+ {
1504
+ "epoch": 0.03338116927684533,
1505
+ "grad_norm": 4.043938636779785,
1506
+ "learning_rate": 0.00019777562862669244,
1507
+ "loss": 0.4148,
1508
+ "step": 2140
1509
+ },
1510
+ {
1511
+ "epoch": 0.03353715604916703,
1512
+ "grad_norm": 4.227573871612549,
1513
+ "learning_rate": 0.0001977652295085377,
1514
+ "loss": 0.4785,
1515
+ "step": 2150
1516
+ },
1517
+ {
1518
+ "epoch": 0.03369314282148874,
1519
+ "grad_norm": 2.0808868408203125,
1520
+ "learning_rate": 0.00019775483039038288,
1521
+ "loss": 0.3073,
1522
+ "step": 2160
1523
+ },
1524
+ {
1525
+ "epoch": 0.033849129593810444,
1526
+ "grad_norm": 4.518038749694824,
1527
+ "learning_rate": 0.00019774443127222813,
1528
+ "loss": 0.4197,
1529
+ "step": 2170
1530
+ },
1531
+ {
1532
+ "epoch": 0.034005116366132154,
1533
+ "grad_norm": 2.3731303215026855,
1534
+ "learning_rate": 0.00019773403215407332,
1535
+ "loss": 0.2753,
1536
+ "step": 2180
1537
+ },
1538
+ {
1539
+ "epoch": 0.03416110313845386,
1540
+ "grad_norm": 1.7613986730575562,
1541
+ "learning_rate": 0.00019772363303591857,
1542
+ "loss": 0.3834,
1543
+ "step": 2190
1544
+ },
1545
+ {
1546
+ "epoch": 0.03431708991077557,
1547
+ "grad_norm": 2.9518485069274902,
1548
+ "learning_rate": 0.00019771323391776376,
1549
+ "loss": 0.3544,
1550
+ "step": 2200
1551
+ },
1552
+ {
1553
+ "epoch": 0.03447307668309727,
1554
+ "grad_norm": 4.424034118652344,
1555
+ "learning_rate": 0.000197702834799609,
1556
+ "loss": 0.4046,
1557
+ "step": 2210
1558
+ },
1559
+ {
1560
+ "epoch": 0.03462906345541898,
1561
+ "grad_norm": 2.307690382003784,
1562
+ "learning_rate": 0.0001976924356814542,
1563
+ "loss": 0.1445,
1564
+ "step": 2220
1565
+ },
1566
+ {
1567
+ "epoch": 0.034785050227740685,
1568
+ "grad_norm": 2.536533832550049,
1569
+ "learning_rate": 0.00019768203656329945,
1570
+ "loss": 0.2815,
1571
+ "step": 2230
1572
+ },
1573
+ {
1574
+ "epoch": 0.034941037000062396,
1575
+ "grad_norm": 2.006728172302246,
1576
+ "learning_rate": 0.00019767163744514464,
1577
+ "loss": 0.5314,
1578
+ "step": 2240
1579
+ },
1580
+ {
1581
+ "epoch": 0.0350970237723841,
1582
+ "grad_norm": 4.729264259338379,
1583
+ "learning_rate": 0.0001976612383269899,
1584
+ "loss": 0.4547,
1585
+ "step": 2250
1586
+ },
1587
+ {
1588
+ "epoch": 0.03525301054470581,
1589
+ "grad_norm": 4.600561141967773,
1590
+ "learning_rate": 0.00019765083920883508,
1591
+ "loss": 0.3967,
1592
+ "step": 2260
1593
+ },
1594
+ {
1595
+ "epoch": 0.03540899731702751,
1596
+ "grad_norm": 0.6419815421104431,
1597
+ "learning_rate": 0.00019764044009068033,
1598
+ "loss": 0.336,
1599
+ "step": 2270
1600
+ },
1601
+ {
1602
+ "epoch": 0.03556498408934922,
1603
+ "grad_norm": 5.567695140838623,
1604
+ "learning_rate": 0.00019763004097252552,
1605
+ "loss": 0.5644,
1606
+ "step": 2280
1607
+ },
1608
+ {
1609
+ "epoch": 0.035720970861670934,
1610
+ "grad_norm": 0.5767773389816284,
1611
+ "learning_rate": 0.00019761964185437077,
1612
+ "loss": 0.2643,
1613
+ "step": 2290
1614
+ },
1615
+ {
1616
+ "epoch": 0.03587695763399264,
1617
+ "grad_norm": 4.91269063949585,
1618
+ "learning_rate": 0.00019760924273621596,
1619
+ "loss": 0.4198,
1620
+ "step": 2300
1621
+ },
1622
+ {
1623
+ "epoch": 0.03603294440631435,
1624
+ "grad_norm": 4.422239780426025,
1625
+ "learning_rate": 0.0001975988436180612,
1626
+ "loss": 0.5338,
1627
+ "step": 2310
1628
+ },
1629
+ {
1630
+ "epoch": 0.03618893117863605,
1631
+ "grad_norm": 1.672713041305542,
1632
+ "learning_rate": 0.0001975884444999064,
1633
+ "loss": 0.3214,
1634
+ "step": 2320
1635
+ },
1636
+ {
1637
+ "epoch": 0.03634491795095776,
1638
+ "grad_norm": 4.033807277679443,
1639
+ "learning_rate": 0.00019757804538175164,
1640
+ "loss": 0.3593,
1641
+ "step": 2330
1642
+ },
1643
+ {
1644
+ "epoch": 0.036500904723279465,
1645
+ "grad_norm": 3.198880910873413,
1646
+ "learning_rate": 0.00019756764626359684,
1647
+ "loss": 0.3657,
1648
+ "step": 2340
1649
+ },
1650
+ {
1651
+ "epoch": 0.036656891495601175,
1652
+ "grad_norm": 2.640443801879883,
1653
+ "learning_rate": 0.00019755724714544208,
1654
+ "loss": 0.3243,
1655
+ "step": 2350
1656
+ },
1657
+ {
1658
+ "epoch": 0.03681287826792288,
1659
+ "grad_norm": 1.562350869178772,
1660
+ "learning_rate": 0.00019754684802728728,
1661
+ "loss": 0.3362,
1662
+ "step": 2360
1663
+ },
1664
+ {
1665
+ "epoch": 0.03696886504024459,
1666
+ "grad_norm": 4.379998683929443,
1667
+ "learning_rate": 0.00019753644890913252,
1668
+ "loss": 0.541,
1669
+ "step": 2370
1670
+ },
1671
+ {
1672
+ "epoch": 0.03712485181256629,
1673
+ "grad_norm": 2.529301881790161,
1674
+ "learning_rate": 0.00019752604979097774,
1675
+ "loss": 0.5396,
1676
+ "step": 2380
1677
+ },
1678
+ {
1679
+ "epoch": 0.037280838584888,
1680
+ "grad_norm": 1.7864131927490234,
1681
+ "learning_rate": 0.00019751565067282296,
1682
+ "loss": 0.3165,
1683
+ "step": 2390
1684
+ },
1685
+ {
1686
+ "epoch": 0.037436825357209706,
1687
+ "grad_norm": 1.0566092729568481,
1688
+ "learning_rate": 0.00019750525155466818,
1689
+ "loss": 0.4225,
1690
+ "step": 2400
1691
+ },
1692
+ {
1693
+ "epoch": 0.03759281212953142,
1694
+ "grad_norm": 2.4530510902404785,
1695
+ "learning_rate": 0.0001974948524365134,
1696
+ "loss": 0.4281,
1697
+ "step": 2410
1698
+ },
1699
+ {
1700
+ "epoch": 0.03774879890185312,
1701
+ "grad_norm": 5.363385200500488,
1702
+ "learning_rate": 0.00019748445331835862,
1703
+ "loss": 0.4746,
1704
+ "step": 2420
1705
+ },
1706
+ {
1707
+ "epoch": 0.03790478567417483,
1708
+ "grad_norm": 6.038578510284424,
1709
+ "learning_rate": 0.00019747405420020384,
1710
+ "loss": 0.3182,
1711
+ "step": 2430
1712
+ },
1713
+ {
1714
+ "epoch": 0.038060772446496534,
1715
+ "grad_norm": 9.37709903717041,
1716
+ "learning_rate": 0.00019746365508204906,
1717
+ "loss": 0.3288,
1718
+ "step": 2440
1719
+ },
1720
+ {
1721
+ "epoch": 0.038216759218818244,
1722
+ "grad_norm": 28.81134033203125,
1723
+ "learning_rate": 0.00019745325596389428,
1724
+ "loss": 0.2646,
1725
+ "step": 2450
1726
+ },
1727
+ {
1728
+ "epoch": 0.038372745991139955,
1729
+ "grad_norm": 0.616172730922699,
1730
+ "learning_rate": 0.0001974428568457395,
1731
+ "loss": 0.3848,
1732
+ "step": 2460
1733
+ },
1734
+ {
1735
+ "epoch": 0.03852873276346166,
1736
+ "grad_norm": 2.9943106174468994,
1737
+ "learning_rate": 0.00019743245772758472,
1738
+ "loss": 0.4097,
1739
+ "step": 2470
1740
+ },
1741
+ {
1742
+ "epoch": 0.03868471953578337,
1743
+ "grad_norm": 4.001022815704346,
1744
+ "learning_rate": 0.00019742205860942994,
1745
+ "loss": 0.2359,
1746
+ "step": 2480
1747
+ },
1748
+ {
1749
+ "epoch": 0.03884070630810507,
1750
+ "grad_norm": 3.587982416152954,
1751
+ "learning_rate": 0.00019741165949127516,
1752
+ "loss": 0.5829,
1753
+ "step": 2490
1754
+ },
1755
+ {
1756
+ "epoch": 0.03899669308042678,
1757
+ "grad_norm": 1.906042218208313,
1758
+ "learning_rate": 0.00019740126037312038,
1759
+ "loss": 0.336,
1760
+ "step": 2500
1761
+ }
1762
+ ],
1763
+ "logging_steps": 10,
1764
+ "max_steps": 192324,
1765
+ "num_input_tokens_seen": 0,
1766
+ "num_train_epochs": 3,
1767
+ "save_steps": 500,
1768
+ "stateful_callbacks": {
1769
+ "TrainerControl": {
1770
+ "args": {
1771
+ "should_epoch_stop": false,
1772
+ "should_evaluate": false,
1773
+ "should_log": false,
1774
+ "should_save": true,
1775
+ "should_training_stop": false
1776
+ },
1777
+ "attributes": {}
1778
+ }
1779
+ },
1780
+ "total_flos": 1522394726400000.0,
1781
+ "train_batch_size": 1,
1782
+ "trial_name": null,
1783
+ "trial_params": null
1784
+ }
HVU_QA/t5-viet-qg-finetuned/checkpoint-2500/training_args.bin ADDED
Binary file (5.71 kB). View file
 
HVU_QA/t5-viet-qg-finetuned/checkpoint-3000/added_tokens.json ADDED
@@ -0,0 +1,98 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "<extra_id_0>": 36095,
3
+ "<extra_id_10>": 36085,
4
+ "<extra_id_11>": 36084,
5
+ "<extra_id_12>": 36083,
6
+ "<extra_id_13>": 36082,
7
+ "<extra_id_14>": 36081,
8
+ "<extra_id_15>": 36080,
9
+ "<extra_id_16>": 36079,
10
+ "<extra_id_17>": 36078,
11
+ "<extra_id_18>": 36077,
12
+ "<extra_id_19>": 36076,
13
+ "<extra_id_1>": 36094,
14
+ "<extra_id_20>": 36075,
15
+ "<extra_id_21>": 36074,
16
+ "<extra_id_22>": 36073,
17
+ "<extra_id_23>": 36072,
18
+ "<extra_id_24>": 36071,
19
+ "<extra_id_25>": 36070,
20
+ "<extra_id_26>": 36069,
21
+ "<extra_id_27>": 36068,
22
+ "<extra_id_28>": 36067,
23
+ "<extra_id_29>": 36066,
24
+ "<extra_id_2>": 36093,
25
+ "<extra_id_30>": 36065,
26
+ "<extra_id_31>": 36064,
27
+ "<extra_id_32>": 36063,
28
+ "<extra_id_33>": 36062,
29
+ "<extra_id_34>": 36061,
30
+ "<extra_id_35>": 36060,
31
+ "<extra_id_36>": 36059,
32
+ "<extra_id_37>": 36058,
33
+ "<extra_id_38>": 36057,
34
+ "<extra_id_39>": 36056,
35
+ "<extra_id_3>": 36092,
36
+ "<extra_id_40>": 36055,
37
+ "<extra_id_41>": 36054,
38
+ "<extra_id_42>": 36053,
39
+ "<extra_id_43>": 36052,
40
+ "<extra_id_44>": 36051,
41
+ "<extra_id_45>": 36050,
42
+ "<extra_id_46>": 36049,
43
+ "<extra_id_47>": 36048,
44
+ "<extra_id_48>": 36047,
45
+ "<extra_id_49>": 36046,
46
+ "<extra_id_4>": 36091,
47
+ "<extra_id_50>": 36045,
48
+ "<extra_id_51>": 36044,
49
+ "<extra_id_52>": 36043,
50
+ "<extra_id_53>": 36042,
51
+ "<extra_id_54>": 36041,
52
+ "<extra_id_55>": 36040,
53
+ "<extra_id_56>": 36039,
54
+ "<extra_id_57>": 36038,
55
+ "<extra_id_58>": 36037,
56
+ "<extra_id_59>": 36036,
57
+ "<extra_id_5>": 36090,
58
+ "<extra_id_60>": 36035,
59
+ "<extra_id_61>": 36034,
60
+ "<extra_id_62>": 36033,
61
+ "<extra_id_63>": 36032,
62
+ "<extra_id_64>": 36031,
63
+ "<extra_id_65>": 36030,
64
+ "<extra_id_66>": 36029,
65
+ "<extra_id_67>": 36028,
66
+ "<extra_id_68>": 36027,
67
+ "<extra_id_69>": 36026,
68
+ "<extra_id_6>": 36089,
69
+ "<extra_id_70>": 36025,
70
+ "<extra_id_71>": 36024,
71
+ "<extra_id_72>": 36023,
72
+ "<extra_id_73>": 36022,
73
+ "<extra_id_74>": 36021,
74
+ "<extra_id_75>": 36020,
75
+ "<extra_id_76>": 36019,
76
+ "<extra_id_77>": 36018,
77
+ "<extra_id_78>": 36017,
78
+ "<extra_id_79>": 36016,
79
+ "<extra_id_7>": 36088,
80
+ "<extra_id_80>": 36015,
81
+ "<extra_id_81>": 36014,
82
+ "<extra_id_82>": 36013,
83
+ "<extra_id_83>": 36012,
84
+ "<extra_id_84>": 36011,
85
+ "<extra_id_85>": 36010,
86
+ "<extra_id_86>": 36009,
87
+ "<extra_id_87>": 36008,
88
+ "<extra_id_88>": 36007,
89
+ "<extra_id_89>": 36006,
90
+ "<extra_id_8>": 36087,
91
+ "<extra_id_90>": 36005,
92
+ "<extra_id_91>": 36004,
93
+ "<extra_id_92>": 36003,
94
+ "<extra_id_93>": 36002,
95
+ "<extra_id_94>": 36001,
96
+ "<extra_id_95>": 36000,
97
+ "<extra_id_9>": 36086
98
+ }
HVU_QA/t5-viet-qg-finetuned/checkpoint-3000/config.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "T5ForConditionalGeneration"
4
+ ],
5
+ "classifier_dropout": 0.0,
6
+ "d_ff": 3072,
7
+ "d_kv": 64,
8
+ "d_model": 768,
9
+ "decoder_start_token_id": 0,
10
+ "dense_act_fn": "relu",
11
+ "dropout_rate": 0.1,
12
+ "eos_token_id": 1,
13
+ "feed_forward_proj": "relu",
14
+ "initializer_factor": 1.0,
15
+ "is_encoder_decoder": true,
16
+ "is_gated_act": false,
17
+ "layer_norm_epsilon": 1e-06,
18
+ "model_type": "t5",
19
+ "n_positions": 512,
20
+ "num_decoder_layers": 12,
21
+ "num_heads": 12,
22
+ "num_layers": 12,
23
+ "output_past": true,
24
+ "pad_token_id": 0,
25
+ "relative_attention_max_distance": 128,
26
+ "relative_attention_num_buckets": 32,
27
+ "torch_dtype": "float32",
28
+ "transformers_version": "4.53.2",
29
+ "use_cache": true,
30
+ "vocab_size": 36096
31
+ }
HVU_QA/t5-viet-qg-finetuned/checkpoint-3000/generation_config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "decoder_start_token_id": 0,
4
+ "eos_token_id": 1,
5
+ "pad_token_id": 0,
6
+ "transformers_version": "4.53.2"
7
+ }
HVU_QA/t5-viet-qg-finetuned/checkpoint-3000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0d5970a71e58b4a80272242ac42733b813a0fbee6d4d7ed752e7f50690215dfe
3
+ size 62652416
HVU_QA/t5-viet-qg-finetuned/checkpoint-3000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aa2041ed607e889082c58cbda54aa4994566ddab43586738a19ccc347c8b6020
3
+ size 62652416
HVU_QA/t5-viet-qg-finetuned/checkpoint-3000/rng_state.pth ADDED
Binary file (14.6 kB). View file
 
HVU_QA/t5-viet-qg-finetuned/checkpoint-3000/scheduler.pt ADDED
Binary file (1.47 kB). View file
 
HVU_QA/t5-viet-qg-finetuned/checkpoint-3000/special_tokens_map.json ADDED
@@ -0,0 +1,121 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<extra_id_0>",
4
+ "<extra_id_1>",
5
+ "<extra_id_2>",
6
+ "<extra_id_3>",
7
+ "<extra_id_4>",
8
+ "<extra_id_5>",
9
+ "<extra_id_6>",
10
+ "<extra_id_7>",
11
+ "<extra_id_8>",
12
+ "<extra_id_9>",
13
+ "<extra_id_10>",
14
+ "<extra_id_11>",
15
+ "<extra_id_12>",
16
+ "<extra_id_13>",
17
+ "<extra_id_14>",
18
+ "<extra_id_15>",
19
+ "<extra_id_16>",
20
+ "<extra_id_17>",
21
+ "<extra_id_18>",
22
+ "<extra_id_19>",
23
+ "<extra_id_20>",
24
+ "<extra_id_21>",
25
+ "<extra_id_22>",
26
+ "<extra_id_23>",
27
+ "<extra_id_24>",
28
+ "<extra_id_25>",
29
+ "<extra_id_26>",
30
+ "<extra_id_27>",
31
+ "<extra_id_28>",
32
+ "<extra_id_29>",
33
+ "<extra_id_30>",
34
+ "<extra_id_31>",
35
+ "<extra_id_32>",
36
+ "<extra_id_33>",
37
+ "<extra_id_34>",
38
+ "<extra_id_35>",
39
+ "<extra_id_36>",
40
+ "<extra_id_37>",
41
+ "<extra_id_38>",
42
+ "<extra_id_39>",
43
+ "<extra_id_40>",
44
+ "<extra_id_41>",
45
+ "<extra_id_42>",
46
+ "<extra_id_43>",
47
+ "<extra_id_44>",
48
+ "<extra_id_45>",
49
+ "<extra_id_46>",
50
+ "<extra_id_47>",
51
+ "<extra_id_48>",
52
+ "<extra_id_49>",
53
+ "<extra_id_50>",
54
+ "<extra_id_51>",
55
+ "<extra_id_52>",
56
+ "<extra_id_53>",
57
+ "<extra_id_54>",
58
+ "<extra_id_55>",
59
+ "<extra_id_56>",
60
+ "<extra_id_57>",
61
+ "<extra_id_58>",
62
+ "<extra_id_59>",
63
+ "<extra_id_60>",
64
+ "<extra_id_61>",
65
+ "<extra_id_62>",
66
+ "<extra_id_63>",
67
+ "<extra_id_64>",
68
+ "<extra_id_65>",
69
+ "<extra_id_66>",
70
+ "<extra_id_67>",
71
+ "<extra_id_68>",
72
+ "<extra_id_69>",
73
+ "<extra_id_70>",
74
+ "<extra_id_71>",
75
+ "<extra_id_72>",
76
+ "<extra_id_73>",
77
+ "<extra_id_74>",
78
+ "<extra_id_75>",
79
+ "<extra_id_76>",
80
+ "<extra_id_77>",
81
+ "<extra_id_78>",
82
+ "<extra_id_79>",
83
+ "<extra_id_80>",
84
+ "<extra_id_81>",
85
+ "<extra_id_82>",
86
+ "<extra_id_83>",
87
+ "<extra_id_84>",
88
+ "<extra_id_85>",
89
+ "<extra_id_86>",
90
+ "<extra_id_87>",
91
+ "<extra_id_88>",
92
+ "<extra_id_89>",
93
+ "<extra_id_90>",
94
+ "<extra_id_91>",
95
+ "<extra_id_92>",
96
+ "<extra_id_93>",
97
+ "<extra_id_94>",
98
+ "<extra_id_95>"
99
+ ],
100
+ "eos_token": {
101
+ "content": "</s>",
102
+ "lstrip": false,
103
+ "normalized": false,
104
+ "rstrip": false,
105
+ "single_word": false
106
+ },
107
+ "pad_token": {
108
+ "content": "<pad>",
109
+ "lstrip": false,
110
+ "normalized": false,
111
+ "rstrip": false,
112
+ "single_word": false
113
+ },
114
+ "unk_token": {
115
+ "content": "<unk>",
116
+ "lstrip": false,
117
+ "normalized": false,
118
+ "rstrip": false,
119
+ "single_word": false
120
+ }
121
+ }
HVU_QA/t5-viet-qg-finetuned/checkpoint-3000/spiece.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6
3
+ size 820370
HVU_QA/t5-viet-qg-finetuned/checkpoint-3000/tokenizer_config.json ADDED
@@ -0,0 +1,905 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": true,
3
+ "added_tokens_decoder": {
4
+ "0": {
5
+ "content": "<pad>",
6
+ "lstrip": false,
7
+ "normalized": false,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "1": {
13
+ "content": "</s>",
14
+ "lstrip": false,
15
+ "normalized": false,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
+ },
20
+ "2": {
21
+ "content": "<unk>",
22
+ "lstrip": false,
23
+ "normalized": false,
24
+ "rstrip": false,
25
+ "single_word": false,
26
+ "special": true
27
+ },
28
+ "36000": {
29
+ "content": "<extra_id_95>",
30
+ "lstrip": true,
31
+ "normalized": false,
32
+ "rstrip": true,
33
+ "single_word": false,
34
+ "special": true
35
+ },
36
+ "36001": {
37
+ "content": "<extra_id_94>",
38
+ "lstrip": true,
39
+ "normalized": false,
40
+ "rstrip": true,
41
+ "single_word": false,
42
+ "special": true
43
+ },
44
+ "36002": {
45
+ "content": "<extra_id_93>",
46
+ "lstrip": true,
47
+ "normalized": false,
48
+ "rstrip": true,
49
+ "single_word": false,
50
+ "special": true
51
+ },
52
+ "36003": {
53
+ "content": "<extra_id_92>",
54
+ "lstrip": true,
55
+ "normalized": false,
56
+ "rstrip": true,
57
+ "single_word": false,
58
+ "special": true
59
+ },
60
+ "36004": {
61
+ "content": "<extra_id_91>",
62
+ "lstrip": true,
63
+ "normalized": false,
64
+ "rstrip": true,
65
+ "single_word": false,
66
+ "special": true
67
+ },
68
+ "36005": {
69
+ "content": "<extra_id_90>",
70
+ "lstrip": true,
71
+ "normalized": false,
72
+ "rstrip": true,
73
+ "single_word": false,
74
+ "special": true
75
+ },
76
+ "36006": {
77
+ "content": "<extra_id_89>",
78
+ "lstrip": true,
79
+ "normalized": false,
80
+ "rstrip": true,
81
+ "single_word": false,
82
+ "special": true
83
+ },
84
+ "36007": {
85
+ "content": "<extra_id_88>",
86
+ "lstrip": true,
87
+ "normalized": false,
88
+ "rstrip": true,
89
+ "single_word": false,
90
+ "special": true
91
+ },
92
+ "36008": {
93
+ "content": "<extra_id_87>",
94
+ "lstrip": true,
95
+ "normalized": false,
96
+ "rstrip": true,
97
+ "single_word": false,
98
+ "special": true
99
+ },
100
+ "36009": {
101
+ "content": "<extra_id_86>",
102
+ "lstrip": true,
103
+ "normalized": false,
104
+ "rstrip": true,
105
+ "single_word": false,
106
+ "special": true
107
+ },
108
+ "36010": {
109
+ "content": "<extra_id_85>",
110
+ "lstrip": true,
111
+ "normalized": false,
112
+ "rstrip": true,
113
+ "single_word": false,
114
+ "special": true
115
+ },
116
+ "36011": {
117
+ "content": "<extra_id_84>",
118
+ "lstrip": true,
119
+ "normalized": false,
120
+ "rstrip": true,
121
+ "single_word": false,
122
+ "special": true
123
+ },
124
+ "36012": {
125
+ "content": "<extra_id_83>",
126
+ "lstrip": true,
127
+ "normalized": false,
128
+ "rstrip": true,
129
+ "single_word": false,
130
+ "special": true
131
+ },
132
+ "36013": {
133
+ "content": "<extra_id_82>",
134
+ "lstrip": true,
135
+ "normalized": false,
136
+ "rstrip": true,
137
+ "single_word": false,
138
+ "special": true
139
+ },
140
+ "36014": {
141
+ "content": "<extra_id_81>",
142
+ "lstrip": true,
143
+ "normalized": false,
144
+ "rstrip": true,
145
+ "single_word": false,
146
+ "special": true
147
+ },
148
+ "36015": {
149
+ "content": "<extra_id_80>",
150
+ "lstrip": true,
151
+ "normalized": false,
152
+ "rstrip": true,
153
+ "single_word": false,
154
+ "special": true
155
+ },
156
+ "36016": {
157
+ "content": "<extra_id_79>",
158
+ "lstrip": true,
159
+ "normalized": false,
160
+ "rstrip": true,
161
+ "single_word": false,
162
+ "special": true
163
+ },
164
+ "36017": {
165
+ "content": "<extra_id_78>",
166
+ "lstrip": true,
167
+ "normalized": false,
168
+ "rstrip": true,
169
+ "single_word": false,
170
+ "special": true
171
+ },
172
+ "36018": {
173
+ "content": "<extra_id_77>",
174
+ "lstrip": true,
175
+ "normalized": false,
176
+ "rstrip": true,
177
+ "single_word": false,
178
+ "special": true
179
+ },
180
+ "36019": {
181
+ "content": "<extra_id_76>",
182
+ "lstrip": true,
183
+ "normalized": false,
184
+ "rstrip": true,
185
+ "single_word": false,
186
+ "special": true
187
+ },
188
+ "36020": {
189
+ "content": "<extra_id_75>",
190
+ "lstrip": true,
191
+ "normalized": false,
192
+ "rstrip": true,
193
+ "single_word": false,
194
+ "special": true
195
+ },
196
+ "36021": {
197
+ "content": "<extra_id_74>",
198
+ "lstrip": true,
199
+ "normalized": false,
200
+ "rstrip": true,
201
+ "single_word": false,
202
+ "special": true
203
+ },
204
+ "36022": {
205
+ "content": "<extra_id_73>",
206
+ "lstrip": true,
207
+ "normalized": false,
208
+ "rstrip": true,
209
+ "single_word": false,
210
+ "special": true
211
+ },
212
+ "36023": {
213
+ "content": "<extra_id_72>",
214
+ "lstrip": true,
215
+ "normalized": false,
216
+ "rstrip": true,
217
+ "single_word": false,
218
+ "special": true
219
+ },
220
+ "36024": {
221
+ "content": "<extra_id_71>",
222
+ "lstrip": true,
223
+ "normalized": false,
224
+ "rstrip": true,
225
+ "single_word": false,
226
+ "special": true
227
+ },
228
+ "36025": {
229
+ "content": "<extra_id_70>",
230
+ "lstrip": true,
231
+ "normalized": false,
232
+ "rstrip": true,
233
+ "single_word": false,
234
+ "special": true
235
+ },
236
+ "36026": {
237
+ "content": "<extra_id_69>",
238
+ "lstrip": true,
239
+ "normalized": false,
240
+ "rstrip": true,
241
+ "single_word": false,
242
+ "special": true
243
+ },
244
+ "36027": {
245
+ "content": "<extra_id_68>",
246
+ "lstrip": true,
247
+ "normalized": false,
248
+ "rstrip": true,
249
+ "single_word": false,
250
+ "special": true
251
+ },
252
+ "36028": {
253
+ "content": "<extra_id_67>",
254
+ "lstrip": true,
255
+ "normalized": false,
256
+ "rstrip": true,
257
+ "single_word": false,
258
+ "special": true
259
+ },
260
+ "36029": {
261
+ "content": "<extra_id_66>",
262
+ "lstrip": true,
263
+ "normalized": false,
264
+ "rstrip": true,
265
+ "single_word": false,
266
+ "special": true
267
+ },
268
+ "36030": {
269
+ "content": "<extra_id_65>",
270
+ "lstrip": true,
271
+ "normalized": false,
272
+ "rstrip": true,
273
+ "single_word": false,
274
+ "special": true
275
+ },
276
+ "36031": {
277
+ "content": "<extra_id_64>",
278
+ "lstrip": true,
279
+ "normalized": false,
280
+ "rstrip": true,
281
+ "single_word": false,
282
+ "special": true
283
+ },
284
+ "36032": {
285
+ "content": "<extra_id_63>",
286
+ "lstrip": true,
287
+ "normalized": false,
288
+ "rstrip": true,
289
+ "single_word": false,
290
+ "special": true
291
+ },
292
+ "36033": {
293
+ "content": "<extra_id_62>",
294
+ "lstrip": true,
295
+ "normalized": false,
296
+ "rstrip": true,
297
+ "single_word": false,
298
+ "special": true
299
+ },
300
+ "36034": {
301
+ "content": "<extra_id_61>",
302
+ "lstrip": true,
303
+ "normalized": false,
304
+ "rstrip": true,
305
+ "single_word": false,
306
+ "special": true
307
+ },
308
+ "36035": {
309
+ "content": "<extra_id_60>",
310
+ "lstrip": true,
311
+ "normalized": false,
312
+ "rstrip": true,
313
+ "single_word": false,
314
+ "special": true
315
+ },
316
+ "36036": {
317
+ "content": "<extra_id_59>",
318
+ "lstrip": true,
319
+ "normalized": false,
320
+ "rstrip": true,
321
+ "single_word": false,
322
+ "special": true
323
+ },
324
+ "36037": {
325
+ "content": "<extra_id_58>",
326
+ "lstrip": true,
327
+ "normalized": false,
328
+ "rstrip": true,
329
+ "single_word": false,
330
+ "special": true
331
+ },
332
+ "36038": {
333
+ "content": "<extra_id_57>",
334
+ "lstrip": true,
335
+ "normalized": false,
336
+ "rstrip": true,
337
+ "single_word": false,
338
+ "special": true
339
+ },
340
+ "36039": {
341
+ "content": "<extra_id_56>",
342
+ "lstrip": true,
343
+ "normalized": false,
344
+ "rstrip": true,
345
+ "single_word": false,
346
+ "special": true
347
+ },
348
+ "36040": {
349
+ "content": "<extra_id_55>",
350
+ "lstrip": true,
351
+ "normalized": false,
352
+ "rstrip": true,
353
+ "single_word": false,
354
+ "special": true
355
+ },
356
+ "36041": {
357
+ "content": "<extra_id_54>",
358
+ "lstrip": true,
359
+ "normalized": false,
360
+ "rstrip": true,
361
+ "single_word": false,
362
+ "special": true
363
+ },
364
+ "36042": {
365
+ "content": "<extra_id_53>",
366
+ "lstrip": true,
367
+ "normalized": false,
368
+ "rstrip": true,
369
+ "single_word": false,
370
+ "special": true
371
+ },
372
+ "36043": {
373
+ "content": "<extra_id_52>",
374
+ "lstrip": true,
375
+ "normalized": false,
376
+ "rstrip": true,
377
+ "single_word": false,
378
+ "special": true
379
+ },
380
+ "36044": {
381
+ "content": "<extra_id_51>",
382
+ "lstrip": true,
383
+ "normalized": false,
384
+ "rstrip": true,
385
+ "single_word": false,
386
+ "special": true
387
+ },
388
+ "36045": {
389
+ "content": "<extra_id_50>",
390
+ "lstrip": true,
391
+ "normalized": false,
392
+ "rstrip": true,
393
+ "single_word": false,
394
+ "special": true
395
+ },
396
+ "36046": {
397
+ "content": "<extra_id_49>",
398
+ "lstrip": true,
399
+ "normalized": false,
400
+ "rstrip": true,
401
+ "single_word": false,
402
+ "special": true
403
+ },
404
+ "36047": {
405
+ "content": "<extra_id_48>",
406
+ "lstrip": true,
407
+ "normalized": false,
408
+ "rstrip": true,
409
+ "single_word": false,
410
+ "special": true
411
+ },
412
+ "36048": {
413
+ "content": "<extra_id_47>",
414
+ "lstrip": true,
415
+ "normalized": false,
416
+ "rstrip": true,
417
+ "single_word": false,
418
+ "special": true
419
+ },
420
+ "36049": {
421
+ "content": "<extra_id_46>",
422
+ "lstrip": true,
423
+ "normalized": false,
424
+ "rstrip": true,
425
+ "single_word": false,
426
+ "special": true
427
+ },
428
+ "36050": {
429
+ "content": "<extra_id_45>",
430
+ "lstrip": true,
431
+ "normalized": false,
432
+ "rstrip": true,
433
+ "single_word": false,
434
+ "special": true
435
+ },
436
+ "36051": {
437
+ "content": "<extra_id_44>",
438
+ "lstrip": true,
439
+ "normalized": false,
440
+ "rstrip": true,
441
+ "single_word": false,
442
+ "special": true
443
+ },
444
+ "36052": {
445
+ "content": "<extra_id_43>",
446
+ "lstrip": true,
447
+ "normalized": false,
448
+ "rstrip": true,
449
+ "single_word": false,
450
+ "special": true
451
+ },
452
+ "36053": {
453
+ "content": "<extra_id_42>",
454
+ "lstrip": true,
455
+ "normalized": false,
456
+ "rstrip": true,
457
+ "single_word": false,
458
+ "special": true
459
+ },
460
+ "36054": {
461
+ "content": "<extra_id_41>",
462
+ "lstrip": true,
463
+ "normalized": false,
464
+ "rstrip": true,
465
+ "single_word": false,
466
+ "special": true
467
+ },
468
+ "36055": {
469
+ "content": "<extra_id_40>",
470
+ "lstrip": true,
471
+ "normalized": false,
472
+ "rstrip": true,
473
+ "single_word": false,
474
+ "special": true
475
+ },
476
+ "36056": {
477
+ "content": "<extra_id_39>",
478
+ "lstrip": true,
479
+ "normalized": false,
480
+ "rstrip": true,
481
+ "single_word": false,
482
+ "special": true
483
+ },
484
+ "36057": {
485
+ "content": "<extra_id_38>",
486
+ "lstrip": true,
487
+ "normalized": false,
488
+ "rstrip": true,
489
+ "single_word": false,
490
+ "special": true
491
+ },
492
+ "36058": {
493
+ "content": "<extra_id_37>",
494
+ "lstrip": true,
495
+ "normalized": false,
496
+ "rstrip": true,
497
+ "single_word": false,
498
+ "special": true
499
+ },
500
+ "36059": {
501
+ "content": "<extra_id_36>",
502
+ "lstrip": true,
503
+ "normalized": false,
504
+ "rstrip": true,
505
+ "single_word": false,
506
+ "special": true
507
+ },
508
+ "36060": {
509
+ "content": "<extra_id_35>",
510
+ "lstrip": true,
511
+ "normalized": false,
512
+ "rstrip": true,
513
+ "single_word": false,
514
+ "special": true
515
+ },
516
+ "36061": {
517
+ "content": "<extra_id_34>",
518
+ "lstrip": true,
519
+ "normalized": false,
520
+ "rstrip": true,
521
+ "single_word": false,
522
+ "special": true
523
+ },
524
+ "36062": {
525
+ "content": "<extra_id_33>",
526
+ "lstrip": true,
527
+ "normalized": false,
528
+ "rstrip": true,
529
+ "single_word": false,
530
+ "special": true
531
+ },
532
+ "36063": {
533
+ "content": "<extra_id_32>",
534
+ "lstrip": true,
535
+ "normalized": false,
536
+ "rstrip": true,
537
+ "single_word": false,
538
+ "special": true
539
+ },
540
+ "36064": {
541
+ "content": "<extra_id_31>",
542
+ "lstrip": true,
543
+ "normalized": false,
544
+ "rstrip": true,
545
+ "single_word": false,
546
+ "special": true
547
+ },
548
+ "36065": {
549
+ "content": "<extra_id_30>",
550
+ "lstrip": true,
551
+ "normalized": false,
552
+ "rstrip": true,
553
+ "single_word": false,
554
+ "special": true
555
+ },
556
+ "36066": {
557
+ "content": "<extra_id_29>",
558
+ "lstrip": true,
559
+ "normalized": false,
560
+ "rstrip": true,
561
+ "single_word": false,
562
+ "special": true
563
+ },
564
+ "36067": {
565
+ "content": "<extra_id_28>",
566
+ "lstrip": true,
567
+ "normalized": false,
568
+ "rstrip": true,
569
+ "single_word": false,
570
+ "special": true
571
+ },
572
+ "36068": {
573
+ "content": "<extra_id_27>",
574
+ "lstrip": true,
575
+ "normalized": false,
576
+ "rstrip": true,
577
+ "single_word": false,
578
+ "special": true
579
+ },
580
+ "36069": {
581
+ "content": "<extra_id_26>",
582
+ "lstrip": true,
583
+ "normalized": false,
584
+ "rstrip": true,
585
+ "single_word": false,
586
+ "special": true
587
+ },
588
+ "36070": {
589
+ "content": "<extra_id_25>",
590
+ "lstrip": true,
591
+ "normalized": false,
592
+ "rstrip": true,
593
+ "single_word": false,
594
+ "special": true
595
+ },
596
+ "36071": {
597
+ "content": "<extra_id_24>",
598
+ "lstrip": true,
599
+ "normalized": false,
600
+ "rstrip": true,
601
+ "single_word": false,
602
+ "special": true
603
+ },
604
+ "36072": {
605
+ "content": "<extra_id_23>",
606
+ "lstrip": true,
607
+ "normalized": false,
608
+ "rstrip": true,
609
+ "single_word": false,
610
+ "special": true
611
+ },
612
+ "36073": {
613
+ "content": "<extra_id_22>",
614
+ "lstrip": true,
615
+ "normalized": false,
616
+ "rstrip": true,
617
+ "single_word": false,
618
+ "special": true
619
+ },
620
+ "36074": {
621
+ "content": "<extra_id_21>",
622
+ "lstrip": true,
623
+ "normalized": false,
624
+ "rstrip": true,
625
+ "single_word": false,
626
+ "special": true
627
+ },
628
+ "36075": {
629
+ "content": "<extra_id_20>",
630
+ "lstrip": true,
631
+ "normalized": false,
632
+ "rstrip": true,
633
+ "single_word": false,
634
+ "special": true
635
+ },
636
+ "36076": {
637
+ "content": "<extra_id_19>",
638
+ "lstrip": true,
639
+ "normalized": false,
640
+ "rstrip": true,
641
+ "single_word": false,
642
+ "special": true
643
+ },
644
+ "36077": {
645
+ "content": "<extra_id_18>",
646
+ "lstrip": true,
647
+ "normalized": false,
648
+ "rstrip": true,
649
+ "single_word": false,
650
+ "special": true
651
+ },
652
+ "36078": {
653
+ "content": "<extra_id_17>",
654
+ "lstrip": true,
655
+ "normalized": false,
656
+ "rstrip": true,
657
+ "single_word": false,
658
+ "special": true
659
+ },
660
+ "36079": {
661
+ "content": "<extra_id_16>",
662
+ "lstrip": true,
663
+ "normalized": false,
664
+ "rstrip": true,
665
+ "single_word": false,
666
+ "special": true
667
+ },
668
+ "36080": {
669
+ "content": "<extra_id_15>",
670
+ "lstrip": true,
671
+ "normalized": false,
672
+ "rstrip": true,
673
+ "single_word": false,
674
+ "special": true
675
+ },
676
+ "36081": {
677
+ "content": "<extra_id_14>",
678
+ "lstrip": true,
679
+ "normalized": false,
680
+ "rstrip": true,
681
+ "single_word": false,
682
+ "special": true
683
+ },
684
+ "36082": {
685
+ "content": "<extra_id_13>",
686
+ "lstrip": true,
687
+ "normalized": false,
688
+ "rstrip": true,
689
+ "single_word": false,
690
+ "special": true
691
+ },
692
+ "36083": {
693
+ "content": "<extra_id_12>",
694
+ "lstrip": true,
695
+ "normalized": false,
696
+ "rstrip": true,
697
+ "single_word": false,
698
+ "special": true
699
+ },
700
+ "36084": {
701
+ "content": "<extra_id_11>",
702
+ "lstrip": true,
703
+ "normalized": false,
704
+ "rstrip": true,
705
+ "single_word": false,
706
+ "special": true
707
+ },
708
+ "36085": {
709
+ "content": "<extra_id_10>",
710
+ "lstrip": true,
711
+ "normalized": false,
712
+ "rstrip": true,
713
+ "single_word": false,
714
+ "special": true
715
+ },
716
+ "36086": {
717
+ "content": "<extra_id_9>",
718
+ "lstrip": true,
719
+ "normalized": false,
720
+ "rstrip": true,
721
+ "single_word": false,
722
+ "special": true
723
+ },
724
+ "36087": {
725
+ "content": "<extra_id_8>",
726
+ "lstrip": true,
727
+ "normalized": false,
728
+ "rstrip": true,
729
+ "single_word": false,
730
+ "special": true
731
+ },
732
+ "36088": {
733
+ "content": "<extra_id_7>",
734
+ "lstrip": true,
735
+ "normalized": false,
736
+ "rstrip": true,
737
+ "single_word": false,
738
+ "special": true
739
+ },
740
+ "36089": {
741
+ "content": "<extra_id_6>",
742
+ "lstrip": true,
743
+ "normalized": false,
744
+ "rstrip": true,
745
+ "single_word": false,
746
+ "special": true
747
+ },
748
+ "36090": {
749
+ "content": "<extra_id_5>",
750
+ "lstrip": true,
751
+ "normalized": false,
752
+ "rstrip": true,
753
+ "single_word": false,
754
+ "special": true
755
+ },
756
+ "36091": {
757
+ "content": "<extra_id_4>",
758
+ "lstrip": true,
759
+ "normalized": false,
760
+ "rstrip": true,
761
+ "single_word": false,
762
+ "special": true
763
+ },
764
+ "36092": {
765
+ "content": "<extra_id_3>",
766
+ "lstrip": true,
767
+ "normalized": false,
768
+ "rstrip": true,
769
+ "single_word": false,
770
+ "special": true
771
+ },
772
+ "36093": {
773
+ "content": "<extra_id_2>",
774
+ "lstrip": true,
775
+ "normalized": false,
776
+ "rstrip": true,
777
+ "single_word": false,
778
+ "special": true
779
+ },
780
+ "36094": {
781
+ "content": "<extra_id_1>",
782
+ "lstrip": true,
783
+ "normalized": false,
784
+ "rstrip": true,
785
+ "single_word": false,
786
+ "special": true
787
+ },
788
+ "36095": {
789
+ "content": "<extra_id_0>",
790
+ "lstrip": true,
791
+ "normalized": false,
792
+ "rstrip": true,
793
+ "single_word": false,
794
+ "special": true
795
+ }
796
+ },
797
+ "additional_special_tokens": [
798
+ "<extra_id_0>",
799
+ "<extra_id_1>",
800
+ "<extra_id_2>",
801
+ "<extra_id_3>",
802
+ "<extra_id_4>",
803
+ "<extra_id_5>",
804
+ "<extra_id_6>",
805
+ "<extra_id_7>",
806
+ "<extra_id_8>",
807
+ "<extra_id_9>",
808
+ "<extra_id_10>",
809
+ "<extra_id_11>",
810
+ "<extra_id_12>",
811
+ "<extra_id_13>",
812
+ "<extra_id_14>",
813
+ "<extra_id_15>",
814
+ "<extra_id_16>",
815
+ "<extra_id_17>",
816
+ "<extra_id_18>",
817
+ "<extra_id_19>",
818
+ "<extra_id_20>",
819
+ "<extra_id_21>",
820
+ "<extra_id_22>",
821
+ "<extra_id_23>",
822
+ "<extra_id_24>",
823
+ "<extra_id_25>",
824
+ "<extra_id_26>",
825
+ "<extra_id_27>",
826
+ "<extra_id_28>",
827
+ "<extra_id_29>",
828
+ "<extra_id_30>",
829
+ "<extra_id_31>",
830
+ "<extra_id_32>",
831
+ "<extra_id_33>",
832
+ "<extra_id_34>",
833
+ "<extra_id_35>",
834
+ "<extra_id_36>",
835
+ "<extra_id_37>",
836
+ "<extra_id_38>",
837
+ "<extra_id_39>",
838
+ "<extra_id_40>",
839
+ "<extra_id_41>",
840
+ "<extra_id_42>",
841
+ "<extra_id_43>",
842
+ "<extra_id_44>",
843
+ "<extra_id_45>",
844
+ "<extra_id_46>",
845
+ "<extra_id_47>",
846
+ "<extra_id_48>",
847
+ "<extra_id_49>",
848
+ "<extra_id_50>",
849
+ "<extra_id_51>",
850
+ "<extra_id_52>",
851
+ "<extra_id_53>",
852
+ "<extra_id_54>",
853
+ "<extra_id_55>",
854
+ "<extra_id_56>",
855
+ "<extra_id_57>",
856
+ "<extra_id_58>",
857
+ "<extra_id_59>",
858
+ "<extra_id_60>",
859
+ "<extra_id_61>",
860
+ "<extra_id_62>",
861
+ "<extra_id_63>",
862
+ "<extra_id_64>",
863
+ "<extra_id_65>",
864
+ "<extra_id_66>",
865
+ "<extra_id_67>",
866
+ "<extra_id_68>",
867
+ "<extra_id_69>",
868
+ "<extra_id_70>",
869
+ "<extra_id_71>",
870
+ "<extra_id_72>",
871
+ "<extra_id_73>",
872
+ "<extra_id_74>",
873
+ "<extra_id_75>",
874
+ "<extra_id_76>",
875
+ "<extra_id_77>",
876
+ "<extra_id_78>",
877
+ "<extra_id_79>",
878
+ "<extra_id_80>",
879
+ "<extra_id_81>",
880
+ "<extra_id_82>",
881
+ "<extra_id_83>",
882
+ "<extra_id_84>",
883
+ "<extra_id_85>",
884
+ "<extra_id_86>",
885
+ "<extra_id_87>",
886
+ "<extra_id_88>",
887
+ "<extra_id_89>",
888
+ "<extra_id_90>",
889
+ "<extra_id_91>",
890
+ "<extra_id_92>",
891
+ "<extra_id_93>",
892
+ "<extra_id_94>",
893
+ "<extra_id_95>"
894
+ ],
895
+ "clean_up_tokenization_spaces": false,
896
+ "eos_token": "</s>",
897
+ "extra_ids": 96,
898
+ "extra_special_tokens": {},
899
+ "legacy": true,
900
+ "model_max_length": 1000000000000000019884624838656,
901
+ "pad_token": "<pad>",
902
+ "sp_model_kwargs": {},
903
+ "tokenizer_class": "T5Tokenizer",
904
+ "unk_token": "<unk>"
905
+ }
HVU_QA/t5-viet-qg-finetuned/checkpoint-3000/trainer_state.json ADDED
@@ -0,0 +1,2134 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 0.046796031696512135,
6
+ "eval_steps": 500,
7
+ "global_step": 3000,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.00015598677232170712,
14
+ "grad_norm": 3.216968059539795,
15
+ "learning_rate": 0.0001999906407936607,
16
+ "loss": 4.7744,
17
+ "step": 10
18
+ },
19
+ {
20
+ "epoch": 0.00031197354464341425,
21
+ "grad_norm": 9.366360664367676,
22
+ "learning_rate": 0.00019998024167550594,
23
+ "loss": 0.9639,
24
+ "step": 20
25
+ },
26
+ {
27
+ "epoch": 0.00046796031696512137,
28
+ "grad_norm": 4.382473945617676,
29
+ "learning_rate": 0.00019996984255735113,
30
+ "loss": 0.6338,
31
+ "step": 30
32
+ },
33
+ {
34
+ "epoch": 0.0006239470892868285,
35
+ "grad_norm": 14.415914535522461,
36
+ "learning_rate": 0.00019995944343919637,
37
+ "loss": 0.4962,
38
+ "step": 40
39
+ },
40
+ {
41
+ "epoch": 0.0007799338616085356,
42
+ "grad_norm": 7.440260410308838,
43
+ "learning_rate": 0.00019994904432104157,
44
+ "loss": 0.4619,
45
+ "step": 50
46
+ },
47
+ {
48
+ "epoch": 0.0009359206339302427,
49
+ "grad_norm": 27.18022918701172,
50
+ "learning_rate": 0.00019993864520288681,
51
+ "loss": 0.5335,
52
+ "step": 60
53
+ },
54
+ {
55
+ "epoch": 0.0010919074062519499,
56
+ "grad_norm": 7.954453468322754,
57
+ "learning_rate": 0.000199928246084732,
58
+ "loss": 0.8085,
59
+ "step": 70
60
+ },
61
+ {
62
+ "epoch": 0.001247894178573657,
63
+ "grad_norm": 4.762608528137207,
64
+ "learning_rate": 0.00019991784696657725,
65
+ "loss": 0.6333,
66
+ "step": 80
67
+ },
68
+ {
69
+ "epoch": 0.0014038809508953641,
70
+ "grad_norm": 9.577717781066895,
71
+ "learning_rate": 0.00019990744784842245,
72
+ "loss": 0.4558,
73
+ "step": 90
74
+ },
75
+ {
76
+ "epoch": 0.0015598677232170712,
77
+ "grad_norm": 7.586806297302246,
78
+ "learning_rate": 0.0001998970487302677,
79
+ "loss": 0.3227,
80
+ "step": 100
81
+ },
82
+ {
83
+ "epoch": 0.0017158544955387784,
84
+ "grad_norm": 4.262249946594238,
85
+ "learning_rate": 0.00019988664961211288,
86
+ "loss": 0.3261,
87
+ "step": 110
88
+ },
89
+ {
90
+ "epoch": 0.0018718412678604855,
91
+ "grad_norm": 6.639514923095703,
92
+ "learning_rate": 0.00019987625049395813,
93
+ "loss": 0.6788,
94
+ "step": 120
95
+ },
96
+ {
97
+ "epoch": 0.0020278280401821924,
98
+ "grad_norm": 3.7448127269744873,
99
+ "learning_rate": 0.00019986585137580332,
100
+ "loss": 0.4224,
101
+ "step": 130
102
+ },
103
+ {
104
+ "epoch": 0.0021838148125038997,
105
+ "grad_norm": 13.205582618713379,
106
+ "learning_rate": 0.00019985545225764857,
107
+ "loss": 0.4074,
108
+ "step": 140
109
+ },
110
+ {
111
+ "epoch": 0.0023398015848256066,
112
+ "grad_norm": 5.55623722076416,
113
+ "learning_rate": 0.00019984505313949376,
114
+ "loss": 0.3408,
115
+ "step": 150
116
+ },
117
+ {
118
+ "epoch": 0.002495788357147314,
119
+ "grad_norm": 10.253294944763184,
120
+ "learning_rate": 0.000199834654021339,
121
+ "loss": 0.4513,
122
+ "step": 160
123
+ },
124
+ {
125
+ "epoch": 0.002651775129469021,
126
+ "grad_norm": 12.778948783874512,
127
+ "learning_rate": 0.0001998242549031842,
128
+ "loss": 0.4287,
129
+ "step": 170
130
+ },
131
+ {
132
+ "epoch": 0.0028077619017907282,
133
+ "grad_norm": 8.961956977844238,
134
+ "learning_rate": 0.00019981385578502945,
135
+ "loss": 0.6746,
136
+ "step": 180
137
+ },
138
+ {
139
+ "epoch": 0.002963748674112435,
140
+ "grad_norm": 6.611206531524658,
141
+ "learning_rate": 0.00019980345666687467,
142
+ "loss": 0.4106,
143
+ "step": 190
144
+ },
145
+ {
146
+ "epoch": 0.0031197354464341425,
147
+ "grad_norm": 4.179671764373779,
148
+ "learning_rate": 0.0001997930575487199,
149
+ "loss": 0.4234,
150
+ "step": 200
151
+ },
152
+ {
153
+ "epoch": 0.0032757222187558494,
154
+ "grad_norm": 7.957104206085205,
155
+ "learning_rate": 0.0001997826584305651,
156
+ "loss": 0.4062,
157
+ "step": 210
158
+ },
159
+ {
160
+ "epoch": 0.0034317089910775567,
161
+ "grad_norm": 10.04617977142334,
162
+ "learning_rate": 0.00019977225931241033,
163
+ "loss": 0.4792,
164
+ "step": 220
165
+ },
166
+ {
167
+ "epoch": 0.0035876957633992636,
168
+ "grad_norm": 5.119971752166748,
169
+ "learning_rate": 0.00019976186019425555,
170
+ "loss": 0.3385,
171
+ "step": 230
172
+ },
173
+ {
174
+ "epoch": 0.003743682535720971,
175
+ "grad_norm": 6.039366722106934,
176
+ "learning_rate": 0.00019975146107610077,
177
+ "loss": 0.4406,
178
+ "step": 240
179
+ },
180
+ {
181
+ "epoch": 0.003899669308042678,
182
+ "grad_norm": 9.345914840698242,
183
+ "learning_rate": 0.000199741061957946,
184
+ "loss": 0.5163,
185
+ "step": 250
186
+ },
187
+ {
188
+ "epoch": 0.004055656080364385,
189
+ "grad_norm": 3.707460641860962,
190
+ "learning_rate": 0.0001997306628397912,
191
+ "loss": 0.2971,
192
+ "step": 260
193
+ },
194
+ {
195
+ "epoch": 0.0042116428526860926,
196
+ "grad_norm": 11.500152587890625,
197
+ "learning_rate": 0.00019972026372163643,
198
+ "loss": 0.3659,
199
+ "step": 270
200
+ },
201
+ {
202
+ "epoch": 0.0043676296250077995,
203
+ "grad_norm": 0.7893226742744446,
204
+ "learning_rate": 0.00019970986460348165,
205
+ "loss": 0.5052,
206
+ "step": 280
207
+ },
208
+ {
209
+ "epoch": 0.004523616397329506,
210
+ "grad_norm": 3.2410888671875,
211
+ "learning_rate": 0.00019969946548532687,
212
+ "loss": 0.4038,
213
+ "step": 290
214
+ },
215
+ {
216
+ "epoch": 0.004679603169651213,
217
+ "grad_norm": 4.468445777893066,
218
+ "learning_rate": 0.00019968906636717208,
219
+ "loss": 0.6076,
220
+ "step": 300
221
+ },
222
+ {
223
+ "epoch": 0.004835589941972921,
224
+ "grad_norm": 8.308616638183594,
225
+ "learning_rate": 0.0001996786672490173,
226
+ "loss": 0.4465,
227
+ "step": 310
228
+ },
229
+ {
230
+ "epoch": 0.004991576714294628,
231
+ "grad_norm": 3.736629009246826,
232
+ "learning_rate": 0.0001996682681308625,
233
+ "loss": 0.4133,
234
+ "step": 320
235
+ },
236
+ {
237
+ "epoch": 0.005147563486616335,
238
+ "grad_norm": 2.157092809677124,
239
+ "learning_rate": 0.00019965786901270774,
240
+ "loss": 0.3663,
241
+ "step": 330
242
+ },
243
+ {
244
+ "epoch": 0.005303550258938042,
245
+ "grad_norm": 5.357859134674072,
246
+ "learning_rate": 0.00019964746989455294,
247
+ "loss": 0.5599,
248
+ "step": 340
249
+ },
250
+ {
251
+ "epoch": 0.0054595370312597495,
252
+ "grad_norm": 5.352982044219971,
253
+ "learning_rate": 0.00019963707077639818,
254
+ "loss": 0.6796,
255
+ "step": 350
256
+ },
257
+ {
258
+ "epoch": 0.0056155238035814565,
259
+ "grad_norm": 3.7645204067230225,
260
+ "learning_rate": 0.00019962667165824338,
261
+ "loss": 0.5774,
262
+ "step": 360
263
+ },
264
+ {
265
+ "epoch": 0.005771510575903163,
266
+ "grad_norm": 6.222738265991211,
267
+ "learning_rate": 0.00019961627254008862,
268
+ "loss": 0.641,
269
+ "step": 370
270
+ },
271
+ {
272
+ "epoch": 0.00592749734822487,
273
+ "grad_norm": 6.460178852081299,
274
+ "learning_rate": 0.00019960587342193381,
275
+ "loss": 0.3702,
276
+ "step": 380
277
+ },
278
+ {
279
+ "epoch": 0.006083484120546578,
280
+ "grad_norm": 3.1196858882904053,
281
+ "learning_rate": 0.00019959547430377906,
282
+ "loss": 0.5036,
283
+ "step": 390
284
+ },
285
+ {
286
+ "epoch": 0.006239470892868285,
287
+ "grad_norm": 7.03672981262207,
288
+ "learning_rate": 0.00019958507518562425,
289
+ "loss": 0.41,
290
+ "step": 400
291
+ },
292
+ {
293
+ "epoch": 0.006395457665189992,
294
+ "grad_norm": 15.546775817871094,
295
+ "learning_rate": 0.0001995746760674695,
296
+ "loss": 0.5623,
297
+ "step": 410
298
+ },
299
+ {
300
+ "epoch": 0.006551444437511699,
301
+ "grad_norm": 4.868813514709473,
302
+ "learning_rate": 0.0001995642769493147,
303
+ "loss": 0.4,
304
+ "step": 420
305
+ },
306
+ {
307
+ "epoch": 0.0067074312098334065,
308
+ "grad_norm": 3.5767409801483154,
309
+ "learning_rate": 0.00019955387783115994,
310
+ "loss": 0.5193,
311
+ "step": 430
312
+ },
313
+ {
314
+ "epoch": 0.0068634179821551134,
315
+ "grad_norm": 10.07199764251709,
316
+ "learning_rate": 0.00019954347871300513,
317
+ "loss": 0.5692,
318
+ "step": 440
319
+ },
320
+ {
321
+ "epoch": 0.00701940475447682,
322
+ "grad_norm": 2.809025764465332,
323
+ "learning_rate": 0.00019953307959485038,
324
+ "loss": 0.4361,
325
+ "step": 450
326
+ },
327
+ {
328
+ "epoch": 0.007175391526798527,
329
+ "grad_norm": 14.244531631469727,
330
+ "learning_rate": 0.00019952268047669557,
331
+ "loss": 0.3646,
332
+ "step": 460
333
+ },
334
+ {
335
+ "epoch": 0.007331378299120235,
336
+ "grad_norm": 5.827071666717529,
337
+ "learning_rate": 0.00019951228135854082,
338
+ "loss": 0.5302,
339
+ "step": 470
340
+ },
341
+ {
342
+ "epoch": 0.007487365071441942,
343
+ "grad_norm": 9.067925453186035,
344
+ "learning_rate": 0.000199501882240386,
345
+ "loss": 0.5045,
346
+ "step": 480
347
+ },
348
+ {
349
+ "epoch": 0.007643351843763649,
350
+ "grad_norm": 8.674382209777832,
351
+ "learning_rate": 0.00019949148312223126,
352
+ "loss": 0.5778,
353
+ "step": 490
354
+ },
355
+ {
356
+ "epoch": 0.007799338616085356,
357
+ "grad_norm": 8.94734001159668,
358
+ "learning_rate": 0.00019948108400407645,
359
+ "loss": 0.4141,
360
+ "step": 500
361
+ },
362
+ {
363
+ "epoch": 0.007955325388407063,
364
+ "grad_norm": 10.810057640075684,
365
+ "learning_rate": 0.0001994706848859217,
366
+ "loss": 0.4069,
367
+ "step": 510
368
+ },
369
+ {
370
+ "epoch": 0.00811131216072877,
371
+ "grad_norm": 9.531766891479492,
372
+ "learning_rate": 0.0001994602857677669,
373
+ "loss": 0.5935,
374
+ "step": 520
375
+ },
376
+ {
377
+ "epoch": 0.008267298933050476,
378
+ "grad_norm": 13.168464660644531,
379
+ "learning_rate": 0.00019944988664961214,
380
+ "loss": 0.4725,
381
+ "step": 530
382
+ },
383
+ {
384
+ "epoch": 0.008423285705372185,
385
+ "grad_norm": 4.697941780090332,
386
+ "learning_rate": 0.00019943948753145733,
387
+ "loss": 0.4516,
388
+ "step": 540
389
+ },
390
+ {
391
+ "epoch": 0.008579272477693892,
392
+ "grad_norm": 2.92124342918396,
393
+ "learning_rate": 0.00019942908841330258,
394
+ "loss": 0.4562,
395
+ "step": 550
396
+ },
397
+ {
398
+ "epoch": 0.008735259250015599,
399
+ "grad_norm": 9.23039722442627,
400
+ "learning_rate": 0.00019941868929514777,
401
+ "loss": 0.4051,
402
+ "step": 560
403
+ },
404
+ {
405
+ "epoch": 0.008891246022337306,
406
+ "grad_norm": 10.528753280639648,
407
+ "learning_rate": 0.00019940829017699302,
408
+ "loss": 0.3475,
409
+ "step": 570
410
+ },
411
+ {
412
+ "epoch": 0.009047232794659013,
413
+ "grad_norm": 6.963693618774414,
414
+ "learning_rate": 0.0001993978910588382,
415
+ "loss": 0.5338,
416
+ "step": 580
417
+ },
418
+ {
419
+ "epoch": 0.00920321956698072,
420
+ "grad_norm": 6.09390926361084,
421
+ "learning_rate": 0.00019938749194068345,
422
+ "loss": 0.6501,
423
+ "step": 590
424
+ },
425
+ {
426
+ "epoch": 0.009359206339302427,
427
+ "grad_norm": 4.319295406341553,
428
+ "learning_rate": 0.00019937709282252865,
429
+ "loss": 0.4163,
430
+ "step": 600
431
+ },
432
+ {
433
+ "epoch": 0.009515193111624133,
434
+ "grad_norm": 7.514628887176514,
435
+ "learning_rate": 0.0001993666937043739,
436
+ "loss": 0.5193,
437
+ "step": 610
438
+ },
439
+ {
440
+ "epoch": 0.009671179883945842,
441
+ "grad_norm": 5.011287212371826,
442
+ "learning_rate": 0.00019935629458621909,
443
+ "loss": 0.5743,
444
+ "step": 620
445
+ },
446
+ {
447
+ "epoch": 0.009827166656267549,
448
+ "grad_norm": 6.090816974639893,
449
+ "learning_rate": 0.00019934589546806433,
450
+ "loss": 0.3902,
451
+ "step": 630
452
+ },
453
+ {
454
+ "epoch": 0.009983153428589256,
455
+ "grad_norm": 10.161015510559082,
456
+ "learning_rate": 0.00019933549634990953,
457
+ "loss": 0.4463,
458
+ "step": 640
459
+ },
460
+ {
461
+ "epoch": 0.010139140200910963,
462
+ "grad_norm": 6.14881706237793,
463
+ "learning_rate": 0.00019932509723175477,
464
+ "loss": 0.5009,
465
+ "step": 650
466
+ },
467
+ {
468
+ "epoch": 0.01029512697323267,
469
+ "grad_norm": 3.2960498332977295,
470
+ "learning_rate": 0.00019931469811359996,
471
+ "loss": 0.4597,
472
+ "step": 660
473
+ },
474
+ {
475
+ "epoch": 0.010451113745554377,
476
+ "grad_norm": 7.595890998840332,
477
+ "learning_rate": 0.0001993042989954452,
478
+ "loss": 0.5948,
479
+ "step": 670
480
+ },
481
+ {
482
+ "epoch": 0.010607100517876084,
483
+ "grad_norm": 7.178264617919922,
484
+ "learning_rate": 0.0001992938998772904,
485
+ "loss": 0.7185,
486
+ "step": 680
487
+ },
488
+ {
489
+ "epoch": 0.01076308729019779,
490
+ "grad_norm": 4.092066764831543,
491
+ "learning_rate": 0.00019928350075913565,
492
+ "loss": 0.3544,
493
+ "step": 690
494
+ },
495
+ {
496
+ "epoch": 0.010919074062519499,
497
+ "grad_norm": 13.203527450561523,
498
+ "learning_rate": 0.00019927310164098084,
499
+ "loss": 0.5719,
500
+ "step": 700
501
+ },
502
+ {
503
+ "epoch": 0.011075060834841206,
504
+ "grad_norm": 5.39072847366333,
505
+ "learning_rate": 0.0001992627025228261,
506
+ "loss": 0.5527,
507
+ "step": 710
508
+ },
509
+ {
510
+ "epoch": 0.011231047607162913,
511
+ "grad_norm": 5.769185543060303,
512
+ "learning_rate": 0.00019925230340467128,
513
+ "loss": 0.4555,
514
+ "step": 720
515
+ },
516
+ {
517
+ "epoch": 0.01138703437948462,
518
+ "grad_norm": 3.0080807209014893,
519
+ "learning_rate": 0.00019924190428651653,
520
+ "loss": 0.277,
521
+ "step": 730
522
+ },
523
+ {
524
+ "epoch": 0.011543021151806327,
525
+ "grad_norm": 1.0591176748275757,
526
+ "learning_rate": 0.00019923150516836172,
527
+ "loss": 0.513,
528
+ "step": 740
529
+ },
530
+ {
531
+ "epoch": 0.011699007924128034,
532
+ "grad_norm": 8.458100318908691,
533
+ "learning_rate": 0.00019922110605020697,
534
+ "loss": 0.5038,
535
+ "step": 750
536
+ },
537
+ {
538
+ "epoch": 0.01185499469644974,
539
+ "grad_norm": 3.121551990509033,
540
+ "learning_rate": 0.00019921070693205216,
541
+ "loss": 0.2639,
542
+ "step": 760
543
+ },
544
+ {
545
+ "epoch": 0.012010981468771447,
546
+ "grad_norm": 6.6408610343933105,
547
+ "learning_rate": 0.00019920030781389738,
548
+ "loss": 0.6632,
549
+ "step": 770
550
+ },
551
+ {
552
+ "epoch": 0.012166968241093156,
553
+ "grad_norm": 7.180611610412598,
554
+ "learning_rate": 0.0001991899086957426,
555
+ "loss": 0.5173,
556
+ "step": 780
557
+ },
558
+ {
559
+ "epoch": 0.012322955013414863,
560
+ "grad_norm": 4.404491901397705,
561
+ "learning_rate": 0.00019917950957758782,
562
+ "loss": 0.3628,
563
+ "step": 790
564
+ },
565
+ {
566
+ "epoch": 0.01247894178573657,
567
+ "grad_norm": 5.945431232452393,
568
+ "learning_rate": 0.00019916911045943304,
569
+ "loss": 0.3574,
570
+ "step": 800
571
+ },
572
+ {
573
+ "epoch": 0.012634928558058277,
574
+ "grad_norm": 5.105412483215332,
575
+ "learning_rate": 0.00019915871134127826,
576
+ "loss": 0.7038,
577
+ "step": 810
578
+ },
579
+ {
580
+ "epoch": 0.012790915330379984,
581
+ "grad_norm": 3.7237210273742676,
582
+ "learning_rate": 0.00019914831222312348,
583
+ "loss": 0.4426,
584
+ "step": 820
585
+ },
586
+ {
587
+ "epoch": 0.01294690210270169,
588
+ "grad_norm": 10.387605667114258,
589
+ "learning_rate": 0.0001991379131049687,
590
+ "loss": 0.3152,
591
+ "step": 830
592
+ },
593
+ {
594
+ "epoch": 0.013102888875023398,
595
+ "grad_norm": 2.875211477279663,
596
+ "learning_rate": 0.00019912751398681392,
597
+ "loss": 0.424,
598
+ "step": 840
599
+ },
600
+ {
601
+ "epoch": 0.013258875647345104,
602
+ "grad_norm": 1.0501248836517334,
603
+ "learning_rate": 0.00019911711486865914,
604
+ "loss": 0.5571,
605
+ "step": 850
606
+ },
607
+ {
608
+ "epoch": 0.013414862419666813,
609
+ "grad_norm": 6.748145580291748,
610
+ "learning_rate": 0.00019910671575050436,
611
+ "loss": 0.5226,
612
+ "step": 860
613
+ },
614
+ {
615
+ "epoch": 0.01357084919198852,
616
+ "grad_norm": 6.29374885559082,
617
+ "learning_rate": 0.00019909631663234958,
618
+ "loss": 0.3036,
619
+ "step": 870
620
+ },
621
+ {
622
+ "epoch": 0.013726835964310227,
623
+ "grad_norm": 5.8310699462890625,
624
+ "learning_rate": 0.0001990859175141948,
625
+ "loss": 0.4704,
626
+ "step": 880
627
+ },
628
+ {
629
+ "epoch": 0.013882822736631934,
630
+ "grad_norm": 1.4619086980819702,
631
+ "learning_rate": 0.00019907551839604002,
632
+ "loss": 0.3639,
633
+ "step": 890
634
+ },
635
+ {
636
+ "epoch": 0.01403880950895364,
637
+ "grad_norm": 8.177449226379395,
638
+ "learning_rate": 0.00019906511927788524,
639
+ "loss": 0.3853,
640
+ "step": 900
641
+ },
642
+ {
643
+ "epoch": 0.014194796281275348,
644
+ "grad_norm": 2.045549154281616,
645
+ "learning_rate": 0.00019905472015973046,
646
+ "loss": 0.3728,
647
+ "step": 910
648
+ },
649
+ {
650
+ "epoch": 0.014350783053597055,
651
+ "grad_norm": 6.4138875007629395,
652
+ "learning_rate": 0.00019904432104157568,
653
+ "loss": 0.4615,
654
+ "step": 920
655
+ },
656
+ {
657
+ "epoch": 0.014506769825918761,
658
+ "grad_norm": 2.442739486694336,
659
+ "learning_rate": 0.0001990339219234209,
660
+ "loss": 0.3592,
661
+ "step": 930
662
+ },
663
+ {
664
+ "epoch": 0.01466275659824047,
665
+ "grad_norm": 3.0741333961486816,
666
+ "learning_rate": 0.00019902352280526611,
667
+ "loss": 0.4118,
668
+ "step": 940
669
+ },
670
+ {
671
+ "epoch": 0.014818743370562177,
672
+ "grad_norm": 5.163224220275879,
673
+ "learning_rate": 0.00019901312368711133,
674
+ "loss": 0.7433,
675
+ "step": 950
676
+ },
677
+ {
678
+ "epoch": 0.014974730142883884,
679
+ "grad_norm": 5.6979193687438965,
680
+ "learning_rate": 0.00019900272456895655,
681
+ "loss": 0.3094,
682
+ "step": 960
683
+ },
684
+ {
685
+ "epoch": 0.01513071691520559,
686
+ "grad_norm": 3.2027347087860107,
687
+ "learning_rate": 0.00019899232545080177,
688
+ "loss": 0.3749,
689
+ "step": 970
690
+ },
691
+ {
692
+ "epoch": 0.015286703687527298,
693
+ "grad_norm": 9.097466468811035,
694
+ "learning_rate": 0.000198981926332647,
695
+ "loss": 0.5344,
696
+ "step": 980
697
+ },
698
+ {
699
+ "epoch": 0.015442690459849005,
700
+ "grad_norm": 3.730250835418701,
701
+ "learning_rate": 0.0001989715272144922,
702
+ "loss": 0.4085,
703
+ "step": 990
704
+ },
705
+ {
706
+ "epoch": 0.015598677232170712,
707
+ "grad_norm": 4.03507661819458,
708
+ "learning_rate": 0.00019896112809633743,
709
+ "loss": 0.7935,
710
+ "step": 1000
711
+ },
712
+ {
713
+ "epoch": 0.01575466400449242,
714
+ "grad_norm": 6.676806449890137,
715
+ "learning_rate": 0.00019895072897818265,
716
+ "loss": 0.4276,
717
+ "step": 1010
718
+ },
719
+ {
720
+ "epoch": 0.015910650776814125,
721
+ "grad_norm": 2.294931173324585,
722
+ "learning_rate": 0.00019894032986002787,
723
+ "loss": 0.2544,
724
+ "step": 1020
725
+ },
726
+ {
727
+ "epoch": 0.016066637549135832,
728
+ "grad_norm": 15.660760879516602,
729
+ "learning_rate": 0.0001989299307418731,
730
+ "loss": 0.2441,
731
+ "step": 1030
732
+ },
733
+ {
734
+ "epoch": 0.01622262432145754,
735
+ "grad_norm": 3.4223055839538574,
736
+ "learning_rate": 0.00019891953162371834,
737
+ "loss": 0.4861,
738
+ "step": 1040
739
+ },
740
+ {
741
+ "epoch": 0.016378611093779246,
742
+ "grad_norm": 8.601675987243652,
743
+ "learning_rate": 0.00019890913250556353,
744
+ "loss": 0.7212,
745
+ "step": 1050
746
+ },
747
+ {
748
+ "epoch": 0.016534597866100953,
749
+ "grad_norm": 3.814286708831787,
750
+ "learning_rate": 0.00019889873338740878,
751
+ "loss": 0.4785,
752
+ "step": 1060
753
+ },
754
+ {
755
+ "epoch": 0.016690584638422663,
756
+ "grad_norm": 10.6279296875,
757
+ "learning_rate": 0.00019888833426925397,
758
+ "loss": 0.278,
759
+ "step": 1070
760
+ },
761
+ {
762
+ "epoch": 0.01684657141074437,
763
+ "grad_norm": 2.3790910243988037,
764
+ "learning_rate": 0.00019887793515109922,
765
+ "loss": 0.3589,
766
+ "step": 1080
767
+ },
768
+ {
769
+ "epoch": 0.017002558183066077,
770
+ "grad_norm": 8.939804077148438,
771
+ "learning_rate": 0.0001988675360329444,
772
+ "loss": 0.5794,
773
+ "step": 1090
774
+ },
775
+ {
776
+ "epoch": 0.017158544955387784,
777
+ "grad_norm": 7.628971576690674,
778
+ "learning_rate": 0.00019885713691478966,
779
+ "loss": 0.5317,
780
+ "step": 1100
781
+ },
782
+ {
783
+ "epoch": 0.01731453172770949,
784
+ "grad_norm": 1.8617459535598755,
785
+ "learning_rate": 0.00019884673779663485,
786
+ "loss": 0.433,
787
+ "step": 1110
788
+ },
789
+ {
790
+ "epoch": 0.017470518500031198,
791
+ "grad_norm": 6.928651809692383,
792
+ "learning_rate": 0.0001988363386784801,
793
+ "loss": 0.216,
794
+ "step": 1120
795
+ },
796
+ {
797
+ "epoch": 0.017626505272352905,
798
+ "grad_norm": 3.6206703186035156,
799
+ "learning_rate": 0.0001988259395603253,
800
+ "loss": 0.4963,
801
+ "step": 1130
802
+ },
803
+ {
804
+ "epoch": 0.01778249204467461,
805
+ "grad_norm": 4.615293025970459,
806
+ "learning_rate": 0.00019881554044217053,
807
+ "loss": 0.3579,
808
+ "step": 1140
809
+ },
810
+ {
811
+ "epoch": 0.01793847881699632,
812
+ "grad_norm": 3.3206489086151123,
813
+ "learning_rate": 0.00019880514132401573,
814
+ "loss": 0.3717,
815
+ "step": 1150
816
+ },
817
+ {
818
+ "epoch": 0.018094465589318025,
819
+ "grad_norm": 7.5789408683776855,
820
+ "learning_rate": 0.00019879474220586097,
821
+ "loss": 0.2646,
822
+ "step": 1160
823
+ },
824
+ {
825
+ "epoch": 0.018250452361639732,
826
+ "grad_norm": 4.467035293579102,
827
+ "learning_rate": 0.00019878434308770617,
828
+ "loss": 0.1804,
829
+ "step": 1170
830
+ },
831
+ {
832
+ "epoch": 0.01840643913396144,
833
+ "grad_norm": 9.318495750427246,
834
+ "learning_rate": 0.0001987739439695514,
835
+ "loss": 0.3257,
836
+ "step": 1180
837
+ },
838
+ {
839
+ "epoch": 0.018562425906283146,
840
+ "grad_norm": 3.1046931743621826,
841
+ "learning_rate": 0.0001987635448513966,
842
+ "loss": 0.4267,
843
+ "step": 1190
844
+ },
845
+ {
846
+ "epoch": 0.018718412678604853,
847
+ "grad_norm": 4.199711799621582,
848
+ "learning_rate": 0.00019875314573324185,
849
+ "loss": 0.6676,
850
+ "step": 1200
851
+ },
852
+ {
853
+ "epoch": 0.01887439945092656,
854
+ "grad_norm": 12.733593940734863,
855
+ "learning_rate": 0.00019874274661508705,
856
+ "loss": 0.5991,
857
+ "step": 1210
858
+ },
859
+ {
860
+ "epoch": 0.019030386223248267,
861
+ "grad_norm": 5.1266984939575195,
862
+ "learning_rate": 0.0001987323474969323,
863
+ "loss": 0.2945,
864
+ "step": 1220
865
+ },
866
+ {
867
+ "epoch": 0.019186372995569977,
868
+ "grad_norm": 6.397432804107666,
869
+ "learning_rate": 0.00019872194837877748,
870
+ "loss": 0.4774,
871
+ "step": 1230
872
+ },
873
+ {
874
+ "epoch": 0.019342359767891684,
875
+ "grad_norm": 3.761129379272461,
876
+ "learning_rate": 0.0001987115492606227,
877
+ "loss": 0.4345,
878
+ "step": 1240
879
+ },
880
+ {
881
+ "epoch": 0.01949834654021339,
882
+ "grad_norm": 2.7355475425720215,
883
+ "learning_rate": 0.00019870115014246792,
884
+ "loss": 0.497,
885
+ "step": 1250
886
+ },
887
+ {
888
+ "epoch": 0.019654333312535098,
889
+ "grad_norm": 2.7893755435943604,
890
+ "learning_rate": 0.00019869075102431314,
891
+ "loss": 0.2169,
892
+ "step": 1260
893
+ },
894
+ {
895
+ "epoch": 0.019810320084856805,
896
+ "grad_norm": 10.722111701965332,
897
+ "learning_rate": 0.00019868035190615836,
898
+ "loss": 0.3867,
899
+ "step": 1270
900
+ },
901
+ {
902
+ "epoch": 0.019966306857178512,
903
+ "grad_norm": 11.647889137268066,
904
+ "learning_rate": 0.00019866995278800358,
905
+ "loss": 0.4138,
906
+ "step": 1280
907
+ },
908
+ {
909
+ "epoch": 0.02012229362950022,
910
+ "grad_norm": 1.4313205480575562,
911
+ "learning_rate": 0.0001986595536698488,
912
+ "loss": 0.3581,
913
+ "step": 1290
914
+ },
915
+ {
916
+ "epoch": 0.020278280401821926,
917
+ "grad_norm": 5.193609714508057,
918
+ "learning_rate": 0.00019864915455169402,
919
+ "loss": 0.5259,
920
+ "step": 1300
921
+ },
922
+ {
923
+ "epoch": 0.020434267174143633,
924
+ "grad_norm": 3.1939520835876465,
925
+ "learning_rate": 0.00019863875543353924,
926
+ "loss": 0.5391,
927
+ "step": 1310
928
+ },
929
+ {
930
+ "epoch": 0.02059025394646534,
931
+ "grad_norm": 7.716986179351807,
932
+ "learning_rate": 0.00019862835631538446,
933
+ "loss": 0.4773,
934
+ "step": 1320
935
+ },
936
+ {
937
+ "epoch": 0.020746240718787046,
938
+ "grad_norm": 3.191506862640381,
939
+ "learning_rate": 0.00019861795719722968,
940
+ "loss": 0.6051,
941
+ "step": 1330
942
+ },
943
+ {
944
+ "epoch": 0.020902227491108753,
945
+ "grad_norm": 5.67583703994751,
946
+ "learning_rate": 0.0001986075580790749,
947
+ "loss": 0.2716,
948
+ "step": 1340
949
+ },
950
+ {
951
+ "epoch": 0.02105821426343046,
952
+ "grad_norm": 3.1657073497772217,
953
+ "learning_rate": 0.00019859715896092012,
954
+ "loss": 0.4838,
955
+ "step": 1350
956
+ },
957
+ {
958
+ "epoch": 0.021214201035752167,
959
+ "grad_norm": 5.144644260406494,
960
+ "learning_rate": 0.00019858675984276534,
961
+ "loss": 0.3185,
962
+ "step": 1360
963
+ },
964
+ {
965
+ "epoch": 0.021370187808073874,
966
+ "grad_norm": 0.3853607177734375,
967
+ "learning_rate": 0.00019857636072461056,
968
+ "loss": 0.21,
969
+ "step": 1370
970
+ },
971
+ {
972
+ "epoch": 0.02152617458039558,
973
+ "grad_norm": 15.329313278198242,
974
+ "learning_rate": 0.00019856596160645578,
975
+ "loss": 0.4216,
976
+ "step": 1380
977
+ },
978
+ {
979
+ "epoch": 0.02168216135271729,
980
+ "grad_norm": 1.9162156581878662,
981
+ "learning_rate": 0.000198555562488301,
982
+ "loss": 0.3766,
983
+ "step": 1390
984
+ },
985
+ {
986
+ "epoch": 0.021838148125038998,
987
+ "grad_norm": 2.5800933837890625,
988
+ "learning_rate": 0.00019854516337014622,
989
+ "loss": 0.4478,
990
+ "step": 1400
991
+ },
992
+ {
993
+ "epoch": 0.021994134897360705,
994
+ "grad_norm": 12.898444175720215,
995
+ "learning_rate": 0.00019853476425199144,
996
+ "loss": 0.7923,
997
+ "step": 1410
998
+ },
999
+ {
1000
+ "epoch": 0.022150121669682412,
1001
+ "grad_norm": 11.576621055603027,
1002
+ "learning_rate": 0.00019852436513383666,
1003
+ "loss": 0.3582,
1004
+ "step": 1420
1005
+ },
1006
+ {
1007
+ "epoch": 0.02230610844200412,
1008
+ "grad_norm": 10.80225658416748,
1009
+ "learning_rate": 0.00019851396601568188,
1010
+ "loss": 0.4698,
1011
+ "step": 1430
1012
+ },
1013
+ {
1014
+ "epoch": 0.022462095214325826,
1015
+ "grad_norm": 5.019324779510498,
1016
+ "learning_rate": 0.0001985035668975271,
1017
+ "loss": 0.4463,
1018
+ "step": 1440
1019
+ },
1020
+ {
1021
+ "epoch": 0.022618081986647533,
1022
+ "grad_norm": 5.366154670715332,
1023
+ "learning_rate": 0.00019849316777937232,
1024
+ "loss": 0.4288,
1025
+ "step": 1450
1026
+ },
1027
+ {
1028
+ "epoch": 0.02277406875896924,
1029
+ "grad_norm": 3.945873498916626,
1030
+ "learning_rate": 0.00019848276866121754,
1031
+ "loss": 0.5625,
1032
+ "step": 1460
1033
+ },
1034
+ {
1035
+ "epoch": 0.022930055531290947,
1036
+ "grad_norm": 1.8938435316085815,
1037
+ "learning_rate": 0.00019847236954306276,
1038
+ "loss": 0.4298,
1039
+ "step": 1470
1040
+ },
1041
+ {
1042
+ "epoch": 0.023086042303612653,
1043
+ "grad_norm": 7.040036678314209,
1044
+ "learning_rate": 0.00019846197042490798,
1045
+ "loss": 0.5806,
1046
+ "step": 1480
1047
+ },
1048
+ {
1049
+ "epoch": 0.02324202907593436,
1050
+ "grad_norm": 12.009136199951172,
1051
+ "learning_rate": 0.0001984515713067532,
1052
+ "loss": 0.3603,
1053
+ "step": 1490
1054
+ },
1055
+ {
1056
+ "epoch": 0.023398015848256067,
1057
+ "grad_norm": 3.157111883163452,
1058
+ "learning_rate": 0.00019844117218859841,
1059
+ "loss": 0.3811,
1060
+ "step": 1500
1061
+ },
1062
+ {
1063
+ "epoch": 0.023554002620577774,
1064
+ "grad_norm": 1.0227491855621338,
1065
+ "learning_rate": 0.00019843077307044363,
1066
+ "loss": 0.3459,
1067
+ "step": 1510
1068
+ },
1069
+ {
1070
+ "epoch": 0.02370998939289948,
1071
+ "grad_norm": 2.699263334274292,
1072
+ "learning_rate": 0.00019842037395228885,
1073
+ "loss": 0.4047,
1074
+ "step": 1520
1075
+ },
1076
+ {
1077
+ "epoch": 0.023865976165221188,
1078
+ "grad_norm": 3.879258871078491,
1079
+ "learning_rate": 0.00019840997483413407,
1080
+ "loss": 0.2083,
1081
+ "step": 1530
1082
+ },
1083
+ {
1084
+ "epoch": 0.024021962937542895,
1085
+ "grad_norm": 1.0468460321426392,
1086
+ "learning_rate": 0.0001983995757159793,
1087
+ "loss": 0.3146,
1088
+ "step": 1540
1089
+ },
1090
+ {
1091
+ "epoch": 0.024177949709864602,
1092
+ "grad_norm": 3.5738143920898438,
1093
+ "learning_rate": 0.0001983891765978245,
1094
+ "loss": 0.3181,
1095
+ "step": 1550
1096
+ },
1097
+ {
1098
+ "epoch": 0.024333936482186312,
1099
+ "grad_norm": 6.3198113441467285,
1100
+ "learning_rate": 0.00019837877747966973,
1101
+ "loss": 0.3196,
1102
+ "step": 1560
1103
+ },
1104
+ {
1105
+ "epoch": 0.02448992325450802,
1106
+ "grad_norm": 1.6974947452545166,
1107
+ "learning_rate": 0.00019836837836151495,
1108
+ "loss": 0.3832,
1109
+ "step": 1570
1110
+ },
1111
+ {
1112
+ "epoch": 0.024645910026829726,
1113
+ "grad_norm": 6.220627307891846,
1114
+ "learning_rate": 0.00019835797924336017,
1115
+ "loss": 0.3445,
1116
+ "step": 1580
1117
+ },
1118
+ {
1119
+ "epoch": 0.024801896799151433,
1120
+ "grad_norm": 3.9202497005462646,
1121
+ "learning_rate": 0.0001983475801252054,
1122
+ "loss": 0.4304,
1123
+ "step": 1590
1124
+ },
1125
+ {
1126
+ "epoch": 0.02495788357147314,
1127
+ "grad_norm": 2.2945311069488525,
1128
+ "learning_rate": 0.0001983371810070506,
1129
+ "loss": 0.4487,
1130
+ "step": 1600
1131
+ },
1132
+ {
1133
+ "epoch": 0.025113870343794847,
1134
+ "grad_norm": 9.262955665588379,
1135
+ "learning_rate": 0.00019832678188889583,
1136
+ "loss": 0.3691,
1137
+ "step": 1610
1138
+ },
1139
+ {
1140
+ "epoch": 0.025269857116116554,
1141
+ "grad_norm": 2.6905221939086914,
1142
+ "learning_rate": 0.00019831638277074105,
1143
+ "loss": 0.5117,
1144
+ "step": 1620
1145
+ },
1146
+ {
1147
+ "epoch": 0.02542584388843826,
1148
+ "grad_norm": 4.1330060958862305,
1149
+ "learning_rate": 0.00019830598365258627,
1150
+ "loss": 0.5721,
1151
+ "step": 1630
1152
+ },
1153
+ {
1154
+ "epoch": 0.025581830660759967,
1155
+ "grad_norm": 1.0681806802749634,
1156
+ "learning_rate": 0.0001982955845344315,
1157
+ "loss": 0.4794,
1158
+ "step": 1640
1159
+ },
1160
+ {
1161
+ "epoch": 0.025737817433081674,
1162
+ "grad_norm": 2.6171841621398926,
1163
+ "learning_rate": 0.0001982851854162767,
1164
+ "loss": 0.4596,
1165
+ "step": 1650
1166
+ },
1167
+ {
1168
+ "epoch": 0.02589380420540338,
1169
+ "grad_norm": 3.3451762199401855,
1170
+ "learning_rate": 0.00019827478629812193,
1171
+ "loss": 0.4745,
1172
+ "step": 1660
1173
+ },
1174
+ {
1175
+ "epoch": 0.026049790977725088,
1176
+ "grad_norm": 4.152349472045898,
1177
+ "learning_rate": 0.00019826438717996715,
1178
+ "loss": 0.4452,
1179
+ "step": 1670
1180
+ },
1181
+ {
1182
+ "epoch": 0.026205777750046795,
1183
+ "grad_norm": 2.7109954357147217,
1184
+ "learning_rate": 0.00019825398806181237,
1185
+ "loss": 0.3966,
1186
+ "step": 1680
1187
+ },
1188
+ {
1189
+ "epoch": 0.026361764522368502,
1190
+ "grad_norm": 5.731235980987549,
1191
+ "learning_rate": 0.0001982435889436576,
1192
+ "loss": 0.4423,
1193
+ "step": 1690
1194
+ },
1195
+ {
1196
+ "epoch": 0.02651775129469021,
1197
+ "grad_norm": 2.370852470397949,
1198
+ "learning_rate": 0.0001982331898255028,
1199
+ "loss": 0.5619,
1200
+ "step": 1700
1201
+ },
1202
+ {
1203
+ "epoch": 0.026673738067011916,
1204
+ "grad_norm": 2.5500876903533936,
1205
+ "learning_rate": 0.00019822279070734803,
1206
+ "loss": 0.2658,
1207
+ "step": 1710
1208
+ },
1209
+ {
1210
+ "epoch": 0.026829724839333626,
1211
+ "grad_norm": 0.20157204568386078,
1212
+ "learning_rate": 0.00019821239158919325,
1213
+ "loss": 0.2939,
1214
+ "step": 1720
1215
+ },
1216
+ {
1217
+ "epoch": 0.026985711611655333,
1218
+ "grad_norm": 2.6305084228515625,
1219
+ "learning_rate": 0.00019820199247103847,
1220
+ "loss": 0.5339,
1221
+ "step": 1730
1222
+ },
1223
+ {
1224
+ "epoch": 0.02714169838397704,
1225
+ "grad_norm": 6.241075038909912,
1226
+ "learning_rate": 0.00019819159335288369,
1227
+ "loss": 0.3625,
1228
+ "step": 1740
1229
+ },
1230
+ {
1231
+ "epoch": 0.027297685156298747,
1232
+ "grad_norm": 5.584851264953613,
1233
+ "learning_rate": 0.0001981811942347289,
1234
+ "loss": 0.6158,
1235
+ "step": 1750
1236
+ },
1237
+ {
1238
+ "epoch": 0.027453671928620454,
1239
+ "grad_norm": 6.756530284881592,
1240
+ "learning_rate": 0.00019817079511657413,
1241
+ "loss": 0.3513,
1242
+ "step": 1760
1243
+ },
1244
+ {
1245
+ "epoch": 0.02760965870094216,
1246
+ "grad_norm": 5.153111457824707,
1247
+ "learning_rate": 0.00019816039599841935,
1248
+ "loss": 0.2885,
1249
+ "step": 1770
1250
+ },
1251
+ {
1252
+ "epoch": 0.027765645473263868,
1253
+ "grad_norm": 6.222873210906982,
1254
+ "learning_rate": 0.00019814999688026456,
1255
+ "loss": 0.3291,
1256
+ "step": 1780
1257
+ },
1258
+ {
1259
+ "epoch": 0.027921632245585574,
1260
+ "grad_norm": 4.744391918182373,
1261
+ "learning_rate": 0.00019813959776210978,
1262
+ "loss": 0.5053,
1263
+ "step": 1790
1264
+ },
1265
+ {
1266
+ "epoch": 0.02807761901790728,
1267
+ "grad_norm": 3.7181131839752197,
1268
+ "learning_rate": 0.000198129198643955,
1269
+ "loss": 0.3983,
1270
+ "step": 1800
1271
+ },
1272
+ {
1273
+ "epoch": 0.02823360579022899,
1274
+ "grad_norm": 1.8923834562301636,
1275
+ "learning_rate": 0.00019811879952580022,
1276
+ "loss": 0.4465,
1277
+ "step": 1810
1278
+ },
1279
+ {
1280
+ "epoch": 0.028389592562550695,
1281
+ "grad_norm": 6.049956321716309,
1282
+ "learning_rate": 0.00019810840040764544,
1283
+ "loss": 0.5028,
1284
+ "step": 1820
1285
+ },
1286
+ {
1287
+ "epoch": 0.028545579334872402,
1288
+ "grad_norm": 8.034428596496582,
1289
+ "learning_rate": 0.00019809800128949066,
1290
+ "loss": 0.4799,
1291
+ "step": 1830
1292
+ },
1293
+ {
1294
+ "epoch": 0.02870156610719411,
1295
+ "grad_norm": 3.916890859603882,
1296
+ "learning_rate": 0.00019808760217133588,
1297
+ "loss": 0.3109,
1298
+ "step": 1840
1299
+ },
1300
+ {
1301
+ "epoch": 0.028857552879515816,
1302
+ "grad_norm": 3.351285457611084,
1303
+ "learning_rate": 0.0001980772030531811,
1304
+ "loss": 0.5981,
1305
+ "step": 1850
1306
+ },
1307
+ {
1308
+ "epoch": 0.029013539651837523,
1309
+ "grad_norm": 4.443714141845703,
1310
+ "learning_rate": 0.00019806680393502632,
1311
+ "loss": 0.3243,
1312
+ "step": 1860
1313
+ },
1314
+ {
1315
+ "epoch": 0.02916952642415923,
1316
+ "grad_norm": 0.6998550295829773,
1317
+ "learning_rate": 0.00019805640481687154,
1318
+ "loss": 0.2057,
1319
+ "step": 1870
1320
+ },
1321
+ {
1322
+ "epoch": 0.02932551319648094,
1323
+ "grad_norm": 12.252562522888184,
1324
+ "learning_rate": 0.00019804600569871676,
1325
+ "loss": 0.3806,
1326
+ "step": 1880
1327
+ },
1328
+ {
1329
+ "epoch": 0.029481499968802647,
1330
+ "grad_norm": 9.03868293762207,
1331
+ "learning_rate": 0.00019803560658056198,
1332
+ "loss": 0.3351,
1333
+ "step": 1890
1334
+ },
1335
+ {
1336
+ "epoch": 0.029637486741124354,
1337
+ "grad_norm": 2.3400044441223145,
1338
+ "learning_rate": 0.0001980252074624072,
1339
+ "loss": 0.3103,
1340
+ "step": 1900
1341
+ },
1342
+ {
1343
+ "epoch": 0.02979347351344606,
1344
+ "grad_norm": 2.399489164352417,
1345
+ "learning_rate": 0.00019801480834425242,
1346
+ "loss": 0.4792,
1347
+ "step": 1910
1348
+ },
1349
+ {
1350
+ "epoch": 0.029949460285767768,
1351
+ "grad_norm": 1.5152381658554077,
1352
+ "learning_rate": 0.00019800440922609764,
1353
+ "loss": 0.2873,
1354
+ "step": 1920
1355
+ },
1356
+ {
1357
+ "epoch": 0.030105447058089475,
1358
+ "grad_norm": 2.7013959884643555,
1359
+ "learning_rate": 0.00019799401010794286,
1360
+ "loss": 0.381,
1361
+ "step": 1930
1362
+ },
1363
+ {
1364
+ "epoch": 0.03026143383041118,
1365
+ "grad_norm": 0.7135261297225952,
1366
+ "learning_rate": 0.00019798361098978808,
1367
+ "loss": 0.4001,
1368
+ "step": 1940
1369
+ },
1370
+ {
1371
+ "epoch": 0.03041742060273289,
1372
+ "grad_norm": 15.61109447479248,
1373
+ "learning_rate": 0.0001979732118716333,
1374
+ "loss": 0.474,
1375
+ "step": 1950
1376
+ },
1377
+ {
1378
+ "epoch": 0.030573407375054595,
1379
+ "grad_norm": 3.236668348312378,
1380
+ "learning_rate": 0.00019796281275347852,
1381
+ "loss": 0.2149,
1382
+ "step": 1960
1383
+ },
1384
+ {
1385
+ "epoch": 0.030729394147376302,
1386
+ "grad_norm": 4.722929954528809,
1387
+ "learning_rate": 0.00019795241363532374,
1388
+ "loss": 0.5474,
1389
+ "step": 1970
1390
+ },
1391
+ {
1392
+ "epoch": 0.03088538091969801,
1393
+ "grad_norm": 6.002335548400879,
1394
+ "learning_rate": 0.00019794201451716896,
1395
+ "loss": 0.5196,
1396
+ "step": 1980
1397
+ },
1398
+ {
1399
+ "epoch": 0.031041367692019716,
1400
+ "grad_norm": 5.210022926330566,
1401
+ "learning_rate": 0.00019793161539901418,
1402
+ "loss": 0.4741,
1403
+ "step": 1990
1404
+ },
1405
+ {
1406
+ "epoch": 0.031197354464341423,
1407
+ "grad_norm": 7.984034538269043,
1408
+ "learning_rate": 0.0001979212162808594,
1409
+ "loss": 0.5399,
1410
+ "step": 2000
1411
+ },
1412
+ {
1413
+ "epoch": 0.03135334123666313,
1414
+ "grad_norm": 4.008991241455078,
1415
+ "learning_rate": 0.00019791081716270462,
1416
+ "loss": 0.4788,
1417
+ "step": 2010
1418
+ },
1419
+ {
1420
+ "epoch": 0.03150932800898484,
1421
+ "grad_norm": 4.395316123962402,
1422
+ "learning_rate": 0.00019790041804454984,
1423
+ "loss": 0.3456,
1424
+ "step": 2020
1425
+ },
1426
+ {
1427
+ "epoch": 0.03166531478130655,
1428
+ "grad_norm": 3.9784998893737793,
1429
+ "learning_rate": 0.00019789001892639506,
1430
+ "loss": 0.326,
1431
+ "step": 2030
1432
+ },
1433
+ {
1434
+ "epoch": 0.03182130155362825,
1435
+ "grad_norm": 5.040256023406982,
1436
+ "learning_rate": 0.00019787961980824028,
1437
+ "loss": 0.5264,
1438
+ "step": 2040
1439
+ },
1440
+ {
1441
+ "epoch": 0.03197728832594996,
1442
+ "grad_norm": 2.686203956604004,
1443
+ "learning_rate": 0.0001978692206900855,
1444
+ "loss": 0.5461,
1445
+ "step": 2050
1446
+ },
1447
+ {
1448
+ "epoch": 0.032133275098271664,
1449
+ "grad_norm": 1.8417892456054688,
1450
+ "learning_rate": 0.00019785882157193071,
1451
+ "loss": 0.4736,
1452
+ "step": 2060
1453
+ },
1454
+ {
1455
+ "epoch": 0.032289261870593375,
1456
+ "grad_norm": 3.623887300491333,
1457
+ "learning_rate": 0.00019784842245377593,
1458
+ "loss": 0.3742,
1459
+ "step": 2070
1460
+ },
1461
+ {
1462
+ "epoch": 0.03244524864291508,
1463
+ "grad_norm": 0.5184723734855652,
1464
+ "learning_rate": 0.00019783802333562115,
1465
+ "loss": 0.3941,
1466
+ "step": 2080
1467
+ },
1468
+ {
1469
+ "epoch": 0.03260123541523679,
1470
+ "grad_norm": 7.751046180725098,
1471
+ "learning_rate": 0.00019782762421746637,
1472
+ "loss": 0.4609,
1473
+ "step": 2090
1474
+ },
1475
+ {
1476
+ "epoch": 0.03275722218755849,
1477
+ "grad_norm": 2.547175168991089,
1478
+ "learning_rate": 0.0001978172250993116,
1479
+ "loss": 0.3089,
1480
+ "step": 2100
1481
+ },
1482
+ {
1483
+ "epoch": 0.0329132089598802,
1484
+ "grad_norm": 1.8453168869018555,
1485
+ "learning_rate": 0.0001978068259811568,
1486
+ "loss": 0.4371,
1487
+ "step": 2110
1488
+ },
1489
+ {
1490
+ "epoch": 0.033069195732201906,
1491
+ "grad_norm": 7.033441543579102,
1492
+ "learning_rate": 0.00019779642686300203,
1493
+ "loss": 0.3842,
1494
+ "step": 2120
1495
+ },
1496
+ {
1497
+ "epoch": 0.033225182504523616,
1498
+ "grad_norm": 2.2642929553985596,
1499
+ "learning_rate": 0.00019778602774484725,
1500
+ "loss": 0.2069,
1501
+ "step": 2130
1502
+ },
1503
+ {
1504
+ "epoch": 0.03338116927684533,
1505
+ "grad_norm": 4.043938636779785,
1506
+ "learning_rate": 0.00019777562862669244,
1507
+ "loss": 0.4148,
1508
+ "step": 2140
1509
+ },
1510
+ {
1511
+ "epoch": 0.03353715604916703,
1512
+ "grad_norm": 4.227573871612549,
1513
+ "learning_rate": 0.0001977652295085377,
1514
+ "loss": 0.4785,
1515
+ "step": 2150
1516
+ },
1517
+ {
1518
+ "epoch": 0.03369314282148874,
1519
+ "grad_norm": 2.0808868408203125,
1520
+ "learning_rate": 0.00019775483039038288,
1521
+ "loss": 0.3073,
1522
+ "step": 2160
1523
+ },
1524
+ {
1525
+ "epoch": 0.033849129593810444,
1526
+ "grad_norm": 4.518038749694824,
1527
+ "learning_rate": 0.00019774443127222813,
1528
+ "loss": 0.4197,
1529
+ "step": 2170
1530
+ },
1531
+ {
1532
+ "epoch": 0.034005116366132154,
1533
+ "grad_norm": 2.3731303215026855,
1534
+ "learning_rate": 0.00019773403215407332,
1535
+ "loss": 0.2753,
1536
+ "step": 2180
1537
+ },
1538
+ {
1539
+ "epoch": 0.03416110313845386,
1540
+ "grad_norm": 1.7613986730575562,
1541
+ "learning_rate": 0.00019772363303591857,
1542
+ "loss": 0.3834,
1543
+ "step": 2190
1544
+ },
1545
+ {
1546
+ "epoch": 0.03431708991077557,
1547
+ "grad_norm": 2.9518485069274902,
1548
+ "learning_rate": 0.00019771323391776376,
1549
+ "loss": 0.3544,
1550
+ "step": 2200
1551
+ },
1552
+ {
1553
+ "epoch": 0.03447307668309727,
1554
+ "grad_norm": 4.424034118652344,
1555
+ "learning_rate": 0.000197702834799609,
1556
+ "loss": 0.4046,
1557
+ "step": 2210
1558
+ },
1559
+ {
1560
+ "epoch": 0.03462906345541898,
1561
+ "grad_norm": 2.307690382003784,
1562
+ "learning_rate": 0.0001976924356814542,
1563
+ "loss": 0.1445,
1564
+ "step": 2220
1565
+ },
1566
+ {
1567
+ "epoch": 0.034785050227740685,
1568
+ "grad_norm": 2.536533832550049,
1569
+ "learning_rate": 0.00019768203656329945,
1570
+ "loss": 0.2815,
1571
+ "step": 2230
1572
+ },
1573
+ {
1574
+ "epoch": 0.034941037000062396,
1575
+ "grad_norm": 2.006728172302246,
1576
+ "learning_rate": 0.00019767163744514464,
1577
+ "loss": 0.5314,
1578
+ "step": 2240
1579
+ },
1580
+ {
1581
+ "epoch": 0.0350970237723841,
1582
+ "grad_norm": 4.729264259338379,
1583
+ "learning_rate": 0.0001976612383269899,
1584
+ "loss": 0.4547,
1585
+ "step": 2250
1586
+ },
1587
+ {
1588
+ "epoch": 0.03525301054470581,
1589
+ "grad_norm": 4.600561141967773,
1590
+ "learning_rate": 0.00019765083920883508,
1591
+ "loss": 0.3967,
1592
+ "step": 2260
1593
+ },
1594
+ {
1595
+ "epoch": 0.03540899731702751,
1596
+ "grad_norm": 0.6419815421104431,
1597
+ "learning_rate": 0.00019764044009068033,
1598
+ "loss": 0.336,
1599
+ "step": 2270
1600
+ },
1601
+ {
1602
+ "epoch": 0.03556498408934922,
1603
+ "grad_norm": 5.567695140838623,
1604
+ "learning_rate": 0.00019763004097252552,
1605
+ "loss": 0.5644,
1606
+ "step": 2280
1607
+ },
1608
+ {
1609
+ "epoch": 0.035720970861670934,
1610
+ "grad_norm": 0.5767773389816284,
1611
+ "learning_rate": 0.00019761964185437077,
1612
+ "loss": 0.2643,
1613
+ "step": 2290
1614
+ },
1615
+ {
1616
+ "epoch": 0.03587695763399264,
1617
+ "grad_norm": 4.91269063949585,
1618
+ "learning_rate": 0.00019760924273621596,
1619
+ "loss": 0.4198,
1620
+ "step": 2300
1621
+ },
1622
+ {
1623
+ "epoch": 0.03603294440631435,
1624
+ "grad_norm": 4.422239780426025,
1625
+ "learning_rate": 0.0001975988436180612,
1626
+ "loss": 0.5338,
1627
+ "step": 2310
1628
+ },
1629
+ {
1630
+ "epoch": 0.03618893117863605,
1631
+ "grad_norm": 1.672713041305542,
1632
+ "learning_rate": 0.0001975884444999064,
1633
+ "loss": 0.3214,
1634
+ "step": 2320
1635
+ },
1636
+ {
1637
+ "epoch": 0.03634491795095776,
1638
+ "grad_norm": 4.033807277679443,
1639
+ "learning_rate": 0.00019757804538175164,
1640
+ "loss": 0.3593,
1641
+ "step": 2330
1642
+ },
1643
+ {
1644
+ "epoch": 0.036500904723279465,
1645
+ "grad_norm": 3.198880910873413,
1646
+ "learning_rate": 0.00019756764626359684,
1647
+ "loss": 0.3657,
1648
+ "step": 2340
1649
+ },
1650
+ {
1651
+ "epoch": 0.036656891495601175,
1652
+ "grad_norm": 2.640443801879883,
1653
+ "learning_rate": 0.00019755724714544208,
1654
+ "loss": 0.3243,
1655
+ "step": 2350
1656
+ },
1657
+ {
1658
+ "epoch": 0.03681287826792288,
1659
+ "grad_norm": 1.562350869178772,
1660
+ "learning_rate": 0.00019754684802728728,
1661
+ "loss": 0.3362,
1662
+ "step": 2360
1663
+ },
1664
+ {
1665
+ "epoch": 0.03696886504024459,
1666
+ "grad_norm": 4.379998683929443,
1667
+ "learning_rate": 0.00019753644890913252,
1668
+ "loss": 0.541,
1669
+ "step": 2370
1670
+ },
1671
+ {
1672
+ "epoch": 0.03712485181256629,
1673
+ "grad_norm": 2.529301881790161,
1674
+ "learning_rate": 0.00019752604979097774,
1675
+ "loss": 0.5396,
1676
+ "step": 2380
1677
+ },
1678
+ {
1679
+ "epoch": 0.037280838584888,
1680
+ "grad_norm": 1.7864131927490234,
1681
+ "learning_rate": 0.00019751565067282296,
1682
+ "loss": 0.3165,
1683
+ "step": 2390
1684
+ },
1685
+ {
1686
+ "epoch": 0.037436825357209706,
1687
+ "grad_norm": 1.0566092729568481,
1688
+ "learning_rate": 0.00019750525155466818,
1689
+ "loss": 0.4225,
1690
+ "step": 2400
1691
+ },
1692
+ {
1693
+ "epoch": 0.03759281212953142,
1694
+ "grad_norm": 2.4530510902404785,
1695
+ "learning_rate": 0.0001974948524365134,
1696
+ "loss": 0.4281,
1697
+ "step": 2410
1698
+ },
1699
+ {
1700
+ "epoch": 0.03774879890185312,
1701
+ "grad_norm": 5.363385200500488,
1702
+ "learning_rate": 0.00019748445331835862,
1703
+ "loss": 0.4746,
1704
+ "step": 2420
1705
+ },
1706
+ {
1707
+ "epoch": 0.03790478567417483,
1708
+ "grad_norm": 6.038578510284424,
1709
+ "learning_rate": 0.00019747405420020384,
1710
+ "loss": 0.3182,
1711
+ "step": 2430
1712
+ },
1713
+ {
1714
+ "epoch": 0.038060772446496534,
1715
+ "grad_norm": 9.37709903717041,
1716
+ "learning_rate": 0.00019746365508204906,
1717
+ "loss": 0.3288,
1718
+ "step": 2440
1719
+ },
1720
+ {
1721
+ "epoch": 0.038216759218818244,
1722
+ "grad_norm": 28.81134033203125,
1723
+ "learning_rate": 0.00019745325596389428,
1724
+ "loss": 0.2646,
1725
+ "step": 2450
1726
+ },
1727
+ {
1728
+ "epoch": 0.038372745991139955,
1729
+ "grad_norm": 0.616172730922699,
1730
+ "learning_rate": 0.0001974428568457395,
1731
+ "loss": 0.3848,
1732
+ "step": 2460
1733
+ },
1734
+ {
1735
+ "epoch": 0.03852873276346166,
1736
+ "grad_norm": 2.9943106174468994,
1737
+ "learning_rate": 0.00019743245772758472,
1738
+ "loss": 0.4097,
1739
+ "step": 2470
1740
+ },
1741
+ {
1742
+ "epoch": 0.03868471953578337,
1743
+ "grad_norm": 4.001022815704346,
1744
+ "learning_rate": 0.00019742205860942994,
1745
+ "loss": 0.2359,
1746
+ "step": 2480
1747
+ },
1748
+ {
1749
+ "epoch": 0.03884070630810507,
1750
+ "grad_norm": 3.587982416152954,
1751
+ "learning_rate": 0.00019741165949127516,
1752
+ "loss": 0.5829,
1753
+ "step": 2490
1754
+ },
1755
+ {
1756
+ "epoch": 0.03899669308042678,
1757
+ "grad_norm": 1.906042218208313,
1758
+ "learning_rate": 0.00019740126037312038,
1759
+ "loss": 0.336,
1760
+ "step": 2500
1761
+ },
1762
+ {
1763
+ "epoch": 0.039152679852748486,
1764
+ "grad_norm": 9.536376953125,
1765
+ "learning_rate": 0.0001973908612549656,
1766
+ "loss": 0.3468,
1767
+ "step": 2510
1768
+ },
1769
+ {
1770
+ "epoch": 0.039308666625070196,
1771
+ "grad_norm": 1.582669734954834,
1772
+ "learning_rate": 0.00019738046213681082,
1773
+ "loss": 0.3534,
1774
+ "step": 2520
1775
+ },
1776
+ {
1777
+ "epoch": 0.0394646533973919,
1778
+ "grad_norm": 2.59112286567688,
1779
+ "learning_rate": 0.00019737006301865604,
1780
+ "loss": 0.239,
1781
+ "step": 2530
1782
+ },
1783
+ {
1784
+ "epoch": 0.03962064016971361,
1785
+ "grad_norm": 2.091029167175293,
1786
+ "learning_rate": 0.00019735966390050126,
1787
+ "loss": 0.5045,
1788
+ "step": 2540
1789
+ },
1790
+ {
1791
+ "epoch": 0.03977662694203531,
1792
+ "grad_norm": 5.665968894958496,
1793
+ "learning_rate": 0.00019734926478234648,
1794
+ "loss": 0.3729,
1795
+ "step": 2550
1796
+ },
1797
+ {
1798
+ "epoch": 0.039932613714357024,
1799
+ "grad_norm": 2.2713849544525146,
1800
+ "learning_rate": 0.0001973388656641917,
1801
+ "loss": 0.3598,
1802
+ "step": 2560
1803
+ },
1804
+ {
1805
+ "epoch": 0.04008860048667873,
1806
+ "grad_norm": 3.3854405879974365,
1807
+ "learning_rate": 0.00019732846654603692,
1808
+ "loss": 0.2485,
1809
+ "step": 2570
1810
+ },
1811
+ {
1812
+ "epoch": 0.04024458725900044,
1813
+ "grad_norm": 7.780918598175049,
1814
+ "learning_rate": 0.00019731806742788214,
1815
+ "loss": 0.3557,
1816
+ "step": 2580
1817
+ },
1818
+ {
1819
+ "epoch": 0.04040057403132214,
1820
+ "grad_norm": 2.7871644496917725,
1821
+ "learning_rate": 0.00019730766830972733,
1822
+ "loss": 0.7119,
1823
+ "step": 2590
1824
+ },
1825
+ {
1826
+ "epoch": 0.04055656080364385,
1827
+ "grad_norm": 2.53657603263855,
1828
+ "learning_rate": 0.00019729726919157258,
1829
+ "loss": 0.3575,
1830
+ "step": 2600
1831
+ },
1832
+ {
1833
+ "epoch": 0.040712547575965555,
1834
+ "grad_norm": 0.3270189166069031,
1835
+ "learning_rate": 0.00019728687007341777,
1836
+ "loss": 0.3733,
1837
+ "step": 2610
1838
+ },
1839
+ {
1840
+ "epoch": 0.040868534348287265,
1841
+ "grad_norm": 1.2280256748199463,
1842
+ "learning_rate": 0.00019727647095526301,
1843
+ "loss": 0.458,
1844
+ "step": 2620
1845
+ },
1846
+ {
1847
+ "epoch": 0.041024521120608975,
1848
+ "grad_norm": 0.7863756418228149,
1849
+ "learning_rate": 0.0001972660718371082,
1850
+ "loss": 0.6961,
1851
+ "step": 2630
1852
+ },
1853
+ {
1854
+ "epoch": 0.04118050789293068,
1855
+ "grad_norm": 3.562969207763672,
1856
+ "learning_rate": 0.00019725567271895345,
1857
+ "loss": 0.3823,
1858
+ "step": 2640
1859
+ },
1860
+ {
1861
+ "epoch": 0.04133649466525239,
1862
+ "grad_norm": 2.5304505825042725,
1863
+ "learning_rate": 0.00019724527360079865,
1864
+ "loss": 0.4938,
1865
+ "step": 2650
1866
+ },
1867
+ {
1868
+ "epoch": 0.04149248143757409,
1869
+ "grad_norm": 1.8894665241241455,
1870
+ "learning_rate": 0.0001972348744826439,
1871
+ "loss": 0.4397,
1872
+ "step": 2660
1873
+ },
1874
+ {
1875
+ "epoch": 0.0416484682098958,
1876
+ "grad_norm": 3.460160970687866,
1877
+ "learning_rate": 0.00019722447536448909,
1878
+ "loss": 0.3571,
1879
+ "step": 2670
1880
+ },
1881
+ {
1882
+ "epoch": 0.04180445498221751,
1883
+ "grad_norm": 4.1218061447143555,
1884
+ "learning_rate": 0.00019721407624633433,
1885
+ "loss": 0.2929,
1886
+ "step": 2680
1887
+ },
1888
+ {
1889
+ "epoch": 0.04196044175453922,
1890
+ "grad_norm": 2.9370386600494385,
1891
+ "learning_rate": 0.00019720367712817952,
1892
+ "loss": 0.4142,
1893
+ "step": 2690
1894
+ },
1895
+ {
1896
+ "epoch": 0.04211642852686092,
1897
+ "grad_norm": 5.4618401527404785,
1898
+ "learning_rate": 0.00019719327801002477,
1899
+ "loss": 0.3288,
1900
+ "step": 2700
1901
+ },
1902
+ {
1903
+ "epoch": 0.04227241529918263,
1904
+ "grad_norm": 3.1975386142730713,
1905
+ "learning_rate": 0.00019718287889186996,
1906
+ "loss": 0.3373,
1907
+ "step": 2710
1908
+ },
1909
+ {
1910
+ "epoch": 0.042428402071504334,
1911
+ "grad_norm": 4.580129623413086,
1912
+ "learning_rate": 0.0001971724797737152,
1913
+ "loss": 0.7003,
1914
+ "step": 2720
1915
+ },
1916
+ {
1917
+ "epoch": 0.042584388843826045,
1918
+ "grad_norm": 1.7264137268066406,
1919
+ "learning_rate": 0.0001971620806555604,
1920
+ "loss": 0.3375,
1921
+ "step": 2730
1922
+ },
1923
+ {
1924
+ "epoch": 0.04274037561614775,
1925
+ "grad_norm": 8.90623950958252,
1926
+ "learning_rate": 0.00019715168153740565,
1927
+ "loss": 0.4492,
1928
+ "step": 2740
1929
+ },
1930
+ {
1931
+ "epoch": 0.04289636238846946,
1932
+ "grad_norm": 8.407225608825684,
1933
+ "learning_rate": 0.00019714128241925084,
1934
+ "loss": 0.3827,
1935
+ "step": 2750
1936
+ },
1937
+ {
1938
+ "epoch": 0.04305234916079116,
1939
+ "grad_norm": 5.202430725097656,
1940
+ "learning_rate": 0.0001971308833010961,
1941
+ "loss": 0.2806,
1942
+ "step": 2760
1943
+ },
1944
+ {
1945
+ "epoch": 0.04320833593311287,
1946
+ "grad_norm": 1.9569600820541382,
1947
+ "learning_rate": 0.00019712048418294128,
1948
+ "loss": 0.2506,
1949
+ "step": 2770
1950
+ },
1951
+ {
1952
+ "epoch": 0.04336432270543458,
1953
+ "grad_norm": 3.2171106338500977,
1954
+ "learning_rate": 0.00019711008506478653,
1955
+ "loss": 0.3344,
1956
+ "step": 2780
1957
+ },
1958
+ {
1959
+ "epoch": 0.043520309477756286,
1960
+ "grad_norm": 0.7588744163513184,
1961
+ "learning_rate": 0.00019709968594663172,
1962
+ "loss": 0.3656,
1963
+ "step": 2790
1964
+ },
1965
+ {
1966
+ "epoch": 0.043676296250077996,
1967
+ "grad_norm": 1.4201767444610596,
1968
+ "learning_rate": 0.00019708928682847697,
1969
+ "loss": 0.4033,
1970
+ "step": 2800
1971
+ },
1972
+ {
1973
+ "epoch": 0.0438322830223997,
1974
+ "grad_norm": 2.439276933670044,
1975
+ "learning_rate": 0.00019707888771032216,
1976
+ "loss": 0.2828,
1977
+ "step": 2810
1978
+ },
1979
+ {
1980
+ "epoch": 0.04398826979472141,
1981
+ "grad_norm": 18.436410903930664,
1982
+ "learning_rate": 0.0001970684885921674,
1983
+ "loss": 0.6879,
1984
+ "step": 2820
1985
+ },
1986
+ {
1987
+ "epoch": 0.044144256567043114,
1988
+ "grad_norm": 3.1035194396972656,
1989
+ "learning_rate": 0.0001970580894740126,
1990
+ "loss": 0.3022,
1991
+ "step": 2830
1992
+ },
1993
+ {
1994
+ "epoch": 0.044300243339364824,
1995
+ "grad_norm": 4.849295139312744,
1996
+ "learning_rate": 0.00019704769035585785,
1997
+ "loss": 0.4078,
1998
+ "step": 2840
1999
+ },
2000
+ {
2001
+ "epoch": 0.04445623011168653,
2002
+ "grad_norm": 3.680732488632202,
2003
+ "learning_rate": 0.00019703729123770304,
2004
+ "loss": 0.4738,
2005
+ "step": 2850
2006
+ },
2007
+ {
2008
+ "epoch": 0.04461221688400824,
2009
+ "grad_norm": 2.8644354343414307,
2010
+ "learning_rate": 0.00019702689211954829,
2011
+ "loss": 0.386,
2012
+ "step": 2860
2013
+ },
2014
+ {
2015
+ "epoch": 0.04476820365632994,
2016
+ "grad_norm": 0.7621586918830872,
2017
+ "learning_rate": 0.00019701649300139348,
2018
+ "loss": 0.2032,
2019
+ "step": 2870
2020
+ },
2021
+ {
2022
+ "epoch": 0.04492419042865165,
2023
+ "grad_norm": 0.9556747078895569,
2024
+ "learning_rate": 0.00019700609388323873,
2025
+ "loss": 0.5366,
2026
+ "step": 2880
2027
+ },
2028
+ {
2029
+ "epoch": 0.045080177200973355,
2030
+ "grad_norm": 4.282146453857422,
2031
+ "learning_rate": 0.00019699569476508392,
2032
+ "loss": 0.4814,
2033
+ "step": 2890
2034
+ },
2035
+ {
2036
+ "epoch": 0.045236163973295065,
2037
+ "grad_norm": 1.905647873878479,
2038
+ "learning_rate": 0.00019698529564692916,
2039
+ "loss": 0.5273,
2040
+ "step": 2900
2041
+ },
2042
+ {
2043
+ "epoch": 0.04539215074561677,
2044
+ "grad_norm": 0.7890440821647644,
2045
+ "learning_rate": 0.00019697489652877436,
2046
+ "loss": 0.3434,
2047
+ "step": 2910
2048
+ },
2049
+ {
2050
+ "epoch": 0.04554813751793848,
2051
+ "grad_norm": 1.717529296875,
2052
+ "learning_rate": 0.0001969644974106196,
2053
+ "loss": 0.6094,
2054
+ "step": 2920
2055
+ },
2056
+ {
2057
+ "epoch": 0.04570412429026018,
2058
+ "grad_norm": 3.9639952182769775,
2059
+ "learning_rate": 0.0001969540982924648,
2060
+ "loss": 0.2958,
2061
+ "step": 2930
2062
+ },
2063
+ {
2064
+ "epoch": 0.04586011106258189,
2065
+ "grad_norm": 1.4309637546539307,
2066
+ "learning_rate": 0.00019694369917431004,
2067
+ "loss": 0.2879,
2068
+ "step": 2940
2069
+ },
2070
+ {
2071
+ "epoch": 0.0460160978349036,
2072
+ "grad_norm": 8.747053146362305,
2073
+ "learning_rate": 0.00019693330005615524,
2074
+ "loss": 0.5953,
2075
+ "step": 2950
2076
+ },
2077
+ {
2078
+ "epoch": 0.04617208460722531,
2079
+ "grad_norm": 0.027920352295041084,
2080
+ "learning_rate": 0.00019692290093800048,
2081
+ "loss": 0.2181,
2082
+ "step": 2960
2083
+ },
2084
+ {
2085
+ "epoch": 0.04632807137954702,
2086
+ "grad_norm": 1.9787837266921997,
2087
+ "learning_rate": 0.00019691250181984567,
2088
+ "loss": 0.2764,
2089
+ "step": 2970
2090
+ },
2091
+ {
2092
+ "epoch": 0.04648405815186872,
2093
+ "grad_norm": 10.669599533081055,
2094
+ "learning_rate": 0.00019690210270169092,
2095
+ "loss": 0.5824,
2096
+ "step": 2980
2097
+ },
2098
+ {
2099
+ "epoch": 0.04664004492419043,
2100
+ "grad_norm": 2.9555306434631348,
2101
+ "learning_rate": 0.00019689170358353611,
2102
+ "loss": 0.5993,
2103
+ "step": 2990
2104
+ },
2105
+ {
2106
+ "epoch": 0.046796031696512135,
2107
+ "grad_norm": 2.168975353240967,
2108
+ "learning_rate": 0.00019688130446538136,
2109
+ "loss": 0.2988,
2110
+ "step": 3000
2111
+ }
2112
+ ],
2113
+ "logging_steps": 10,
2114
+ "max_steps": 192324,
2115
+ "num_input_tokens_seen": 0,
2116
+ "num_train_epochs": 3,
2117
+ "save_steps": 500,
2118
+ "stateful_callbacks": {
2119
+ "TrainerControl": {
2120
+ "args": {
2121
+ "should_epoch_stop": false,
2122
+ "should_evaluate": false,
2123
+ "should_log": false,
2124
+ "should_save": true,
2125
+ "should_training_stop": false
2126
+ },
2127
+ "attributes": {}
2128
+ }
2129
+ },
2130
+ "total_flos": 1826873671680000.0,
2131
+ "train_batch_size": 1,
2132
+ "trial_name": null,
2133
+ "trial_params": null
2134
+ }
HVU_QA/t5-viet-qg-finetuned/checkpoint-3000/training_args.bin ADDED
Binary file (5.71 kB). View file
 
HVU_QA/t5-viet-qg-finetuned/checkpoint-3500/added_tokens.json ADDED
@@ -0,0 +1,98 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "<extra_id_0>": 36095,
3
+ "<extra_id_10>": 36085,
4
+ "<extra_id_11>": 36084,
5
+ "<extra_id_12>": 36083,
6
+ "<extra_id_13>": 36082,
7
+ "<extra_id_14>": 36081,
8
+ "<extra_id_15>": 36080,
9
+ "<extra_id_16>": 36079,
10
+ "<extra_id_17>": 36078,
11
+ "<extra_id_18>": 36077,
12
+ "<extra_id_19>": 36076,
13
+ "<extra_id_1>": 36094,
14
+ "<extra_id_20>": 36075,
15
+ "<extra_id_21>": 36074,
16
+ "<extra_id_22>": 36073,
17
+ "<extra_id_23>": 36072,
18
+ "<extra_id_24>": 36071,
19
+ "<extra_id_25>": 36070,
20
+ "<extra_id_26>": 36069,
21
+ "<extra_id_27>": 36068,
22
+ "<extra_id_28>": 36067,
23
+ "<extra_id_29>": 36066,
24
+ "<extra_id_2>": 36093,
25
+ "<extra_id_30>": 36065,
26
+ "<extra_id_31>": 36064,
27
+ "<extra_id_32>": 36063,
28
+ "<extra_id_33>": 36062,
29
+ "<extra_id_34>": 36061,
30
+ "<extra_id_35>": 36060,
31
+ "<extra_id_36>": 36059,
32
+ "<extra_id_37>": 36058,
33
+ "<extra_id_38>": 36057,
34
+ "<extra_id_39>": 36056,
35
+ "<extra_id_3>": 36092,
36
+ "<extra_id_40>": 36055,
37
+ "<extra_id_41>": 36054,
38
+ "<extra_id_42>": 36053,
39
+ "<extra_id_43>": 36052,
40
+ "<extra_id_44>": 36051,
41
+ "<extra_id_45>": 36050,
42
+ "<extra_id_46>": 36049,
43
+ "<extra_id_47>": 36048,
44
+ "<extra_id_48>": 36047,
45
+ "<extra_id_49>": 36046,
46
+ "<extra_id_4>": 36091,
47
+ "<extra_id_50>": 36045,
48
+ "<extra_id_51>": 36044,
49
+ "<extra_id_52>": 36043,
50
+ "<extra_id_53>": 36042,
51
+ "<extra_id_54>": 36041,
52
+ "<extra_id_55>": 36040,
53
+ "<extra_id_56>": 36039,
54
+ "<extra_id_57>": 36038,
55
+ "<extra_id_58>": 36037,
56
+ "<extra_id_59>": 36036,
57
+ "<extra_id_5>": 36090,
58
+ "<extra_id_60>": 36035,
59
+ "<extra_id_61>": 36034,
60
+ "<extra_id_62>": 36033,
61
+ "<extra_id_63>": 36032,
62
+ "<extra_id_64>": 36031,
63
+ "<extra_id_65>": 36030,
64
+ "<extra_id_66>": 36029,
65
+ "<extra_id_67>": 36028,
66
+ "<extra_id_68>": 36027,
67
+ "<extra_id_69>": 36026,
68
+ "<extra_id_6>": 36089,
69
+ "<extra_id_70>": 36025,
70
+ "<extra_id_71>": 36024,
71
+ "<extra_id_72>": 36023,
72
+ "<extra_id_73>": 36022,
73
+ "<extra_id_74>": 36021,
74
+ "<extra_id_75>": 36020,
75
+ "<extra_id_76>": 36019,
76
+ "<extra_id_77>": 36018,
77
+ "<extra_id_78>": 36017,
78
+ "<extra_id_79>": 36016,
79
+ "<extra_id_7>": 36088,
80
+ "<extra_id_80>": 36015,
81
+ "<extra_id_81>": 36014,
82
+ "<extra_id_82>": 36013,
83
+ "<extra_id_83>": 36012,
84
+ "<extra_id_84>": 36011,
85
+ "<extra_id_85>": 36010,
86
+ "<extra_id_86>": 36009,
87
+ "<extra_id_87>": 36008,
88
+ "<extra_id_88>": 36007,
89
+ "<extra_id_89>": 36006,
90
+ "<extra_id_8>": 36087,
91
+ "<extra_id_90>": 36005,
92
+ "<extra_id_91>": 36004,
93
+ "<extra_id_92>": 36003,
94
+ "<extra_id_93>": 36002,
95
+ "<extra_id_94>": 36001,
96
+ "<extra_id_95>": 36000,
97
+ "<extra_id_9>": 36086
98
+ }
HVU_QA/t5-viet-qg-finetuned/checkpoint-3500/config.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "T5ForConditionalGeneration"
4
+ ],
5
+ "classifier_dropout": 0.0,
6
+ "d_ff": 3072,
7
+ "d_kv": 64,
8
+ "d_model": 768,
9
+ "decoder_start_token_id": 0,
10
+ "dense_act_fn": "relu",
11
+ "dropout_rate": 0.1,
12
+ "eos_token_id": 1,
13
+ "feed_forward_proj": "relu",
14
+ "initializer_factor": 1.0,
15
+ "is_encoder_decoder": true,
16
+ "is_gated_act": false,
17
+ "layer_norm_epsilon": 1e-06,
18
+ "model_type": "t5",
19
+ "n_positions": 512,
20
+ "num_decoder_layers": 12,
21
+ "num_heads": 12,
22
+ "num_layers": 12,
23
+ "output_past": true,
24
+ "pad_token_id": 0,
25
+ "relative_attention_max_distance": 128,
26
+ "relative_attention_num_buckets": 32,
27
+ "torch_dtype": "float32",
28
+ "transformers_version": "4.53.2",
29
+ "use_cache": true,
30
+ "vocab_size": 36096
31
+ }
HVU_QA/t5-viet-qg-finetuned/checkpoint-3500/generation_config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "decoder_start_token_id": 0,
4
+ "eos_token_id": 1,
5
+ "pad_token_id": 0,
6
+ "transformers_version": "4.53.2"
7
+ }
HVU_QA/t5-viet-qg-finetuned/checkpoint-3500/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3b8dd9715ff4735133a346b2a1e22e9b240ac2bd3ae4e445f666fbb47a3e5171
3
+ size 62652416
HVU_QA/t5-viet-qg-finetuned/checkpoint-3500/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:71bc47635358979e808042dacfd5322c43a27981481b5a71b59df2b1839314ed
3
+ size 62652416
HVU_QA/t5-viet-qg-finetuned/checkpoint-3500/rng_state.pth ADDED
Binary file (14.6 kB). View file
 
HVU_QA/t5-viet-qg-finetuned/checkpoint-3500/scheduler.pt ADDED
Binary file (1.47 kB). View file
 
HVU_QA/t5-viet-qg-finetuned/checkpoint-3500/special_tokens_map.json ADDED
@@ -0,0 +1,121 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<extra_id_0>",
4
+ "<extra_id_1>",
5
+ "<extra_id_2>",
6
+ "<extra_id_3>",
7
+ "<extra_id_4>",
8
+ "<extra_id_5>",
9
+ "<extra_id_6>",
10
+ "<extra_id_7>",
11
+ "<extra_id_8>",
12
+ "<extra_id_9>",
13
+ "<extra_id_10>",
14
+ "<extra_id_11>",
15
+ "<extra_id_12>",
16
+ "<extra_id_13>",
17
+ "<extra_id_14>",
18
+ "<extra_id_15>",
19
+ "<extra_id_16>",
20
+ "<extra_id_17>",
21
+ "<extra_id_18>",
22
+ "<extra_id_19>",
23
+ "<extra_id_20>",
24
+ "<extra_id_21>",
25
+ "<extra_id_22>",
26
+ "<extra_id_23>",
27
+ "<extra_id_24>",
28
+ "<extra_id_25>",
29
+ "<extra_id_26>",
30
+ "<extra_id_27>",
31
+ "<extra_id_28>",
32
+ "<extra_id_29>",
33
+ "<extra_id_30>",
34
+ "<extra_id_31>",
35
+ "<extra_id_32>",
36
+ "<extra_id_33>",
37
+ "<extra_id_34>",
38
+ "<extra_id_35>",
39
+ "<extra_id_36>",
40
+ "<extra_id_37>",
41
+ "<extra_id_38>",
42
+ "<extra_id_39>",
43
+ "<extra_id_40>",
44
+ "<extra_id_41>",
45
+ "<extra_id_42>",
46
+ "<extra_id_43>",
47
+ "<extra_id_44>",
48
+ "<extra_id_45>",
49
+ "<extra_id_46>",
50
+ "<extra_id_47>",
51
+ "<extra_id_48>",
52
+ "<extra_id_49>",
53
+ "<extra_id_50>",
54
+ "<extra_id_51>",
55
+ "<extra_id_52>",
56
+ "<extra_id_53>",
57
+ "<extra_id_54>",
58
+ "<extra_id_55>",
59
+ "<extra_id_56>",
60
+ "<extra_id_57>",
61
+ "<extra_id_58>",
62
+ "<extra_id_59>",
63
+ "<extra_id_60>",
64
+ "<extra_id_61>",
65
+ "<extra_id_62>",
66
+ "<extra_id_63>",
67
+ "<extra_id_64>",
68
+ "<extra_id_65>",
69
+ "<extra_id_66>",
70
+ "<extra_id_67>",
71
+ "<extra_id_68>",
72
+ "<extra_id_69>",
73
+ "<extra_id_70>",
74
+ "<extra_id_71>",
75
+ "<extra_id_72>",
76
+ "<extra_id_73>",
77
+ "<extra_id_74>",
78
+ "<extra_id_75>",
79
+ "<extra_id_76>",
80
+ "<extra_id_77>",
81
+ "<extra_id_78>",
82
+ "<extra_id_79>",
83
+ "<extra_id_80>",
84
+ "<extra_id_81>",
85
+ "<extra_id_82>",
86
+ "<extra_id_83>",
87
+ "<extra_id_84>",
88
+ "<extra_id_85>",
89
+ "<extra_id_86>",
90
+ "<extra_id_87>",
91
+ "<extra_id_88>",
92
+ "<extra_id_89>",
93
+ "<extra_id_90>",
94
+ "<extra_id_91>",
95
+ "<extra_id_92>",
96
+ "<extra_id_93>",
97
+ "<extra_id_94>",
98
+ "<extra_id_95>"
99
+ ],
100
+ "eos_token": {
101
+ "content": "</s>",
102
+ "lstrip": false,
103
+ "normalized": false,
104
+ "rstrip": false,
105
+ "single_word": false
106
+ },
107
+ "pad_token": {
108
+ "content": "<pad>",
109
+ "lstrip": false,
110
+ "normalized": false,
111
+ "rstrip": false,
112
+ "single_word": false
113
+ },
114
+ "unk_token": {
115
+ "content": "<unk>",
116
+ "lstrip": false,
117
+ "normalized": false,
118
+ "rstrip": false,
119
+ "single_word": false
120
+ }
121
+ }
HVU_QA/t5-viet-qg-finetuned/checkpoint-3500/spiece.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:59986b62f9f0b90edafb9b073ea7b93d21114a5841219a1ea2399ade73f729c6
3
+ size 820370
HVU_QA/t5-viet-qg-finetuned/checkpoint-3500/tokenizer_config.json ADDED
@@ -0,0 +1,905 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": true,
3
+ "added_tokens_decoder": {
4
+ "0": {
5
+ "content": "<pad>",
6
+ "lstrip": false,
7
+ "normalized": false,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "1": {
13
+ "content": "</s>",
14
+ "lstrip": false,
15
+ "normalized": false,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
+ },
20
+ "2": {
21
+ "content": "<unk>",
22
+ "lstrip": false,
23
+ "normalized": false,
24
+ "rstrip": false,
25
+ "single_word": false,
26
+ "special": true
27
+ },
28
+ "36000": {
29
+ "content": "<extra_id_95>",
30
+ "lstrip": true,
31
+ "normalized": false,
32
+ "rstrip": true,
33
+ "single_word": false,
34
+ "special": true
35
+ },
36
+ "36001": {
37
+ "content": "<extra_id_94>",
38
+ "lstrip": true,
39
+ "normalized": false,
40
+ "rstrip": true,
41
+ "single_word": false,
42
+ "special": true
43
+ },
44
+ "36002": {
45
+ "content": "<extra_id_93>",
46
+ "lstrip": true,
47
+ "normalized": false,
48
+ "rstrip": true,
49
+ "single_word": false,
50
+ "special": true
51
+ },
52
+ "36003": {
53
+ "content": "<extra_id_92>",
54
+ "lstrip": true,
55
+ "normalized": false,
56
+ "rstrip": true,
57
+ "single_word": false,
58
+ "special": true
59
+ },
60
+ "36004": {
61
+ "content": "<extra_id_91>",
62
+ "lstrip": true,
63
+ "normalized": false,
64
+ "rstrip": true,
65
+ "single_word": false,
66
+ "special": true
67
+ },
68
+ "36005": {
69
+ "content": "<extra_id_90>",
70
+ "lstrip": true,
71
+ "normalized": false,
72
+ "rstrip": true,
73
+ "single_word": false,
74
+ "special": true
75
+ },
76
+ "36006": {
77
+ "content": "<extra_id_89>",
78
+ "lstrip": true,
79
+ "normalized": false,
80
+ "rstrip": true,
81
+ "single_word": false,
82
+ "special": true
83
+ },
84
+ "36007": {
85
+ "content": "<extra_id_88>",
86
+ "lstrip": true,
87
+ "normalized": false,
88
+ "rstrip": true,
89
+ "single_word": false,
90
+ "special": true
91
+ },
92
+ "36008": {
93
+ "content": "<extra_id_87>",
94
+ "lstrip": true,
95
+ "normalized": false,
96
+ "rstrip": true,
97
+ "single_word": false,
98
+ "special": true
99
+ },
100
+ "36009": {
101
+ "content": "<extra_id_86>",
102
+ "lstrip": true,
103
+ "normalized": false,
104
+ "rstrip": true,
105
+ "single_word": false,
106
+ "special": true
107
+ },
108
+ "36010": {
109
+ "content": "<extra_id_85>",
110
+ "lstrip": true,
111
+ "normalized": false,
112
+ "rstrip": true,
113
+ "single_word": false,
114
+ "special": true
115
+ },
116
+ "36011": {
117
+ "content": "<extra_id_84>",
118
+ "lstrip": true,
119
+ "normalized": false,
120
+ "rstrip": true,
121
+ "single_word": false,
122
+ "special": true
123
+ },
124
+ "36012": {
125
+ "content": "<extra_id_83>",
126
+ "lstrip": true,
127
+ "normalized": false,
128
+ "rstrip": true,
129
+ "single_word": false,
130
+ "special": true
131
+ },
132
+ "36013": {
133
+ "content": "<extra_id_82>",
134
+ "lstrip": true,
135
+ "normalized": false,
136
+ "rstrip": true,
137
+ "single_word": false,
138
+ "special": true
139
+ },
140
+ "36014": {
141
+ "content": "<extra_id_81>",
142
+ "lstrip": true,
143
+ "normalized": false,
144
+ "rstrip": true,
145
+ "single_word": false,
146
+ "special": true
147
+ },
148
+ "36015": {
149
+ "content": "<extra_id_80>",
150
+ "lstrip": true,
151
+ "normalized": false,
152
+ "rstrip": true,
153
+ "single_word": false,
154
+ "special": true
155
+ },
156
+ "36016": {
157
+ "content": "<extra_id_79>",
158
+ "lstrip": true,
159
+ "normalized": false,
160
+ "rstrip": true,
161
+ "single_word": false,
162
+ "special": true
163
+ },
164
+ "36017": {
165
+ "content": "<extra_id_78>",
166
+ "lstrip": true,
167
+ "normalized": false,
168
+ "rstrip": true,
169
+ "single_word": false,
170
+ "special": true
171
+ },
172
+ "36018": {
173
+ "content": "<extra_id_77>",
174
+ "lstrip": true,
175
+ "normalized": false,
176
+ "rstrip": true,
177
+ "single_word": false,
178
+ "special": true
179
+ },
180
+ "36019": {
181
+ "content": "<extra_id_76>",
182
+ "lstrip": true,
183
+ "normalized": false,
184
+ "rstrip": true,
185
+ "single_word": false,
186
+ "special": true
187
+ },
188
+ "36020": {
189
+ "content": "<extra_id_75>",
190
+ "lstrip": true,
191
+ "normalized": false,
192
+ "rstrip": true,
193
+ "single_word": false,
194
+ "special": true
195
+ },
196
+ "36021": {
197
+ "content": "<extra_id_74>",
198
+ "lstrip": true,
199
+ "normalized": false,
200
+ "rstrip": true,
201
+ "single_word": false,
202
+ "special": true
203
+ },
204
+ "36022": {
205
+ "content": "<extra_id_73>",
206
+ "lstrip": true,
207
+ "normalized": false,
208
+ "rstrip": true,
209
+ "single_word": false,
210
+ "special": true
211
+ },
212
+ "36023": {
213
+ "content": "<extra_id_72>",
214
+ "lstrip": true,
215
+ "normalized": false,
216
+ "rstrip": true,
217
+ "single_word": false,
218
+ "special": true
219
+ },
220
+ "36024": {
221
+ "content": "<extra_id_71>",
222
+ "lstrip": true,
223
+ "normalized": false,
224
+ "rstrip": true,
225
+ "single_word": false,
226
+ "special": true
227
+ },
228
+ "36025": {
229
+ "content": "<extra_id_70>",
230
+ "lstrip": true,
231
+ "normalized": false,
232
+ "rstrip": true,
233
+ "single_word": false,
234
+ "special": true
235
+ },
236
+ "36026": {
237
+ "content": "<extra_id_69>",
238
+ "lstrip": true,
239
+ "normalized": false,
240
+ "rstrip": true,
241
+ "single_word": false,
242
+ "special": true
243
+ },
244
+ "36027": {
245
+ "content": "<extra_id_68>",
246
+ "lstrip": true,
247
+ "normalized": false,
248
+ "rstrip": true,
249
+ "single_word": false,
250
+ "special": true
251
+ },
252
+ "36028": {
253
+ "content": "<extra_id_67>",
254
+ "lstrip": true,
255
+ "normalized": false,
256
+ "rstrip": true,
257
+ "single_word": false,
258
+ "special": true
259
+ },
260
+ "36029": {
261
+ "content": "<extra_id_66>",
262
+ "lstrip": true,
263
+ "normalized": false,
264
+ "rstrip": true,
265
+ "single_word": false,
266
+ "special": true
267
+ },
268
+ "36030": {
269
+ "content": "<extra_id_65>",
270
+ "lstrip": true,
271
+ "normalized": false,
272
+ "rstrip": true,
273
+ "single_word": false,
274
+ "special": true
275
+ },
276
+ "36031": {
277
+ "content": "<extra_id_64>",
278
+ "lstrip": true,
279
+ "normalized": false,
280
+ "rstrip": true,
281
+ "single_word": false,
282
+ "special": true
283
+ },
284
+ "36032": {
285
+ "content": "<extra_id_63>",
286
+ "lstrip": true,
287
+ "normalized": false,
288
+ "rstrip": true,
289
+ "single_word": false,
290
+ "special": true
291
+ },
292
+ "36033": {
293
+ "content": "<extra_id_62>",
294
+ "lstrip": true,
295
+ "normalized": false,
296
+ "rstrip": true,
297
+ "single_word": false,
298
+ "special": true
299
+ },
300
+ "36034": {
301
+ "content": "<extra_id_61>",
302
+ "lstrip": true,
303
+ "normalized": false,
304
+ "rstrip": true,
305
+ "single_word": false,
306
+ "special": true
307
+ },
308
+ "36035": {
309
+ "content": "<extra_id_60>",
310
+ "lstrip": true,
311
+ "normalized": false,
312
+ "rstrip": true,
313
+ "single_word": false,
314
+ "special": true
315
+ },
316
+ "36036": {
317
+ "content": "<extra_id_59>",
318
+ "lstrip": true,
319
+ "normalized": false,
320
+ "rstrip": true,
321
+ "single_word": false,
322
+ "special": true
323
+ },
324
+ "36037": {
325
+ "content": "<extra_id_58>",
326
+ "lstrip": true,
327
+ "normalized": false,
328
+ "rstrip": true,
329
+ "single_word": false,
330
+ "special": true
331
+ },
332
+ "36038": {
333
+ "content": "<extra_id_57>",
334
+ "lstrip": true,
335
+ "normalized": false,
336
+ "rstrip": true,
337
+ "single_word": false,
338
+ "special": true
339
+ },
340
+ "36039": {
341
+ "content": "<extra_id_56>",
342
+ "lstrip": true,
343
+ "normalized": false,
344
+ "rstrip": true,
345
+ "single_word": false,
346
+ "special": true
347
+ },
348
+ "36040": {
349
+ "content": "<extra_id_55>",
350
+ "lstrip": true,
351
+ "normalized": false,
352
+ "rstrip": true,
353
+ "single_word": false,
354
+ "special": true
355
+ },
356
+ "36041": {
357
+ "content": "<extra_id_54>",
358
+ "lstrip": true,
359
+ "normalized": false,
360
+ "rstrip": true,
361
+ "single_word": false,
362
+ "special": true
363
+ },
364
+ "36042": {
365
+ "content": "<extra_id_53>",
366
+ "lstrip": true,
367
+ "normalized": false,
368
+ "rstrip": true,
369
+ "single_word": false,
370
+ "special": true
371
+ },
372
+ "36043": {
373
+ "content": "<extra_id_52>",
374
+ "lstrip": true,
375
+ "normalized": false,
376
+ "rstrip": true,
377
+ "single_word": false,
378
+ "special": true
379
+ },
380
+ "36044": {
381
+ "content": "<extra_id_51>",
382
+ "lstrip": true,
383
+ "normalized": false,
384
+ "rstrip": true,
385
+ "single_word": false,
386
+ "special": true
387
+ },
388
+ "36045": {
389
+ "content": "<extra_id_50>",
390
+ "lstrip": true,
391
+ "normalized": false,
392
+ "rstrip": true,
393
+ "single_word": false,
394
+ "special": true
395
+ },
396
+ "36046": {
397
+ "content": "<extra_id_49>",
398
+ "lstrip": true,
399
+ "normalized": false,
400
+ "rstrip": true,
401
+ "single_word": false,
402
+ "special": true
403
+ },
404
+ "36047": {
405
+ "content": "<extra_id_48>",
406
+ "lstrip": true,
407
+ "normalized": false,
408
+ "rstrip": true,
409
+ "single_word": false,
410
+ "special": true
411
+ },
412
+ "36048": {
413
+ "content": "<extra_id_47>",
414
+ "lstrip": true,
415
+ "normalized": false,
416
+ "rstrip": true,
417
+ "single_word": false,
418
+ "special": true
419
+ },
420
+ "36049": {
421
+ "content": "<extra_id_46>",
422
+ "lstrip": true,
423
+ "normalized": false,
424
+ "rstrip": true,
425
+ "single_word": false,
426
+ "special": true
427
+ },
428
+ "36050": {
429
+ "content": "<extra_id_45>",
430
+ "lstrip": true,
431
+ "normalized": false,
432
+ "rstrip": true,
433
+ "single_word": false,
434
+ "special": true
435
+ },
436
+ "36051": {
437
+ "content": "<extra_id_44>",
438
+ "lstrip": true,
439
+ "normalized": false,
440
+ "rstrip": true,
441
+ "single_word": false,
442
+ "special": true
443
+ },
444
+ "36052": {
445
+ "content": "<extra_id_43>",
446
+ "lstrip": true,
447
+ "normalized": false,
448
+ "rstrip": true,
449
+ "single_word": false,
450
+ "special": true
451
+ },
452
+ "36053": {
453
+ "content": "<extra_id_42>",
454
+ "lstrip": true,
455
+ "normalized": false,
456
+ "rstrip": true,
457
+ "single_word": false,
458
+ "special": true
459
+ },
460
+ "36054": {
461
+ "content": "<extra_id_41>",
462
+ "lstrip": true,
463
+ "normalized": false,
464
+ "rstrip": true,
465
+ "single_word": false,
466
+ "special": true
467
+ },
468
+ "36055": {
469
+ "content": "<extra_id_40>",
470
+ "lstrip": true,
471
+ "normalized": false,
472
+ "rstrip": true,
473
+ "single_word": false,
474
+ "special": true
475
+ },
476
+ "36056": {
477
+ "content": "<extra_id_39>",
478
+ "lstrip": true,
479
+ "normalized": false,
480
+ "rstrip": true,
481
+ "single_word": false,
482
+ "special": true
483
+ },
484
+ "36057": {
485
+ "content": "<extra_id_38>",
486
+ "lstrip": true,
487
+ "normalized": false,
488
+ "rstrip": true,
489
+ "single_word": false,
490
+ "special": true
491
+ },
492
+ "36058": {
493
+ "content": "<extra_id_37>",
494
+ "lstrip": true,
495
+ "normalized": false,
496
+ "rstrip": true,
497
+ "single_word": false,
498
+ "special": true
499
+ },
500
+ "36059": {
501
+ "content": "<extra_id_36>",
502
+ "lstrip": true,
503
+ "normalized": false,
504
+ "rstrip": true,
505
+ "single_word": false,
506
+ "special": true
507
+ },
508
+ "36060": {
509
+ "content": "<extra_id_35>",
510
+ "lstrip": true,
511
+ "normalized": false,
512
+ "rstrip": true,
513
+ "single_word": false,
514
+ "special": true
515
+ },
516
+ "36061": {
517
+ "content": "<extra_id_34>",
518
+ "lstrip": true,
519
+ "normalized": false,
520
+ "rstrip": true,
521
+ "single_word": false,
522
+ "special": true
523
+ },
524
+ "36062": {
525
+ "content": "<extra_id_33>",
526
+ "lstrip": true,
527
+ "normalized": false,
528
+ "rstrip": true,
529
+ "single_word": false,
530
+ "special": true
531
+ },
532
+ "36063": {
533
+ "content": "<extra_id_32>",
534
+ "lstrip": true,
535
+ "normalized": false,
536
+ "rstrip": true,
537
+ "single_word": false,
538
+ "special": true
539
+ },
540
+ "36064": {
541
+ "content": "<extra_id_31>",
542
+ "lstrip": true,
543
+ "normalized": false,
544
+ "rstrip": true,
545
+ "single_word": false,
546
+ "special": true
547
+ },
548
+ "36065": {
549
+ "content": "<extra_id_30>",
550
+ "lstrip": true,
551
+ "normalized": false,
552
+ "rstrip": true,
553
+ "single_word": false,
554
+ "special": true
555
+ },
556
+ "36066": {
557
+ "content": "<extra_id_29>",
558
+ "lstrip": true,
559
+ "normalized": false,
560
+ "rstrip": true,
561
+ "single_word": false,
562
+ "special": true
563
+ },
564
+ "36067": {
565
+ "content": "<extra_id_28>",
566
+ "lstrip": true,
567
+ "normalized": false,
568
+ "rstrip": true,
569
+ "single_word": false,
570
+ "special": true
571
+ },
572
+ "36068": {
573
+ "content": "<extra_id_27>",
574
+ "lstrip": true,
575
+ "normalized": false,
576
+ "rstrip": true,
577
+ "single_word": false,
578
+ "special": true
579
+ },
580
+ "36069": {
581
+ "content": "<extra_id_26>",
582
+ "lstrip": true,
583
+ "normalized": false,
584
+ "rstrip": true,
585
+ "single_word": false,
586
+ "special": true
587
+ },
588
+ "36070": {
589
+ "content": "<extra_id_25>",
590
+ "lstrip": true,
591
+ "normalized": false,
592
+ "rstrip": true,
593
+ "single_word": false,
594
+ "special": true
595
+ },
596
+ "36071": {
597
+ "content": "<extra_id_24>",
598
+ "lstrip": true,
599
+ "normalized": false,
600
+ "rstrip": true,
601
+ "single_word": false,
602
+ "special": true
603
+ },
604
+ "36072": {
605
+ "content": "<extra_id_23>",
606
+ "lstrip": true,
607
+ "normalized": false,
608
+ "rstrip": true,
609
+ "single_word": false,
610
+ "special": true
611
+ },
612
+ "36073": {
613
+ "content": "<extra_id_22>",
614
+ "lstrip": true,
615
+ "normalized": false,
616
+ "rstrip": true,
617
+ "single_word": false,
618
+ "special": true
619
+ },
620
+ "36074": {
621
+ "content": "<extra_id_21>",
622
+ "lstrip": true,
623
+ "normalized": false,
624
+ "rstrip": true,
625
+ "single_word": false,
626
+ "special": true
627
+ },
628
+ "36075": {
629
+ "content": "<extra_id_20>",
630
+ "lstrip": true,
631
+ "normalized": false,
632
+ "rstrip": true,
633
+ "single_word": false,
634
+ "special": true
635
+ },
636
+ "36076": {
637
+ "content": "<extra_id_19>",
638
+ "lstrip": true,
639
+ "normalized": false,
640
+ "rstrip": true,
641
+ "single_word": false,
642
+ "special": true
643
+ },
644
+ "36077": {
645
+ "content": "<extra_id_18>",
646
+ "lstrip": true,
647
+ "normalized": false,
648
+ "rstrip": true,
649
+ "single_word": false,
650
+ "special": true
651
+ },
652
+ "36078": {
653
+ "content": "<extra_id_17>",
654
+ "lstrip": true,
655
+ "normalized": false,
656
+ "rstrip": true,
657
+ "single_word": false,
658
+ "special": true
659
+ },
660
+ "36079": {
661
+ "content": "<extra_id_16>",
662
+ "lstrip": true,
663
+ "normalized": false,
664
+ "rstrip": true,
665
+ "single_word": false,
666
+ "special": true
667
+ },
668
+ "36080": {
669
+ "content": "<extra_id_15>",
670
+ "lstrip": true,
671
+ "normalized": false,
672
+ "rstrip": true,
673
+ "single_word": false,
674
+ "special": true
675
+ },
676
+ "36081": {
677
+ "content": "<extra_id_14>",
678
+ "lstrip": true,
679
+ "normalized": false,
680
+ "rstrip": true,
681
+ "single_word": false,
682
+ "special": true
683
+ },
684
+ "36082": {
685
+ "content": "<extra_id_13>",
686
+ "lstrip": true,
687
+ "normalized": false,
688
+ "rstrip": true,
689
+ "single_word": false,
690
+ "special": true
691
+ },
692
+ "36083": {
693
+ "content": "<extra_id_12>",
694
+ "lstrip": true,
695
+ "normalized": false,
696
+ "rstrip": true,
697
+ "single_word": false,
698
+ "special": true
699
+ },
700
+ "36084": {
701
+ "content": "<extra_id_11>",
702
+ "lstrip": true,
703
+ "normalized": false,
704
+ "rstrip": true,
705
+ "single_word": false,
706
+ "special": true
707
+ },
708
+ "36085": {
709
+ "content": "<extra_id_10>",
710
+ "lstrip": true,
711
+ "normalized": false,
712
+ "rstrip": true,
713
+ "single_word": false,
714
+ "special": true
715
+ },
716
+ "36086": {
717
+ "content": "<extra_id_9>",
718
+ "lstrip": true,
719
+ "normalized": false,
720
+ "rstrip": true,
721
+ "single_word": false,
722
+ "special": true
723
+ },
724
+ "36087": {
725
+ "content": "<extra_id_8>",
726
+ "lstrip": true,
727
+ "normalized": false,
728
+ "rstrip": true,
729
+ "single_word": false,
730
+ "special": true
731
+ },
732
+ "36088": {
733
+ "content": "<extra_id_7>",
734
+ "lstrip": true,
735
+ "normalized": false,
736
+ "rstrip": true,
737
+ "single_word": false,
738
+ "special": true
739
+ },
740
+ "36089": {
741
+ "content": "<extra_id_6>",
742
+ "lstrip": true,
743
+ "normalized": false,
744
+ "rstrip": true,
745
+ "single_word": false,
746
+ "special": true
747
+ },
748
+ "36090": {
749
+ "content": "<extra_id_5>",
750
+ "lstrip": true,
751
+ "normalized": false,
752
+ "rstrip": true,
753
+ "single_word": false,
754
+ "special": true
755
+ },
756
+ "36091": {
757
+ "content": "<extra_id_4>",
758
+ "lstrip": true,
759
+ "normalized": false,
760
+ "rstrip": true,
761
+ "single_word": false,
762
+ "special": true
763
+ },
764
+ "36092": {
765
+ "content": "<extra_id_3>",
766
+ "lstrip": true,
767
+ "normalized": false,
768
+ "rstrip": true,
769
+ "single_word": false,
770
+ "special": true
771
+ },
772
+ "36093": {
773
+ "content": "<extra_id_2>",
774
+ "lstrip": true,
775
+ "normalized": false,
776
+ "rstrip": true,
777
+ "single_word": false,
778
+ "special": true
779
+ },
780
+ "36094": {
781
+ "content": "<extra_id_1>",
782
+ "lstrip": true,
783
+ "normalized": false,
784
+ "rstrip": true,
785
+ "single_word": false,
786
+ "special": true
787
+ },
788
+ "36095": {
789
+ "content": "<extra_id_0>",
790
+ "lstrip": true,
791
+ "normalized": false,
792
+ "rstrip": true,
793
+ "single_word": false,
794
+ "special": true
795
+ }
796
+ },
797
+ "additional_special_tokens": [
798
+ "<extra_id_0>",
799
+ "<extra_id_1>",
800
+ "<extra_id_2>",
801
+ "<extra_id_3>",
802
+ "<extra_id_4>",
803
+ "<extra_id_5>",
804
+ "<extra_id_6>",
805
+ "<extra_id_7>",
806
+ "<extra_id_8>",
807
+ "<extra_id_9>",
808
+ "<extra_id_10>",
809
+ "<extra_id_11>",
810
+ "<extra_id_12>",
811
+ "<extra_id_13>",
812
+ "<extra_id_14>",
813
+ "<extra_id_15>",
814
+ "<extra_id_16>",
815
+ "<extra_id_17>",
816
+ "<extra_id_18>",
817
+ "<extra_id_19>",
818
+ "<extra_id_20>",
819
+ "<extra_id_21>",
820
+ "<extra_id_22>",
821
+ "<extra_id_23>",
822
+ "<extra_id_24>",
823
+ "<extra_id_25>",
824
+ "<extra_id_26>",
825
+ "<extra_id_27>",
826
+ "<extra_id_28>",
827
+ "<extra_id_29>",
828
+ "<extra_id_30>",
829
+ "<extra_id_31>",
830
+ "<extra_id_32>",
831
+ "<extra_id_33>",
832
+ "<extra_id_34>",
833
+ "<extra_id_35>",
834
+ "<extra_id_36>",
835
+ "<extra_id_37>",
836
+ "<extra_id_38>",
837
+ "<extra_id_39>",
838
+ "<extra_id_40>",
839
+ "<extra_id_41>",
840
+ "<extra_id_42>",
841
+ "<extra_id_43>",
842
+ "<extra_id_44>",
843
+ "<extra_id_45>",
844
+ "<extra_id_46>",
845
+ "<extra_id_47>",
846
+ "<extra_id_48>",
847
+ "<extra_id_49>",
848
+ "<extra_id_50>",
849
+ "<extra_id_51>",
850
+ "<extra_id_52>",
851
+ "<extra_id_53>",
852
+ "<extra_id_54>",
853
+ "<extra_id_55>",
854
+ "<extra_id_56>",
855
+ "<extra_id_57>",
856
+ "<extra_id_58>",
857
+ "<extra_id_59>",
858
+ "<extra_id_60>",
859
+ "<extra_id_61>",
860
+ "<extra_id_62>",
861
+ "<extra_id_63>",
862
+ "<extra_id_64>",
863
+ "<extra_id_65>",
864
+ "<extra_id_66>",
865
+ "<extra_id_67>",
866
+ "<extra_id_68>",
867
+ "<extra_id_69>",
868
+ "<extra_id_70>",
869
+ "<extra_id_71>",
870
+ "<extra_id_72>",
871
+ "<extra_id_73>",
872
+ "<extra_id_74>",
873
+ "<extra_id_75>",
874
+ "<extra_id_76>",
875
+ "<extra_id_77>",
876
+ "<extra_id_78>",
877
+ "<extra_id_79>",
878
+ "<extra_id_80>",
879
+ "<extra_id_81>",
880
+ "<extra_id_82>",
881
+ "<extra_id_83>",
882
+ "<extra_id_84>",
883
+ "<extra_id_85>",
884
+ "<extra_id_86>",
885
+ "<extra_id_87>",
886
+ "<extra_id_88>",
887
+ "<extra_id_89>",
888
+ "<extra_id_90>",
889
+ "<extra_id_91>",
890
+ "<extra_id_92>",
891
+ "<extra_id_93>",
892
+ "<extra_id_94>",
893
+ "<extra_id_95>"
894
+ ],
895
+ "clean_up_tokenization_spaces": false,
896
+ "eos_token": "</s>",
897
+ "extra_ids": 96,
898
+ "extra_special_tokens": {},
899
+ "legacy": true,
900
+ "model_max_length": 1000000000000000019884624838656,
901
+ "pad_token": "<pad>",
902
+ "sp_model_kwargs": {},
903
+ "tokenizer_class": "T5Tokenizer",
904
+ "unk_token": "<unk>"
905
+ }
HVU_QA/t5-viet-qg-finetuned/checkpoint-3500/trainer_state.json ADDED
@@ -0,0 +1,2484 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 0.054595370312597494,
6
+ "eval_steps": 500,
7
+ "global_step": 3500,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.00015598677232170712,
14
+ "grad_norm": 3.216968059539795,
15
+ "learning_rate": 0.0001999906407936607,
16
+ "loss": 4.7744,
17
+ "step": 10
18
+ },
19
+ {
20
+ "epoch": 0.00031197354464341425,
21
+ "grad_norm": 9.366360664367676,
22
+ "learning_rate": 0.00019998024167550594,
23
+ "loss": 0.9639,
24
+ "step": 20
25
+ },
26
+ {
27
+ "epoch": 0.00046796031696512137,
28
+ "grad_norm": 4.382473945617676,
29
+ "learning_rate": 0.00019996984255735113,
30
+ "loss": 0.6338,
31
+ "step": 30
32
+ },
33
+ {
34
+ "epoch": 0.0006239470892868285,
35
+ "grad_norm": 14.415914535522461,
36
+ "learning_rate": 0.00019995944343919637,
37
+ "loss": 0.4962,
38
+ "step": 40
39
+ },
40
+ {
41
+ "epoch": 0.0007799338616085356,
42
+ "grad_norm": 7.440260410308838,
43
+ "learning_rate": 0.00019994904432104157,
44
+ "loss": 0.4619,
45
+ "step": 50
46
+ },
47
+ {
48
+ "epoch": 0.0009359206339302427,
49
+ "grad_norm": 27.18022918701172,
50
+ "learning_rate": 0.00019993864520288681,
51
+ "loss": 0.5335,
52
+ "step": 60
53
+ },
54
+ {
55
+ "epoch": 0.0010919074062519499,
56
+ "grad_norm": 7.954453468322754,
57
+ "learning_rate": 0.000199928246084732,
58
+ "loss": 0.8085,
59
+ "step": 70
60
+ },
61
+ {
62
+ "epoch": 0.001247894178573657,
63
+ "grad_norm": 4.762608528137207,
64
+ "learning_rate": 0.00019991784696657725,
65
+ "loss": 0.6333,
66
+ "step": 80
67
+ },
68
+ {
69
+ "epoch": 0.0014038809508953641,
70
+ "grad_norm": 9.577717781066895,
71
+ "learning_rate": 0.00019990744784842245,
72
+ "loss": 0.4558,
73
+ "step": 90
74
+ },
75
+ {
76
+ "epoch": 0.0015598677232170712,
77
+ "grad_norm": 7.586806297302246,
78
+ "learning_rate": 0.0001998970487302677,
79
+ "loss": 0.3227,
80
+ "step": 100
81
+ },
82
+ {
83
+ "epoch": 0.0017158544955387784,
84
+ "grad_norm": 4.262249946594238,
85
+ "learning_rate": 0.00019988664961211288,
86
+ "loss": 0.3261,
87
+ "step": 110
88
+ },
89
+ {
90
+ "epoch": 0.0018718412678604855,
91
+ "grad_norm": 6.639514923095703,
92
+ "learning_rate": 0.00019987625049395813,
93
+ "loss": 0.6788,
94
+ "step": 120
95
+ },
96
+ {
97
+ "epoch": 0.0020278280401821924,
98
+ "grad_norm": 3.7448127269744873,
99
+ "learning_rate": 0.00019986585137580332,
100
+ "loss": 0.4224,
101
+ "step": 130
102
+ },
103
+ {
104
+ "epoch": 0.0021838148125038997,
105
+ "grad_norm": 13.205582618713379,
106
+ "learning_rate": 0.00019985545225764857,
107
+ "loss": 0.4074,
108
+ "step": 140
109
+ },
110
+ {
111
+ "epoch": 0.0023398015848256066,
112
+ "grad_norm": 5.55623722076416,
113
+ "learning_rate": 0.00019984505313949376,
114
+ "loss": 0.3408,
115
+ "step": 150
116
+ },
117
+ {
118
+ "epoch": 0.002495788357147314,
119
+ "grad_norm": 10.253294944763184,
120
+ "learning_rate": 0.000199834654021339,
121
+ "loss": 0.4513,
122
+ "step": 160
123
+ },
124
+ {
125
+ "epoch": 0.002651775129469021,
126
+ "grad_norm": 12.778948783874512,
127
+ "learning_rate": 0.0001998242549031842,
128
+ "loss": 0.4287,
129
+ "step": 170
130
+ },
131
+ {
132
+ "epoch": 0.0028077619017907282,
133
+ "grad_norm": 8.961956977844238,
134
+ "learning_rate": 0.00019981385578502945,
135
+ "loss": 0.6746,
136
+ "step": 180
137
+ },
138
+ {
139
+ "epoch": 0.002963748674112435,
140
+ "grad_norm": 6.611206531524658,
141
+ "learning_rate": 0.00019980345666687467,
142
+ "loss": 0.4106,
143
+ "step": 190
144
+ },
145
+ {
146
+ "epoch": 0.0031197354464341425,
147
+ "grad_norm": 4.179671764373779,
148
+ "learning_rate": 0.0001997930575487199,
149
+ "loss": 0.4234,
150
+ "step": 200
151
+ },
152
+ {
153
+ "epoch": 0.0032757222187558494,
154
+ "grad_norm": 7.957104206085205,
155
+ "learning_rate": 0.0001997826584305651,
156
+ "loss": 0.4062,
157
+ "step": 210
158
+ },
159
+ {
160
+ "epoch": 0.0034317089910775567,
161
+ "grad_norm": 10.04617977142334,
162
+ "learning_rate": 0.00019977225931241033,
163
+ "loss": 0.4792,
164
+ "step": 220
165
+ },
166
+ {
167
+ "epoch": 0.0035876957633992636,
168
+ "grad_norm": 5.119971752166748,
169
+ "learning_rate": 0.00019976186019425555,
170
+ "loss": 0.3385,
171
+ "step": 230
172
+ },
173
+ {
174
+ "epoch": 0.003743682535720971,
175
+ "grad_norm": 6.039366722106934,
176
+ "learning_rate": 0.00019975146107610077,
177
+ "loss": 0.4406,
178
+ "step": 240
179
+ },
180
+ {
181
+ "epoch": 0.003899669308042678,
182
+ "grad_norm": 9.345914840698242,
183
+ "learning_rate": 0.000199741061957946,
184
+ "loss": 0.5163,
185
+ "step": 250
186
+ },
187
+ {
188
+ "epoch": 0.004055656080364385,
189
+ "grad_norm": 3.707460641860962,
190
+ "learning_rate": 0.0001997306628397912,
191
+ "loss": 0.2971,
192
+ "step": 260
193
+ },
194
+ {
195
+ "epoch": 0.0042116428526860926,
196
+ "grad_norm": 11.500152587890625,
197
+ "learning_rate": 0.00019972026372163643,
198
+ "loss": 0.3659,
199
+ "step": 270
200
+ },
201
+ {
202
+ "epoch": 0.0043676296250077995,
203
+ "grad_norm": 0.7893226742744446,
204
+ "learning_rate": 0.00019970986460348165,
205
+ "loss": 0.5052,
206
+ "step": 280
207
+ },
208
+ {
209
+ "epoch": 0.004523616397329506,
210
+ "grad_norm": 3.2410888671875,
211
+ "learning_rate": 0.00019969946548532687,
212
+ "loss": 0.4038,
213
+ "step": 290
214
+ },
215
+ {
216
+ "epoch": 0.004679603169651213,
217
+ "grad_norm": 4.468445777893066,
218
+ "learning_rate": 0.00019968906636717208,
219
+ "loss": 0.6076,
220
+ "step": 300
221
+ },
222
+ {
223
+ "epoch": 0.004835589941972921,
224
+ "grad_norm": 8.308616638183594,
225
+ "learning_rate": 0.0001996786672490173,
226
+ "loss": 0.4465,
227
+ "step": 310
228
+ },
229
+ {
230
+ "epoch": 0.004991576714294628,
231
+ "grad_norm": 3.736629009246826,
232
+ "learning_rate": 0.0001996682681308625,
233
+ "loss": 0.4133,
234
+ "step": 320
235
+ },
236
+ {
237
+ "epoch": 0.005147563486616335,
238
+ "grad_norm": 2.157092809677124,
239
+ "learning_rate": 0.00019965786901270774,
240
+ "loss": 0.3663,
241
+ "step": 330
242
+ },
243
+ {
244
+ "epoch": 0.005303550258938042,
245
+ "grad_norm": 5.357859134674072,
246
+ "learning_rate": 0.00019964746989455294,
247
+ "loss": 0.5599,
248
+ "step": 340
249
+ },
250
+ {
251
+ "epoch": 0.0054595370312597495,
252
+ "grad_norm": 5.352982044219971,
253
+ "learning_rate": 0.00019963707077639818,
254
+ "loss": 0.6796,
255
+ "step": 350
256
+ },
257
+ {
258
+ "epoch": 0.0056155238035814565,
259
+ "grad_norm": 3.7645204067230225,
260
+ "learning_rate": 0.00019962667165824338,
261
+ "loss": 0.5774,
262
+ "step": 360
263
+ },
264
+ {
265
+ "epoch": 0.005771510575903163,
266
+ "grad_norm": 6.222738265991211,
267
+ "learning_rate": 0.00019961627254008862,
268
+ "loss": 0.641,
269
+ "step": 370
270
+ },
271
+ {
272
+ "epoch": 0.00592749734822487,
273
+ "grad_norm": 6.460178852081299,
274
+ "learning_rate": 0.00019960587342193381,
275
+ "loss": 0.3702,
276
+ "step": 380
277
+ },
278
+ {
279
+ "epoch": 0.006083484120546578,
280
+ "grad_norm": 3.1196858882904053,
281
+ "learning_rate": 0.00019959547430377906,
282
+ "loss": 0.5036,
283
+ "step": 390
284
+ },
285
+ {
286
+ "epoch": 0.006239470892868285,
287
+ "grad_norm": 7.03672981262207,
288
+ "learning_rate": 0.00019958507518562425,
289
+ "loss": 0.41,
290
+ "step": 400
291
+ },
292
+ {
293
+ "epoch": 0.006395457665189992,
294
+ "grad_norm": 15.546775817871094,
295
+ "learning_rate": 0.0001995746760674695,
296
+ "loss": 0.5623,
297
+ "step": 410
298
+ },
299
+ {
300
+ "epoch": 0.006551444437511699,
301
+ "grad_norm": 4.868813514709473,
302
+ "learning_rate": 0.0001995642769493147,
303
+ "loss": 0.4,
304
+ "step": 420
305
+ },
306
+ {
307
+ "epoch": 0.0067074312098334065,
308
+ "grad_norm": 3.5767409801483154,
309
+ "learning_rate": 0.00019955387783115994,
310
+ "loss": 0.5193,
311
+ "step": 430
312
+ },
313
+ {
314
+ "epoch": 0.0068634179821551134,
315
+ "grad_norm": 10.07199764251709,
316
+ "learning_rate": 0.00019954347871300513,
317
+ "loss": 0.5692,
318
+ "step": 440
319
+ },
320
+ {
321
+ "epoch": 0.00701940475447682,
322
+ "grad_norm": 2.809025764465332,
323
+ "learning_rate": 0.00019953307959485038,
324
+ "loss": 0.4361,
325
+ "step": 450
326
+ },
327
+ {
328
+ "epoch": 0.007175391526798527,
329
+ "grad_norm": 14.244531631469727,
330
+ "learning_rate": 0.00019952268047669557,
331
+ "loss": 0.3646,
332
+ "step": 460
333
+ },
334
+ {
335
+ "epoch": 0.007331378299120235,
336
+ "grad_norm": 5.827071666717529,
337
+ "learning_rate": 0.00019951228135854082,
338
+ "loss": 0.5302,
339
+ "step": 470
340
+ },
341
+ {
342
+ "epoch": 0.007487365071441942,
343
+ "grad_norm": 9.067925453186035,
344
+ "learning_rate": 0.000199501882240386,
345
+ "loss": 0.5045,
346
+ "step": 480
347
+ },
348
+ {
349
+ "epoch": 0.007643351843763649,
350
+ "grad_norm": 8.674382209777832,
351
+ "learning_rate": 0.00019949148312223126,
352
+ "loss": 0.5778,
353
+ "step": 490
354
+ },
355
+ {
356
+ "epoch": 0.007799338616085356,
357
+ "grad_norm": 8.94734001159668,
358
+ "learning_rate": 0.00019948108400407645,
359
+ "loss": 0.4141,
360
+ "step": 500
361
+ },
362
+ {
363
+ "epoch": 0.007955325388407063,
364
+ "grad_norm": 10.810057640075684,
365
+ "learning_rate": 0.0001994706848859217,
366
+ "loss": 0.4069,
367
+ "step": 510
368
+ },
369
+ {
370
+ "epoch": 0.00811131216072877,
371
+ "grad_norm": 9.531766891479492,
372
+ "learning_rate": 0.0001994602857677669,
373
+ "loss": 0.5935,
374
+ "step": 520
375
+ },
376
+ {
377
+ "epoch": 0.008267298933050476,
378
+ "grad_norm": 13.168464660644531,
379
+ "learning_rate": 0.00019944988664961214,
380
+ "loss": 0.4725,
381
+ "step": 530
382
+ },
383
+ {
384
+ "epoch": 0.008423285705372185,
385
+ "grad_norm": 4.697941780090332,
386
+ "learning_rate": 0.00019943948753145733,
387
+ "loss": 0.4516,
388
+ "step": 540
389
+ },
390
+ {
391
+ "epoch": 0.008579272477693892,
392
+ "grad_norm": 2.92124342918396,
393
+ "learning_rate": 0.00019942908841330258,
394
+ "loss": 0.4562,
395
+ "step": 550
396
+ },
397
+ {
398
+ "epoch": 0.008735259250015599,
399
+ "grad_norm": 9.23039722442627,
400
+ "learning_rate": 0.00019941868929514777,
401
+ "loss": 0.4051,
402
+ "step": 560
403
+ },
404
+ {
405
+ "epoch": 0.008891246022337306,
406
+ "grad_norm": 10.528753280639648,
407
+ "learning_rate": 0.00019940829017699302,
408
+ "loss": 0.3475,
409
+ "step": 570
410
+ },
411
+ {
412
+ "epoch": 0.009047232794659013,
413
+ "grad_norm": 6.963693618774414,
414
+ "learning_rate": 0.0001993978910588382,
415
+ "loss": 0.5338,
416
+ "step": 580
417
+ },
418
+ {
419
+ "epoch": 0.00920321956698072,
420
+ "grad_norm": 6.09390926361084,
421
+ "learning_rate": 0.00019938749194068345,
422
+ "loss": 0.6501,
423
+ "step": 590
424
+ },
425
+ {
426
+ "epoch": 0.009359206339302427,
427
+ "grad_norm": 4.319295406341553,
428
+ "learning_rate": 0.00019937709282252865,
429
+ "loss": 0.4163,
430
+ "step": 600
431
+ },
432
+ {
433
+ "epoch": 0.009515193111624133,
434
+ "grad_norm": 7.514628887176514,
435
+ "learning_rate": 0.0001993666937043739,
436
+ "loss": 0.5193,
437
+ "step": 610
438
+ },
439
+ {
440
+ "epoch": 0.009671179883945842,
441
+ "grad_norm": 5.011287212371826,
442
+ "learning_rate": 0.00019935629458621909,
443
+ "loss": 0.5743,
444
+ "step": 620
445
+ },
446
+ {
447
+ "epoch": 0.009827166656267549,
448
+ "grad_norm": 6.090816974639893,
449
+ "learning_rate": 0.00019934589546806433,
450
+ "loss": 0.3902,
451
+ "step": 630
452
+ },
453
+ {
454
+ "epoch": 0.009983153428589256,
455
+ "grad_norm": 10.161015510559082,
456
+ "learning_rate": 0.00019933549634990953,
457
+ "loss": 0.4463,
458
+ "step": 640
459
+ },
460
+ {
461
+ "epoch": 0.010139140200910963,
462
+ "grad_norm": 6.14881706237793,
463
+ "learning_rate": 0.00019932509723175477,
464
+ "loss": 0.5009,
465
+ "step": 650
466
+ },
467
+ {
468
+ "epoch": 0.01029512697323267,
469
+ "grad_norm": 3.2960498332977295,
470
+ "learning_rate": 0.00019931469811359996,
471
+ "loss": 0.4597,
472
+ "step": 660
473
+ },
474
+ {
475
+ "epoch": 0.010451113745554377,
476
+ "grad_norm": 7.595890998840332,
477
+ "learning_rate": 0.0001993042989954452,
478
+ "loss": 0.5948,
479
+ "step": 670
480
+ },
481
+ {
482
+ "epoch": 0.010607100517876084,
483
+ "grad_norm": 7.178264617919922,
484
+ "learning_rate": 0.0001992938998772904,
485
+ "loss": 0.7185,
486
+ "step": 680
487
+ },
488
+ {
489
+ "epoch": 0.01076308729019779,
490
+ "grad_norm": 4.092066764831543,
491
+ "learning_rate": 0.00019928350075913565,
492
+ "loss": 0.3544,
493
+ "step": 690
494
+ },
495
+ {
496
+ "epoch": 0.010919074062519499,
497
+ "grad_norm": 13.203527450561523,
498
+ "learning_rate": 0.00019927310164098084,
499
+ "loss": 0.5719,
500
+ "step": 700
501
+ },
502
+ {
503
+ "epoch": 0.011075060834841206,
504
+ "grad_norm": 5.39072847366333,
505
+ "learning_rate": 0.0001992627025228261,
506
+ "loss": 0.5527,
507
+ "step": 710
508
+ },
509
+ {
510
+ "epoch": 0.011231047607162913,
511
+ "grad_norm": 5.769185543060303,
512
+ "learning_rate": 0.00019925230340467128,
513
+ "loss": 0.4555,
514
+ "step": 720
515
+ },
516
+ {
517
+ "epoch": 0.01138703437948462,
518
+ "grad_norm": 3.0080807209014893,
519
+ "learning_rate": 0.00019924190428651653,
520
+ "loss": 0.277,
521
+ "step": 730
522
+ },
523
+ {
524
+ "epoch": 0.011543021151806327,
525
+ "grad_norm": 1.0591176748275757,
526
+ "learning_rate": 0.00019923150516836172,
527
+ "loss": 0.513,
528
+ "step": 740
529
+ },
530
+ {
531
+ "epoch": 0.011699007924128034,
532
+ "grad_norm": 8.458100318908691,
533
+ "learning_rate": 0.00019922110605020697,
534
+ "loss": 0.5038,
535
+ "step": 750
536
+ },
537
+ {
538
+ "epoch": 0.01185499469644974,
539
+ "grad_norm": 3.121551990509033,
540
+ "learning_rate": 0.00019921070693205216,
541
+ "loss": 0.2639,
542
+ "step": 760
543
+ },
544
+ {
545
+ "epoch": 0.012010981468771447,
546
+ "grad_norm": 6.6408610343933105,
547
+ "learning_rate": 0.00019920030781389738,
548
+ "loss": 0.6632,
549
+ "step": 770
550
+ },
551
+ {
552
+ "epoch": 0.012166968241093156,
553
+ "grad_norm": 7.180611610412598,
554
+ "learning_rate": 0.0001991899086957426,
555
+ "loss": 0.5173,
556
+ "step": 780
557
+ },
558
+ {
559
+ "epoch": 0.012322955013414863,
560
+ "grad_norm": 4.404491901397705,
561
+ "learning_rate": 0.00019917950957758782,
562
+ "loss": 0.3628,
563
+ "step": 790
564
+ },
565
+ {
566
+ "epoch": 0.01247894178573657,
567
+ "grad_norm": 5.945431232452393,
568
+ "learning_rate": 0.00019916911045943304,
569
+ "loss": 0.3574,
570
+ "step": 800
571
+ },
572
+ {
573
+ "epoch": 0.012634928558058277,
574
+ "grad_norm": 5.105412483215332,
575
+ "learning_rate": 0.00019915871134127826,
576
+ "loss": 0.7038,
577
+ "step": 810
578
+ },
579
+ {
580
+ "epoch": 0.012790915330379984,
581
+ "grad_norm": 3.7237210273742676,
582
+ "learning_rate": 0.00019914831222312348,
583
+ "loss": 0.4426,
584
+ "step": 820
585
+ },
586
+ {
587
+ "epoch": 0.01294690210270169,
588
+ "grad_norm": 10.387605667114258,
589
+ "learning_rate": 0.0001991379131049687,
590
+ "loss": 0.3152,
591
+ "step": 830
592
+ },
593
+ {
594
+ "epoch": 0.013102888875023398,
595
+ "grad_norm": 2.875211477279663,
596
+ "learning_rate": 0.00019912751398681392,
597
+ "loss": 0.424,
598
+ "step": 840
599
+ },
600
+ {
601
+ "epoch": 0.013258875647345104,
602
+ "grad_norm": 1.0501248836517334,
603
+ "learning_rate": 0.00019911711486865914,
604
+ "loss": 0.5571,
605
+ "step": 850
606
+ },
607
+ {
608
+ "epoch": 0.013414862419666813,
609
+ "grad_norm": 6.748145580291748,
610
+ "learning_rate": 0.00019910671575050436,
611
+ "loss": 0.5226,
612
+ "step": 860
613
+ },
614
+ {
615
+ "epoch": 0.01357084919198852,
616
+ "grad_norm": 6.29374885559082,
617
+ "learning_rate": 0.00019909631663234958,
618
+ "loss": 0.3036,
619
+ "step": 870
620
+ },
621
+ {
622
+ "epoch": 0.013726835964310227,
623
+ "grad_norm": 5.8310699462890625,
624
+ "learning_rate": 0.0001990859175141948,
625
+ "loss": 0.4704,
626
+ "step": 880
627
+ },
628
+ {
629
+ "epoch": 0.013882822736631934,
630
+ "grad_norm": 1.4619086980819702,
631
+ "learning_rate": 0.00019907551839604002,
632
+ "loss": 0.3639,
633
+ "step": 890
634
+ },
635
+ {
636
+ "epoch": 0.01403880950895364,
637
+ "grad_norm": 8.177449226379395,
638
+ "learning_rate": 0.00019906511927788524,
639
+ "loss": 0.3853,
640
+ "step": 900
641
+ },
642
+ {
643
+ "epoch": 0.014194796281275348,
644
+ "grad_norm": 2.045549154281616,
645
+ "learning_rate": 0.00019905472015973046,
646
+ "loss": 0.3728,
647
+ "step": 910
648
+ },
649
+ {
650
+ "epoch": 0.014350783053597055,
651
+ "grad_norm": 6.4138875007629395,
652
+ "learning_rate": 0.00019904432104157568,
653
+ "loss": 0.4615,
654
+ "step": 920
655
+ },
656
+ {
657
+ "epoch": 0.014506769825918761,
658
+ "grad_norm": 2.442739486694336,
659
+ "learning_rate": 0.0001990339219234209,
660
+ "loss": 0.3592,
661
+ "step": 930
662
+ },
663
+ {
664
+ "epoch": 0.01466275659824047,
665
+ "grad_norm": 3.0741333961486816,
666
+ "learning_rate": 0.00019902352280526611,
667
+ "loss": 0.4118,
668
+ "step": 940
669
+ },
670
+ {
671
+ "epoch": 0.014818743370562177,
672
+ "grad_norm": 5.163224220275879,
673
+ "learning_rate": 0.00019901312368711133,
674
+ "loss": 0.7433,
675
+ "step": 950
676
+ },
677
+ {
678
+ "epoch": 0.014974730142883884,
679
+ "grad_norm": 5.6979193687438965,
680
+ "learning_rate": 0.00019900272456895655,
681
+ "loss": 0.3094,
682
+ "step": 960
683
+ },
684
+ {
685
+ "epoch": 0.01513071691520559,
686
+ "grad_norm": 3.2027347087860107,
687
+ "learning_rate": 0.00019899232545080177,
688
+ "loss": 0.3749,
689
+ "step": 970
690
+ },
691
+ {
692
+ "epoch": 0.015286703687527298,
693
+ "grad_norm": 9.097466468811035,
694
+ "learning_rate": 0.000198981926332647,
695
+ "loss": 0.5344,
696
+ "step": 980
697
+ },
698
+ {
699
+ "epoch": 0.015442690459849005,
700
+ "grad_norm": 3.730250835418701,
701
+ "learning_rate": 0.0001989715272144922,
702
+ "loss": 0.4085,
703
+ "step": 990
704
+ },
705
+ {
706
+ "epoch": 0.015598677232170712,
707
+ "grad_norm": 4.03507661819458,
708
+ "learning_rate": 0.00019896112809633743,
709
+ "loss": 0.7935,
710
+ "step": 1000
711
+ },
712
+ {
713
+ "epoch": 0.01575466400449242,
714
+ "grad_norm": 6.676806449890137,
715
+ "learning_rate": 0.00019895072897818265,
716
+ "loss": 0.4276,
717
+ "step": 1010
718
+ },
719
+ {
720
+ "epoch": 0.015910650776814125,
721
+ "grad_norm": 2.294931173324585,
722
+ "learning_rate": 0.00019894032986002787,
723
+ "loss": 0.2544,
724
+ "step": 1020
725
+ },
726
+ {
727
+ "epoch": 0.016066637549135832,
728
+ "grad_norm": 15.660760879516602,
729
+ "learning_rate": 0.0001989299307418731,
730
+ "loss": 0.2441,
731
+ "step": 1030
732
+ },
733
+ {
734
+ "epoch": 0.01622262432145754,
735
+ "grad_norm": 3.4223055839538574,
736
+ "learning_rate": 0.00019891953162371834,
737
+ "loss": 0.4861,
738
+ "step": 1040
739
+ },
740
+ {
741
+ "epoch": 0.016378611093779246,
742
+ "grad_norm": 8.601675987243652,
743
+ "learning_rate": 0.00019890913250556353,
744
+ "loss": 0.7212,
745
+ "step": 1050
746
+ },
747
+ {
748
+ "epoch": 0.016534597866100953,
749
+ "grad_norm": 3.814286708831787,
750
+ "learning_rate": 0.00019889873338740878,
751
+ "loss": 0.4785,
752
+ "step": 1060
753
+ },
754
+ {
755
+ "epoch": 0.016690584638422663,
756
+ "grad_norm": 10.6279296875,
757
+ "learning_rate": 0.00019888833426925397,
758
+ "loss": 0.278,
759
+ "step": 1070
760
+ },
761
+ {
762
+ "epoch": 0.01684657141074437,
763
+ "grad_norm": 2.3790910243988037,
764
+ "learning_rate": 0.00019887793515109922,
765
+ "loss": 0.3589,
766
+ "step": 1080
767
+ },
768
+ {
769
+ "epoch": 0.017002558183066077,
770
+ "grad_norm": 8.939804077148438,
771
+ "learning_rate": 0.0001988675360329444,
772
+ "loss": 0.5794,
773
+ "step": 1090
774
+ },
775
+ {
776
+ "epoch": 0.017158544955387784,
777
+ "grad_norm": 7.628971576690674,
778
+ "learning_rate": 0.00019885713691478966,
779
+ "loss": 0.5317,
780
+ "step": 1100
781
+ },
782
+ {
783
+ "epoch": 0.01731453172770949,
784
+ "grad_norm": 1.8617459535598755,
785
+ "learning_rate": 0.00019884673779663485,
786
+ "loss": 0.433,
787
+ "step": 1110
788
+ },
789
+ {
790
+ "epoch": 0.017470518500031198,
791
+ "grad_norm": 6.928651809692383,
792
+ "learning_rate": 0.0001988363386784801,
793
+ "loss": 0.216,
794
+ "step": 1120
795
+ },
796
+ {
797
+ "epoch": 0.017626505272352905,
798
+ "grad_norm": 3.6206703186035156,
799
+ "learning_rate": 0.0001988259395603253,
800
+ "loss": 0.4963,
801
+ "step": 1130
802
+ },
803
+ {
804
+ "epoch": 0.01778249204467461,
805
+ "grad_norm": 4.615293025970459,
806
+ "learning_rate": 0.00019881554044217053,
807
+ "loss": 0.3579,
808
+ "step": 1140
809
+ },
810
+ {
811
+ "epoch": 0.01793847881699632,
812
+ "grad_norm": 3.3206489086151123,
813
+ "learning_rate": 0.00019880514132401573,
814
+ "loss": 0.3717,
815
+ "step": 1150
816
+ },
817
+ {
818
+ "epoch": 0.018094465589318025,
819
+ "grad_norm": 7.5789408683776855,
820
+ "learning_rate": 0.00019879474220586097,
821
+ "loss": 0.2646,
822
+ "step": 1160
823
+ },
824
+ {
825
+ "epoch": 0.018250452361639732,
826
+ "grad_norm": 4.467035293579102,
827
+ "learning_rate": 0.00019878434308770617,
828
+ "loss": 0.1804,
829
+ "step": 1170
830
+ },
831
+ {
832
+ "epoch": 0.01840643913396144,
833
+ "grad_norm": 9.318495750427246,
834
+ "learning_rate": 0.0001987739439695514,
835
+ "loss": 0.3257,
836
+ "step": 1180
837
+ },
838
+ {
839
+ "epoch": 0.018562425906283146,
840
+ "grad_norm": 3.1046931743621826,
841
+ "learning_rate": 0.0001987635448513966,
842
+ "loss": 0.4267,
843
+ "step": 1190
844
+ },
845
+ {
846
+ "epoch": 0.018718412678604853,
847
+ "grad_norm": 4.199711799621582,
848
+ "learning_rate": 0.00019875314573324185,
849
+ "loss": 0.6676,
850
+ "step": 1200
851
+ },
852
+ {
853
+ "epoch": 0.01887439945092656,
854
+ "grad_norm": 12.733593940734863,
855
+ "learning_rate": 0.00019874274661508705,
856
+ "loss": 0.5991,
857
+ "step": 1210
858
+ },
859
+ {
860
+ "epoch": 0.019030386223248267,
861
+ "grad_norm": 5.1266984939575195,
862
+ "learning_rate": 0.0001987323474969323,
863
+ "loss": 0.2945,
864
+ "step": 1220
865
+ },
866
+ {
867
+ "epoch": 0.019186372995569977,
868
+ "grad_norm": 6.397432804107666,
869
+ "learning_rate": 0.00019872194837877748,
870
+ "loss": 0.4774,
871
+ "step": 1230
872
+ },
873
+ {
874
+ "epoch": 0.019342359767891684,
875
+ "grad_norm": 3.761129379272461,
876
+ "learning_rate": 0.0001987115492606227,
877
+ "loss": 0.4345,
878
+ "step": 1240
879
+ },
880
+ {
881
+ "epoch": 0.01949834654021339,
882
+ "grad_norm": 2.7355475425720215,
883
+ "learning_rate": 0.00019870115014246792,
884
+ "loss": 0.497,
885
+ "step": 1250
886
+ },
887
+ {
888
+ "epoch": 0.019654333312535098,
889
+ "grad_norm": 2.7893755435943604,
890
+ "learning_rate": 0.00019869075102431314,
891
+ "loss": 0.2169,
892
+ "step": 1260
893
+ },
894
+ {
895
+ "epoch": 0.019810320084856805,
896
+ "grad_norm": 10.722111701965332,
897
+ "learning_rate": 0.00019868035190615836,
898
+ "loss": 0.3867,
899
+ "step": 1270
900
+ },
901
+ {
902
+ "epoch": 0.019966306857178512,
903
+ "grad_norm": 11.647889137268066,
904
+ "learning_rate": 0.00019866995278800358,
905
+ "loss": 0.4138,
906
+ "step": 1280
907
+ },
908
+ {
909
+ "epoch": 0.02012229362950022,
910
+ "grad_norm": 1.4313205480575562,
911
+ "learning_rate": 0.0001986595536698488,
912
+ "loss": 0.3581,
913
+ "step": 1290
914
+ },
915
+ {
916
+ "epoch": 0.020278280401821926,
917
+ "grad_norm": 5.193609714508057,
918
+ "learning_rate": 0.00019864915455169402,
919
+ "loss": 0.5259,
920
+ "step": 1300
921
+ },
922
+ {
923
+ "epoch": 0.020434267174143633,
924
+ "grad_norm": 3.1939520835876465,
925
+ "learning_rate": 0.00019863875543353924,
926
+ "loss": 0.5391,
927
+ "step": 1310
928
+ },
929
+ {
930
+ "epoch": 0.02059025394646534,
931
+ "grad_norm": 7.716986179351807,
932
+ "learning_rate": 0.00019862835631538446,
933
+ "loss": 0.4773,
934
+ "step": 1320
935
+ },
936
+ {
937
+ "epoch": 0.020746240718787046,
938
+ "grad_norm": 3.191506862640381,
939
+ "learning_rate": 0.00019861795719722968,
940
+ "loss": 0.6051,
941
+ "step": 1330
942
+ },
943
+ {
944
+ "epoch": 0.020902227491108753,
945
+ "grad_norm": 5.67583703994751,
946
+ "learning_rate": 0.0001986075580790749,
947
+ "loss": 0.2716,
948
+ "step": 1340
949
+ },
950
+ {
951
+ "epoch": 0.02105821426343046,
952
+ "grad_norm": 3.1657073497772217,
953
+ "learning_rate": 0.00019859715896092012,
954
+ "loss": 0.4838,
955
+ "step": 1350
956
+ },
957
+ {
958
+ "epoch": 0.021214201035752167,
959
+ "grad_norm": 5.144644260406494,
960
+ "learning_rate": 0.00019858675984276534,
961
+ "loss": 0.3185,
962
+ "step": 1360
963
+ },
964
+ {
965
+ "epoch": 0.021370187808073874,
966
+ "grad_norm": 0.3853607177734375,
967
+ "learning_rate": 0.00019857636072461056,
968
+ "loss": 0.21,
969
+ "step": 1370
970
+ },
971
+ {
972
+ "epoch": 0.02152617458039558,
973
+ "grad_norm": 15.329313278198242,
974
+ "learning_rate": 0.00019856596160645578,
975
+ "loss": 0.4216,
976
+ "step": 1380
977
+ },
978
+ {
979
+ "epoch": 0.02168216135271729,
980
+ "grad_norm": 1.9162156581878662,
981
+ "learning_rate": 0.000198555562488301,
982
+ "loss": 0.3766,
983
+ "step": 1390
984
+ },
985
+ {
986
+ "epoch": 0.021838148125038998,
987
+ "grad_norm": 2.5800933837890625,
988
+ "learning_rate": 0.00019854516337014622,
989
+ "loss": 0.4478,
990
+ "step": 1400
991
+ },
992
+ {
993
+ "epoch": 0.021994134897360705,
994
+ "grad_norm": 12.898444175720215,
995
+ "learning_rate": 0.00019853476425199144,
996
+ "loss": 0.7923,
997
+ "step": 1410
998
+ },
999
+ {
1000
+ "epoch": 0.022150121669682412,
1001
+ "grad_norm": 11.576621055603027,
1002
+ "learning_rate": 0.00019852436513383666,
1003
+ "loss": 0.3582,
1004
+ "step": 1420
1005
+ },
1006
+ {
1007
+ "epoch": 0.02230610844200412,
1008
+ "grad_norm": 10.80225658416748,
1009
+ "learning_rate": 0.00019851396601568188,
1010
+ "loss": 0.4698,
1011
+ "step": 1430
1012
+ },
1013
+ {
1014
+ "epoch": 0.022462095214325826,
1015
+ "grad_norm": 5.019324779510498,
1016
+ "learning_rate": 0.0001985035668975271,
1017
+ "loss": 0.4463,
1018
+ "step": 1440
1019
+ },
1020
+ {
1021
+ "epoch": 0.022618081986647533,
1022
+ "grad_norm": 5.366154670715332,
1023
+ "learning_rate": 0.00019849316777937232,
1024
+ "loss": 0.4288,
1025
+ "step": 1450
1026
+ },
1027
+ {
1028
+ "epoch": 0.02277406875896924,
1029
+ "grad_norm": 3.945873498916626,
1030
+ "learning_rate": 0.00019848276866121754,
1031
+ "loss": 0.5625,
1032
+ "step": 1460
1033
+ },
1034
+ {
1035
+ "epoch": 0.022930055531290947,
1036
+ "grad_norm": 1.8938435316085815,
1037
+ "learning_rate": 0.00019847236954306276,
1038
+ "loss": 0.4298,
1039
+ "step": 1470
1040
+ },
1041
+ {
1042
+ "epoch": 0.023086042303612653,
1043
+ "grad_norm": 7.040036678314209,
1044
+ "learning_rate": 0.00019846197042490798,
1045
+ "loss": 0.5806,
1046
+ "step": 1480
1047
+ },
1048
+ {
1049
+ "epoch": 0.02324202907593436,
1050
+ "grad_norm": 12.009136199951172,
1051
+ "learning_rate": 0.0001984515713067532,
1052
+ "loss": 0.3603,
1053
+ "step": 1490
1054
+ },
1055
+ {
1056
+ "epoch": 0.023398015848256067,
1057
+ "grad_norm": 3.157111883163452,
1058
+ "learning_rate": 0.00019844117218859841,
1059
+ "loss": 0.3811,
1060
+ "step": 1500
1061
+ },
1062
+ {
1063
+ "epoch": 0.023554002620577774,
1064
+ "grad_norm": 1.0227491855621338,
1065
+ "learning_rate": 0.00019843077307044363,
1066
+ "loss": 0.3459,
1067
+ "step": 1510
1068
+ },
1069
+ {
1070
+ "epoch": 0.02370998939289948,
1071
+ "grad_norm": 2.699263334274292,
1072
+ "learning_rate": 0.00019842037395228885,
1073
+ "loss": 0.4047,
1074
+ "step": 1520
1075
+ },
1076
+ {
1077
+ "epoch": 0.023865976165221188,
1078
+ "grad_norm": 3.879258871078491,
1079
+ "learning_rate": 0.00019840997483413407,
1080
+ "loss": 0.2083,
1081
+ "step": 1530
1082
+ },
1083
+ {
1084
+ "epoch": 0.024021962937542895,
1085
+ "grad_norm": 1.0468460321426392,
1086
+ "learning_rate": 0.0001983995757159793,
1087
+ "loss": 0.3146,
1088
+ "step": 1540
1089
+ },
1090
+ {
1091
+ "epoch": 0.024177949709864602,
1092
+ "grad_norm": 3.5738143920898438,
1093
+ "learning_rate": 0.0001983891765978245,
1094
+ "loss": 0.3181,
1095
+ "step": 1550
1096
+ },
1097
+ {
1098
+ "epoch": 0.024333936482186312,
1099
+ "grad_norm": 6.3198113441467285,
1100
+ "learning_rate": 0.00019837877747966973,
1101
+ "loss": 0.3196,
1102
+ "step": 1560
1103
+ },
1104
+ {
1105
+ "epoch": 0.02448992325450802,
1106
+ "grad_norm": 1.6974947452545166,
1107
+ "learning_rate": 0.00019836837836151495,
1108
+ "loss": 0.3832,
1109
+ "step": 1570
1110
+ },
1111
+ {
1112
+ "epoch": 0.024645910026829726,
1113
+ "grad_norm": 6.220627307891846,
1114
+ "learning_rate": 0.00019835797924336017,
1115
+ "loss": 0.3445,
1116
+ "step": 1580
1117
+ },
1118
+ {
1119
+ "epoch": 0.024801896799151433,
1120
+ "grad_norm": 3.9202497005462646,
1121
+ "learning_rate": 0.0001983475801252054,
1122
+ "loss": 0.4304,
1123
+ "step": 1590
1124
+ },
1125
+ {
1126
+ "epoch": 0.02495788357147314,
1127
+ "grad_norm": 2.2945311069488525,
1128
+ "learning_rate": 0.0001983371810070506,
1129
+ "loss": 0.4487,
1130
+ "step": 1600
1131
+ },
1132
+ {
1133
+ "epoch": 0.025113870343794847,
1134
+ "grad_norm": 9.262955665588379,
1135
+ "learning_rate": 0.00019832678188889583,
1136
+ "loss": 0.3691,
1137
+ "step": 1610
1138
+ },
1139
+ {
1140
+ "epoch": 0.025269857116116554,
1141
+ "grad_norm": 2.6905221939086914,
1142
+ "learning_rate": 0.00019831638277074105,
1143
+ "loss": 0.5117,
1144
+ "step": 1620
1145
+ },
1146
+ {
1147
+ "epoch": 0.02542584388843826,
1148
+ "grad_norm": 4.1330060958862305,
1149
+ "learning_rate": 0.00019830598365258627,
1150
+ "loss": 0.5721,
1151
+ "step": 1630
1152
+ },
1153
+ {
1154
+ "epoch": 0.025581830660759967,
1155
+ "grad_norm": 1.0681806802749634,
1156
+ "learning_rate": 0.0001982955845344315,
1157
+ "loss": 0.4794,
1158
+ "step": 1640
1159
+ },
1160
+ {
1161
+ "epoch": 0.025737817433081674,
1162
+ "grad_norm": 2.6171841621398926,
1163
+ "learning_rate": 0.0001982851854162767,
1164
+ "loss": 0.4596,
1165
+ "step": 1650
1166
+ },
1167
+ {
1168
+ "epoch": 0.02589380420540338,
1169
+ "grad_norm": 3.3451762199401855,
1170
+ "learning_rate": 0.00019827478629812193,
1171
+ "loss": 0.4745,
1172
+ "step": 1660
1173
+ },
1174
+ {
1175
+ "epoch": 0.026049790977725088,
1176
+ "grad_norm": 4.152349472045898,
1177
+ "learning_rate": 0.00019826438717996715,
1178
+ "loss": 0.4452,
1179
+ "step": 1670
1180
+ },
1181
+ {
1182
+ "epoch": 0.026205777750046795,
1183
+ "grad_norm": 2.7109954357147217,
1184
+ "learning_rate": 0.00019825398806181237,
1185
+ "loss": 0.3966,
1186
+ "step": 1680
1187
+ },
1188
+ {
1189
+ "epoch": 0.026361764522368502,
1190
+ "grad_norm": 5.731235980987549,
1191
+ "learning_rate": 0.0001982435889436576,
1192
+ "loss": 0.4423,
1193
+ "step": 1690
1194
+ },
1195
+ {
1196
+ "epoch": 0.02651775129469021,
1197
+ "grad_norm": 2.370852470397949,
1198
+ "learning_rate": 0.0001982331898255028,
1199
+ "loss": 0.5619,
1200
+ "step": 1700
1201
+ },
1202
+ {
1203
+ "epoch": 0.026673738067011916,
1204
+ "grad_norm": 2.5500876903533936,
1205
+ "learning_rate": 0.00019822279070734803,
1206
+ "loss": 0.2658,
1207
+ "step": 1710
1208
+ },
1209
+ {
1210
+ "epoch": 0.026829724839333626,
1211
+ "grad_norm": 0.20157204568386078,
1212
+ "learning_rate": 0.00019821239158919325,
1213
+ "loss": 0.2939,
1214
+ "step": 1720
1215
+ },
1216
+ {
1217
+ "epoch": 0.026985711611655333,
1218
+ "grad_norm": 2.6305084228515625,
1219
+ "learning_rate": 0.00019820199247103847,
1220
+ "loss": 0.5339,
1221
+ "step": 1730
1222
+ },
1223
+ {
1224
+ "epoch": 0.02714169838397704,
1225
+ "grad_norm": 6.241075038909912,
1226
+ "learning_rate": 0.00019819159335288369,
1227
+ "loss": 0.3625,
1228
+ "step": 1740
1229
+ },
1230
+ {
1231
+ "epoch": 0.027297685156298747,
1232
+ "grad_norm": 5.584851264953613,
1233
+ "learning_rate": 0.0001981811942347289,
1234
+ "loss": 0.6158,
1235
+ "step": 1750
1236
+ },
1237
+ {
1238
+ "epoch": 0.027453671928620454,
1239
+ "grad_norm": 6.756530284881592,
1240
+ "learning_rate": 0.00019817079511657413,
1241
+ "loss": 0.3513,
1242
+ "step": 1760
1243
+ },
1244
+ {
1245
+ "epoch": 0.02760965870094216,
1246
+ "grad_norm": 5.153111457824707,
1247
+ "learning_rate": 0.00019816039599841935,
1248
+ "loss": 0.2885,
1249
+ "step": 1770
1250
+ },
1251
+ {
1252
+ "epoch": 0.027765645473263868,
1253
+ "grad_norm": 6.222873210906982,
1254
+ "learning_rate": 0.00019814999688026456,
1255
+ "loss": 0.3291,
1256
+ "step": 1780
1257
+ },
1258
+ {
1259
+ "epoch": 0.027921632245585574,
1260
+ "grad_norm": 4.744391918182373,
1261
+ "learning_rate": 0.00019813959776210978,
1262
+ "loss": 0.5053,
1263
+ "step": 1790
1264
+ },
1265
+ {
1266
+ "epoch": 0.02807761901790728,
1267
+ "grad_norm": 3.7181131839752197,
1268
+ "learning_rate": 0.000198129198643955,
1269
+ "loss": 0.3983,
1270
+ "step": 1800
1271
+ },
1272
+ {
1273
+ "epoch": 0.02823360579022899,
1274
+ "grad_norm": 1.8923834562301636,
1275
+ "learning_rate": 0.00019811879952580022,
1276
+ "loss": 0.4465,
1277
+ "step": 1810
1278
+ },
1279
+ {
1280
+ "epoch": 0.028389592562550695,
1281
+ "grad_norm": 6.049956321716309,
1282
+ "learning_rate": 0.00019810840040764544,
1283
+ "loss": 0.5028,
1284
+ "step": 1820
1285
+ },
1286
+ {
1287
+ "epoch": 0.028545579334872402,
1288
+ "grad_norm": 8.034428596496582,
1289
+ "learning_rate": 0.00019809800128949066,
1290
+ "loss": 0.4799,
1291
+ "step": 1830
1292
+ },
1293
+ {
1294
+ "epoch": 0.02870156610719411,
1295
+ "grad_norm": 3.916890859603882,
1296
+ "learning_rate": 0.00019808760217133588,
1297
+ "loss": 0.3109,
1298
+ "step": 1840
1299
+ },
1300
+ {
1301
+ "epoch": 0.028857552879515816,
1302
+ "grad_norm": 3.351285457611084,
1303
+ "learning_rate": 0.0001980772030531811,
1304
+ "loss": 0.5981,
1305
+ "step": 1850
1306
+ },
1307
+ {
1308
+ "epoch": 0.029013539651837523,
1309
+ "grad_norm": 4.443714141845703,
1310
+ "learning_rate": 0.00019806680393502632,
1311
+ "loss": 0.3243,
1312
+ "step": 1860
1313
+ },
1314
+ {
1315
+ "epoch": 0.02916952642415923,
1316
+ "grad_norm": 0.6998550295829773,
1317
+ "learning_rate": 0.00019805640481687154,
1318
+ "loss": 0.2057,
1319
+ "step": 1870
1320
+ },
1321
+ {
1322
+ "epoch": 0.02932551319648094,
1323
+ "grad_norm": 12.252562522888184,
1324
+ "learning_rate": 0.00019804600569871676,
1325
+ "loss": 0.3806,
1326
+ "step": 1880
1327
+ },
1328
+ {
1329
+ "epoch": 0.029481499968802647,
1330
+ "grad_norm": 9.03868293762207,
1331
+ "learning_rate": 0.00019803560658056198,
1332
+ "loss": 0.3351,
1333
+ "step": 1890
1334
+ },
1335
+ {
1336
+ "epoch": 0.029637486741124354,
1337
+ "grad_norm": 2.3400044441223145,
1338
+ "learning_rate": 0.0001980252074624072,
1339
+ "loss": 0.3103,
1340
+ "step": 1900
1341
+ },
1342
+ {
1343
+ "epoch": 0.02979347351344606,
1344
+ "grad_norm": 2.399489164352417,
1345
+ "learning_rate": 0.00019801480834425242,
1346
+ "loss": 0.4792,
1347
+ "step": 1910
1348
+ },
1349
+ {
1350
+ "epoch": 0.029949460285767768,
1351
+ "grad_norm": 1.5152381658554077,
1352
+ "learning_rate": 0.00019800440922609764,
1353
+ "loss": 0.2873,
1354
+ "step": 1920
1355
+ },
1356
+ {
1357
+ "epoch": 0.030105447058089475,
1358
+ "grad_norm": 2.7013959884643555,
1359
+ "learning_rate": 0.00019799401010794286,
1360
+ "loss": 0.381,
1361
+ "step": 1930
1362
+ },
1363
+ {
1364
+ "epoch": 0.03026143383041118,
1365
+ "grad_norm": 0.7135261297225952,
1366
+ "learning_rate": 0.00019798361098978808,
1367
+ "loss": 0.4001,
1368
+ "step": 1940
1369
+ },
1370
+ {
1371
+ "epoch": 0.03041742060273289,
1372
+ "grad_norm": 15.61109447479248,
1373
+ "learning_rate": 0.0001979732118716333,
1374
+ "loss": 0.474,
1375
+ "step": 1950
1376
+ },
1377
+ {
1378
+ "epoch": 0.030573407375054595,
1379
+ "grad_norm": 3.236668348312378,
1380
+ "learning_rate": 0.00019796281275347852,
1381
+ "loss": 0.2149,
1382
+ "step": 1960
1383
+ },
1384
+ {
1385
+ "epoch": 0.030729394147376302,
1386
+ "grad_norm": 4.722929954528809,
1387
+ "learning_rate": 0.00019795241363532374,
1388
+ "loss": 0.5474,
1389
+ "step": 1970
1390
+ },
1391
+ {
1392
+ "epoch": 0.03088538091969801,
1393
+ "grad_norm": 6.002335548400879,
1394
+ "learning_rate": 0.00019794201451716896,
1395
+ "loss": 0.5196,
1396
+ "step": 1980
1397
+ },
1398
+ {
1399
+ "epoch": 0.031041367692019716,
1400
+ "grad_norm": 5.210022926330566,
1401
+ "learning_rate": 0.00019793161539901418,
1402
+ "loss": 0.4741,
1403
+ "step": 1990
1404
+ },
1405
+ {
1406
+ "epoch": 0.031197354464341423,
1407
+ "grad_norm": 7.984034538269043,
1408
+ "learning_rate": 0.0001979212162808594,
1409
+ "loss": 0.5399,
1410
+ "step": 2000
1411
+ },
1412
+ {
1413
+ "epoch": 0.03135334123666313,
1414
+ "grad_norm": 4.008991241455078,
1415
+ "learning_rate": 0.00019791081716270462,
1416
+ "loss": 0.4788,
1417
+ "step": 2010
1418
+ },
1419
+ {
1420
+ "epoch": 0.03150932800898484,
1421
+ "grad_norm": 4.395316123962402,
1422
+ "learning_rate": 0.00019790041804454984,
1423
+ "loss": 0.3456,
1424
+ "step": 2020
1425
+ },
1426
+ {
1427
+ "epoch": 0.03166531478130655,
1428
+ "grad_norm": 3.9784998893737793,
1429
+ "learning_rate": 0.00019789001892639506,
1430
+ "loss": 0.326,
1431
+ "step": 2030
1432
+ },
1433
+ {
1434
+ "epoch": 0.03182130155362825,
1435
+ "grad_norm": 5.040256023406982,
1436
+ "learning_rate": 0.00019787961980824028,
1437
+ "loss": 0.5264,
1438
+ "step": 2040
1439
+ },
1440
+ {
1441
+ "epoch": 0.03197728832594996,
1442
+ "grad_norm": 2.686203956604004,
1443
+ "learning_rate": 0.0001978692206900855,
1444
+ "loss": 0.5461,
1445
+ "step": 2050
1446
+ },
1447
+ {
1448
+ "epoch": 0.032133275098271664,
1449
+ "grad_norm": 1.8417892456054688,
1450
+ "learning_rate": 0.00019785882157193071,
1451
+ "loss": 0.4736,
1452
+ "step": 2060
1453
+ },
1454
+ {
1455
+ "epoch": 0.032289261870593375,
1456
+ "grad_norm": 3.623887300491333,
1457
+ "learning_rate": 0.00019784842245377593,
1458
+ "loss": 0.3742,
1459
+ "step": 2070
1460
+ },
1461
+ {
1462
+ "epoch": 0.03244524864291508,
1463
+ "grad_norm": 0.5184723734855652,
1464
+ "learning_rate": 0.00019783802333562115,
1465
+ "loss": 0.3941,
1466
+ "step": 2080
1467
+ },
1468
+ {
1469
+ "epoch": 0.03260123541523679,
1470
+ "grad_norm": 7.751046180725098,
1471
+ "learning_rate": 0.00019782762421746637,
1472
+ "loss": 0.4609,
1473
+ "step": 2090
1474
+ },
1475
+ {
1476
+ "epoch": 0.03275722218755849,
1477
+ "grad_norm": 2.547175168991089,
1478
+ "learning_rate": 0.0001978172250993116,
1479
+ "loss": 0.3089,
1480
+ "step": 2100
1481
+ },
1482
+ {
1483
+ "epoch": 0.0329132089598802,
1484
+ "grad_norm": 1.8453168869018555,
1485
+ "learning_rate": 0.0001978068259811568,
1486
+ "loss": 0.4371,
1487
+ "step": 2110
1488
+ },
1489
+ {
1490
+ "epoch": 0.033069195732201906,
1491
+ "grad_norm": 7.033441543579102,
1492
+ "learning_rate": 0.00019779642686300203,
1493
+ "loss": 0.3842,
1494
+ "step": 2120
1495
+ },
1496
+ {
1497
+ "epoch": 0.033225182504523616,
1498
+ "grad_norm": 2.2642929553985596,
1499
+ "learning_rate": 0.00019778602774484725,
1500
+ "loss": 0.2069,
1501
+ "step": 2130
1502
+ },
1503
+ {
1504
+ "epoch": 0.03338116927684533,
1505
+ "grad_norm": 4.043938636779785,
1506
+ "learning_rate": 0.00019777562862669244,
1507
+ "loss": 0.4148,
1508
+ "step": 2140
1509
+ },
1510
+ {
1511
+ "epoch": 0.03353715604916703,
1512
+ "grad_norm": 4.227573871612549,
1513
+ "learning_rate": 0.0001977652295085377,
1514
+ "loss": 0.4785,
1515
+ "step": 2150
1516
+ },
1517
+ {
1518
+ "epoch": 0.03369314282148874,
1519
+ "grad_norm": 2.0808868408203125,
1520
+ "learning_rate": 0.00019775483039038288,
1521
+ "loss": 0.3073,
1522
+ "step": 2160
1523
+ },
1524
+ {
1525
+ "epoch": 0.033849129593810444,
1526
+ "grad_norm": 4.518038749694824,
1527
+ "learning_rate": 0.00019774443127222813,
1528
+ "loss": 0.4197,
1529
+ "step": 2170
1530
+ },
1531
+ {
1532
+ "epoch": 0.034005116366132154,
1533
+ "grad_norm": 2.3731303215026855,
1534
+ "learning_rate": 0.00019773403215407332,
1535
+ "loss": 0.2753,
1536
+ "step": 2180
1537
+ },
1538
+ {
1539
+ "epoch": 0.03416110313845386,
1540
+ "grad_norm": 1.7613986730575562,
1541
+ "learning_rate": 0.00019772363303591857,
1542
+ "loss": 0.3834,
1543
+ "step": 2190
1544
+ },
1545
+ {
1546
+ "epoch": 0.03431708991077557,
1547
+ "grad_norm": 2.9518485069274902,
1548
+ "learning_rate": 0.00019771323391776376,
1549
+ "loss": 0.3544,
1550
+ "step": 2200
1551
+ },
1552
+ {
1553
+ "epoch": 0.03447307668309727,
1554
+ "grad_norm": 4.424034118652344,
1555
+ "learning_rate": 0.000197702834799609,
1556
+ "loss": 0.4046,
1557
+ "step": 2210
1558
+ },
1559
+ {
1560
+ "epoch": 0.03462906345541898,
1561
+ "grad_norm": 2.307690382003784,
1562
+ "learning_rate": 0.0001976924356814542,
1563
+ "loss": 0.1445,
1564
+ "step": 2220
1565
+ },
1566
+ {
1567
+ "epoch": 0.034785050227740685,
1568
+ "grad_norm": 2.536533832550049,
1569
+ "learning_rate": 0.00019768203656329945,
1570
+ "loss": 0.2815,
1571
+ "step": 2230
1572
+ },
1573
+ {
1574
+ "epoch": 0.034941037000062396,
1575
+ "grad_norm": 2.006728172302246,
1576
+ "learning_rate": 0.00019767163744514464,
1577
+ "loss": 0.5314,
1578
+ "step": 2240
1579
+ },
1580
+ {
1581
+ "epoch": 0.0350970237723841,
1582
+ "grad_norm": 4.729264259338379,
1583
+ "learning_rate": 0.0001976612383269899,
1584
+ "loss": 0.4547,
1585
+ "step": 2250
1586
+ },
1587
+ {
1588
+ "epoch": 0.03525301054470581,
1589
+ "grad_norm": 4.600561141967773,
1590
+ "learning_rate": 0.00019765083920883508,
1591
+ "loss": 0.3967,
1592
+ "step": 2260
1593
+ },
1594
+ {
1595
+ "epoch": 0.03540899731702751,
1596
+ "grad_norm": 0.6419815421104431,
1597
+ "learning_rate": 0.00019764044009068033,
1598
+ "loss": 0.336,
1599
+ "step": 2270
1600
+ },
1601
+ {
1602
+ "epoch": 0.03556498408934922,
1603
+ "grad_norm": 5.567695140838623,
1604
+ "learning_rate": 0.00019763004097252552,
1605
+ "loss": 0.5644,
1606
+ "step": 2280
1607
+ },
1608
+ {
1609
+ "epoch": 0.035720970861670934,
1610
+ "grad_norm": 0.5767773389816284,
1611
+ "learning_rate": 0.00019761964185437077,
1612
+ "loss": 0.2643,
1613
+ "step": 2290
1614
+ },
1615
+ {
1616
+ "epoch": 0.03587695763399264,
1617
+ "grad_norm": 4.91269063949585,
1618
+ "learning_rate": 0.00019760924273621596,
1619
+ "loss": 0.4198,
1620
+ "step": 2300
1621
+ },
1622
+ {
1623
+ "epoch": 0.03603294440631435,
1624
+ "grad_norm": 4.422239780426025,
1625
+ "learning_rate": 0.0001975988436180612,
1626
+ "loss": 0.5338,
1627
+ "step": 2310
1628
+ },
1629
+ {
1630
+ "epoch": 0.03618893117863605,
1631
+ "grad_norm": 1.672713041305542,
1632
+ "learning_rate": 0.0001975884444999064,
1633
+ "loss": 0.3214,
1634
+ "step": 2320
1635
+ },
1636
+ {
1637
+ "epoch": 0.03634491795095776,
1638
+ "grad_norm": 4.033807277679443,
1639
+ "learning_rate": 0.00019757804538175164,
1640
+ "loss": 0.3593,
1641
+ "step": 2330
1642
+ },
1643
+ {
1644
+ "epoch": 0.036500904723279465,
1645
+ "grad_norm": 3.198880910873413,
1646
+ "learning_rate": 0.00019756764626359684,
1647
+ "loss": 0.3657,
1648
+ "step": 2340
1649
+ },
1650
+ {
1651
+ "epoch": 0.036656891495601175,
1652
+ "grad_norm": 2.640443801879883,
1653
+ "learning_rate": 0.00019755724714544208,
1654
+ "loss": 0.3243,
1655
+ "step": 2350
1656
+ },
1657
+ {
1658
+ "epoch": 0.03681287826792288,
1659
+ "grad_norm": 1.562350869178772,
1660
+ "learning_rate": 0.00019754684802728728,
1661
+ "loss": 0.3362,
1662
+ "step": 2360
1663
+ },
1664
+ {
1665
+ "epoch": 0.03696886504024459,
1666
+ "grad_norm": 4.379998683929443,
1667
+ "learning_rate": 0.00019753644890913252,
1668
+ "loss": 0.541,
1669
+ "step": 2370
1670
+ },
1671
+ {
1672
+ "epoch": 0.03712485181256629,
1673
+ "grad_norm": 2.529301881790161,
1674
+ "learning_rate": 0.00019752604979097774,
1675
+ "loss": 0.5396,
1676
+ "step": 2380
1677
+ },
1678
+ {
1679
+ "epoch": 0.037280838584888,
1680
+ "grad_norm": 1.7864131927490234,
1681
+ "learning_rate": 0.00019751565067282296,
1682
+ "loss": 0.3165,
1683
+ "step": 2390
1684
+ },
1685
+ {
1686
+ "epoch": 0.037436825357209706,
1687
+ "grad_norm": 1.0566092729568481,
1688
+ "learning_rate": 0.00019750525155466818,
1689
+ "loss": 0.4225,
1690
+ "step": 2400
1691
+ },
1692
+ {
1693
+ "epoch": 0.03759281212953142,
1694
+ "grad_norm": 2.4530510902404785,
1695
+ "learning_rate": 0.0001974948524365134,
1696
+ "loss": 0.4281,
1697
+ "step": 2410
1698
+ },
1699
+ {
1700
+ "epoch": 0.03774879890185312,
1701
+ "grad_norm": 5.363385200500488,
1702
+ "learning_rate": 0.00019748445331835862,
1703
+ "loss": 0.4746,
1704
+ "step": 2420
1705
+ },
1706
+ {
1707
+ "epoch": 0.03790478567417483,
1708
+ "grad_norm": 6.038578510284424,
1709
+ "learning_rate": 0.00019747405420020384,
1710
+ "loss": 0.3182,
1711
+ "step": 2430
1712
+ },
1713
+ {
1714
+ "epoch": 0.038060772446496534,
1715
+ "grad_norm": 9.37709903717041,
1716
+ "learning_rate": 0.00019746365508204906,
1717
+ "loss": 0.3288,
1718
+ "step": 2440
1719
+ },
1720
+ {
1721
+ "epoch": 0.038216759218818244,
1722
+ "grad_norm": 28.81134033203125,
1723
+ "learning_rate": 0.00019745325596389428,
1724
+ "loss": 0.2646,
1725
+ "step": 2450
1726
+ },
1727
+ {
1728
+ "epoch": 0.038372745991139955,
1729
+ "grad_norm": 0.616172730922699,
1730
+ "learning_rate": 0.0001974428568457395,
1731
+ "loss": 0.3848,
1732
+ "step": 2460
1733
+ },
1734
+ {
1735
+ "epoch": 0.03852873276346166,
1736
+ "grad_norm": 2.9943106174468994,
1737
+ "learning_rate": 0.00019743245772758472,
1738
+ "loss": 0.4097,
1739
+ "step": 2470
1740
+ },
1741
+ {
1742
+ "epoch": 0.03868471953578337,
1743
+ "grad_norm": 4.001022815704346,
1744
+ "learning_rate": 0.00019742205860942994,
1745
+ "loss": 0.2359,
1746
+ "step": 2480
1747
+ },
1748
+ {
1749
+ "epoch": 0.03884070630810507,
1750
+ "grad_norm": 3.587982416152954,
1751
+ "learning_rate": 0.00019741165949127516,
1752
+ "loss": 0.5829,
1753
+ "step": 2490
1754
+ },
1755
+ {
1756
+ "epoch": 0.03899669308042678,
1757
+ "grad_norm": 1.906042218208313,
1758
+ "learning_rate": 0.00019740126037312038,
1759
+ "loss": 0.336,
1760
+ "step": 2500
1761
+ },
1762
+ {
1763
+ "epoch": 0.039152679852748486,
1764
+ "grad_norm": 9.536376953125,
1765
+ "learning_rate": 0.0001973908612549656,
1766
+ "loss": 0.3468,
1767
+ "step": 2510
1768
+ },
1769
+ {
1770
+ "epoch": 0.039308666625070196,
1771
+ "grad_norm": 1.582669734954834,
1772
+ "learning_rate": 0.00019738046213681082,
1773
+ "loss": 0.3534,
1774
+ "step": 2520
1775
+ },
1776
+ {
1777
+ "epoch": 0.0394646533973919,
1778
+ "grad_norm": 2.59112286567688,
1779
+ "learning_rate": 0.00019737006301865604,
1780
+ "loss": 0.239,
1781
+ "step": 2530
1782
+ },
1783
+ {
1784
+ "epoch": 0.03962064016971361,
1785
+ "grad_norm": 2.091029167175293,
1786
+ "learning_rate": 0.00019735966390050126,
1787
+ "loss": 0.5045,
1788
+ "step": 2540
1789
+ },
1790
+ {
1791
+ "epoch": 0.03977662694203531,
1792
+ "grad_norm": 5.665968894958496,
1793
+ "learning_rate": 0.00019734926478234648,
1794
+ "loss": 0.3729,
1795
+ "step": 2550
1796
+ },
1797
+ {
1798
+ "epoch": 0.039932613714357024,
1799
+ "grad_norm": 2.2713849544525146,
1800
+ "learning_rate": 0.0001973388656641917,
1801
+ "loss": 0.3598,
1802
+ "step": 2560
1803
+ },
1804
+ {
1805
+ "epoch": 0.04008860048667873,
1806
+ "grad_norm": 3.3854405879974365,
1807
+ "learning_rate": 0.00019732846654603692,
1808
+ "loss": 0.2485,
1809
+ "step": 2570
1810
+ },
1811
+ {
1812
+ "epoch": 0.04024458725900044,
1813
+ "grad_norm": 7.780918598175049,
1814
+ "learning_rate": 0.00019731806742788214,
1815
+ "loss": 0.3557,
1816
+ "step": 2580
1817
+ },
1818
+ {
1819
+ "epoch": 0.04040057403132214,
1820
+ "grad_norm": 2.7871644496917725,
1821
+ "learning_rate": 0.00019730766830972733,
1822
+ "loss": 0.7119,
1823
+ "step": 2590
1824
+ },
1825
+ {
1826
+ "epoch": 0.04055656080364385,
1827
+ "grad_norm": 2.53657603263855,
1828
+ "learning_rate": 0.00019729726919157258,
1829
+ "loss": 0.3575,
1830
+ "step": 2600
1831
+ },
1832
+ {
1833
+ "epoch": 0.040712547575965555,
1834
+ "grad_norm": 0.3270189166069031,
1835
+ "learning_rate": 0.00019728687007341777,
1836
+ "loss": 0.3733,
1837
+ "step": 2610
1838
+ },
1839
+ {
1840
+ "epoch": 0.040868534348287265,
1841
+ "grad_norm": 1.2280256748199463,
1842
+ "learning_rate": 0.00019727647095526301,
1843
+ "loss": 0.458,
1844
+ "step": 2620
1845
+ },
1846
+ {
1847
+ "epoch": 0.041024521120608975,
1848
+ "grad_norm": 0.7863756418228149,
1849
+ "learning_rate": 0.0001972660718371082,
1850
+ "loss": 0.6961,
1851
+ "step": 2630
1852
+ },
1853
+ {
1854
+ "epoch": 0.04118050789293068,
1855
+ "grad_norm": 3.562969207763672,
1856
+ "learning_rate": 0.00019725567271895345,
1857
+ "loss": 0.3823,
1858
+ "step": 2640
1859
+ },
1860
+ {
1861
+ "epoch": 0.04133649466525239,
1862
+ "grad_norm": 2.5304505825042725,
1863
+ "learning_rate": 0.00019724527360079865,
1864
+ "loss": 0.4938,
1865
+ "step": 2650
1866
+ },
1867
+ {
1868
+ "epoch": 0.04149248143757409,
1869
+ "grad_norm": 1.8894665241241455,
1870
+ "learning_rate": 0.0001972348744826439,
1871
+ "loss": 0.4397,
1872
+ "step": 2660
1873
+ },
1874
+ {
1875
+ "epoch": 0.0416484682098958,
1876
+ "grad_norm": 3.460160970687866,
1877
+ "learning_rate": 0.00019722447536448909,
1878
+ "loss": 0.3571,
1879
+ "step": 2670
1880
+ },
1881
+ {
1882
+ "epoch": 0.04180445498221751,
1883
+ "grad_norm": 4.1218061447143555,
1884
+ "learning_rate": 0.00019721407624633433,
1885
+ "loss": 0.2929,
1886
+ "step": 2680
1887
+ },
1888
+ {
1889
+ "epoch": 0.04196044175453922,
1890
+ "grad_norm": 2.9370386600494385,
1891
+ "learning_rate": 0.00019720367712817952,
1892
+ "loss": 0.4142,
1893
+ "step": 2690
1894
+ },
1895
+ {
1896
+ "epoch": 0.04211642852686092,
1897
+ "grad_norm": 5.4618401527404785,
1898
+ "learning_rate": 0.00019719327801002477,
1899
+ "loss": 0.3288,
1900
+ "step": 2700
1901
+ },
1902
+ {
1903
+ "epoch": 0.04227241529918263,
1904
+ "grad_norm": 3.1975386142730713,
1905
+ "learning_rate": 0.00019718287889186996,
1906
+ "loss": 0.3373,
1907
+ "step": 2710
1908
+ },
1909
+ {
1910
+ "epoch": 0.042428402071504334,
1911
+ "grad_norm": 4.580129623413086,
1912
+ "learning_rate": 0.0001971724797737152,
1913
+ "loss": 0.7003,
1914
+ "step": 2720
1915
+ },
1916
+ {
1917
+ "epoch": 0.042584388843826045,
1918
+ "grad_norm": 1.7264137268066406,
1919
+ "learning_rate": 0.0001971620806555604,
1920
+ "loss": 0.3375,
1921
+ "step": 2730
1922
+ },
1923
+ {
1924
+ "epoch": 0.04274037561614775,
1925
+ "grad_norm": 8.90623950958252,
1926
+ "learning_rate": 0.00019715168153740565,
1927
+ "loss": 0.4492,
1928
+ "step": 2740
1929
+ },
1930
+ {
1931
+ "epoch": 0.04289636238846946,
1932
+ "grad_norm": 8.407225608825684,
1933
+ "learning_rate": 0.00019714128241925084,
1934
+ "loss": 0.3827,
1935
+ "step": 2750
1936
+ },
1937
+ {
1938
+ "epoch": 0.04305234916079116,
1939
+ "grad_norm": 5.202430725097656,
1940
+ "learning_rate": 0.0001971308833010961,
1941
+ "loss": 0.2806,
1942
+ "step": 2760
1943
+ },
1944
+ {
1945
+ "epoch": 0.04320833593311287,
1946
+ "grad_norm": 1.9569600820541382,
1947
+ "learning_rate": 0.00019712048418294128,
1948
+ "loss": 0.2506,
1949
+ "step": 2770
1950
+ },
1951
+ {
1952
+ "epoch": 0.04336432270543458,
1953
+ "grad_norm": 3.2171106338500977,
1954
+ "learning_rate": 0.00019711008506478653,
1955
+ "loss": 0.3344,
1956
+ "step": 2780
1957
+ },
1958
+ {
1959
+ "epoch": 0.043520309477756286,
1960
+ "grad_norm": 0.7588744163513184,
1961
+ "learning_rate": 0.00019709968594663172,
1962
+ "loss": 0.3656,
1963
+ "step": 2790
1964
+ },
1965
+ {
1966
+ "epoch": 0.043676296250077996,
1967
+ "grad_norm": 1.4201767444610596,
1968
+ "learning_rate": 0.00019708928682847697,
1969
+ "loss": 0.4033,
1970
+ "step": 2800
1971
+ },
1972
+ {
1973
+ "epoch": 0.0438322830223997,
1974
+ "grad_norm": 2.439276933670044,
1975
+ "learning_rate": 0.00019707888771032216,
1976
+ "loss": 0.2828,
1977
+ "step": 2810
1978
+ },
1979
+ {
1980
+ "epoch": 0.04398826979472141,
1981
+ "grad_norm": 18.436410903930664,
1982
+ "learning_rate": 0.0001970684885921674,
1983
+ "loss": 0.6879,
1984
+ "step": 2820
1985
+ },
1986
+ {
1987
+ "epoch": 0.044144256567043114,
1988
+ "grad_norm": 3.1035194396972656,
1989
+ "learning_rate": 0.0001970580894740126,
1990
+ "loss": 0.3022,
1991
+ "step": 2830
1992
+ },
1993
+ {
1994
+ "epoch": 0.044300243339364824,
1995
+ "grad_norm": 4.849295139312744,
1996
+ "learning_rate": 0.00019704769035585785,
1997
+ "loss": 0.4078,
1998
+ "step": 2840
1999
+ },
2000
+ {
2001
+ "epoch": 0.04445623011168653,
2002
+ "grad_norm": 3.680732488632202,
2003
+ "learning_rate": 0.00019703729123770304,
2004
+ "loss": 0.4738,
2005
+ "step": 2850
2006
+ },
2007
+ {
2008
+ "epoch": 0.04461221688400824,
2009
+ "grad_norm": 2.8644354343414307,
2010
+ "learning_rate": 0.00019702689211954829,
2011
+ "loss": 0.386,
2012
+ "step": 2860
2013
+ },
2014
+ {
2015
+ "epoch": 0.04476820365632994,
2016
+ "grad_norm": 0.7621586918830872,
2017
+ "learning_rate": 0.00019701649300139348,
2018
+ "loss": 0.2032,
2019
+ "step": 2870
2020
+ },
2021
+ {
2022
+ "epoch": 0.04492419042865165,
2023
+ "grad_norm": 0.9556747078895569,
2024
+ "learning_rate": 0.00019700609388323873,
2025
+ "loss": 0.5366,
2026
+ "step": 2880
2027
+ },
2028
+ {
2029
+ "epoch": 0.045080177200973355,
2030
+ "grad_norm": 4.282146453857422,
2031
+ "learning_rate": 0.00019699569476508392,
2032
+ "loss": 0.4814,
2033
+ "step": 2890
2034
+ },
2035
+ {
2036
+ "epoch": 0.045236163973295065,
2037
+ "grad_norm": 1.905647873878479,
2038
+ "learning_rate": 0.00019698529564692916,
2039
+ "loss": 0.5273,
2040
+ "step": 2900
2041
+ },
2042
+ {
2043
+ "epoch": 0.04539215074561677,
2044
+ "grad_norm": 0.7890440821647644,
2045
+ "learning_rate": 0.00019697489652877436,
2046
+ "loss": 0.3434,
2047
+ "step": 2910
2048
+ },
2049
+ {
2050
+ "epoch": 0.04554813751793848,
2051
+ "grad_norm": 1.717529296875,
2052
+ "learning_rate": 0.0001969644974106196,
2053
+ "loss": 0.6094,
2054
+ "step": 2920
2055
+ },
2056
+ {
2057
+ "epoch": 0.04570412429026018,
2058
+ "grad_norm": 3.9639952182769775,
2059
+ "learning_rate": 0.0001969540982924648,
2060
+ "loss": 0.2958,
2061
+ "step": 2930
2062
+ },
2063
+ {
2064
+ "epoch": 0.04586011106258189,
2065
+ "grad_norm": 1.4309637546539307,
2066
+ "learning_rate": 0.00019694369917431004,
2067
+ "loss": 0.2879,
2068
+ "step": 2940
2069
+ },
2070
+ {
2071
+ "epoch": 0.0460160978349036,
2072
+ "grad_norm": 8.747053146362305,
2073
+ "learning_rate": 0.00019693330005615524,
2074
+ "loss": 0.5953,
2075
+ "step": 2950
2076
+ },
2077
+ {
2078
+ "epoch": 0.04617208460722531,
2079
+ "grad_norm": 0.027920352295041084,
2080
+ "learning_rate": 0.00019692290093800048,
2081
+ "loss": 0.2181,
2082
+ "step": 2960
2083
+ },
2084
+ {
2085
+ "epoch": 0.04632807137954702,
2086
+ "grad_norm": 1.9787837266921997,
2087
+ "learning_rate": 0.00019691250181984567,
2088
+ "loss": 0.2764,
2089
+ "step": 2970
2090
+ },
2091
+ {
2092
+ "epoch": 0.04648405815186872,
2093
+ "grad_norm": 10.669599533081055,
2094
+ "learning_rate": 0.00019690210270169092,
2095
+ "loss": 0.5824,
2096
+ "step": 2980
2097
+ },
2098
+ {
2099
+ "epoch": 0.04664004492419043,
2100
+ "grad_norm": 2.9555306434631348,
2101
+ "learning_rate": 0.00019689170358353611,
2102
+ "loss": 0.5993,
2103
+ "step": 2990
2104
+ },
2105
+ {
2106
+ "epoch": 0.046796031696512135,
2107
+ "grad_norm": 2.168975353240967,
2108
+ "learning_rate": 0.00019688130446538136,
2109
+ "loss": 0.2988,
2110
+ "step": 3000
2111
+ },
2112
+ {
2113
+ "epoch": 0.046952018468833845,
2114
+ "grad_norm": 1.4161510467529297,
2115
+ "learning_rate": 0.00019687090534722655,
2116
+ "loss": 0.2984,
2117
+ "step": 3010
2118
+ },
2119
+ {
2120
+ "epoch": 0.04710800524115555,
2121
+ "grad_norm": 4.282196521759033,
2122
+ "learning_rate": 0.0001968605062290718,
2123
+ "loss": 0.2822,
2124
+ "step": 3020
2125
+ },
2126
+ {
2127
+ "epoch": 0.04726399201347726,
2128
+ "grad_norm": 2.5621113777160645,
2129
+ "learning_rate": 0.000196850107110917,
2130
+ "loss": 0.3886,
2131
+ "step": 3030
2132
+ },
2133
+ {
2134
+ "epoch": 0.04741997878579896,
2135
+ "grad_norm": 0.7100577354431152,
2136
+ "learning_rate": 0.0001968397079927622,
2137
+ "loss": 0.2256,
2138
+ "step": 3040
2139
+ },
2140
+ {
2141
+ "epoch": 0.04757596555812067,
2142
+ "grad_norm": 2.2435333728790283,
2143
+ "learning_rate": 0.00019682930887460743,
2144
+ "loss": 0.2155,
2145
+ "step": 3050
2146
+ },
2147
+ {
2148
+ "epoch": 0.047731952330442376,
2149
+ "grad_norm": 1.9456003904342651,
2150
+ "learning_rate": 0.00019681890975645265,
2151
+ "loss": 0.2938,
2152
+ "step": 3060
2153
+ },
2154
+ {
2155
+ "epoch": 0.047887939102764086,
2156
+ "grad_norm": 8.851899147033691,
2157
+ "learning_rate": 0.00019680851063829787,
2158
+ "loss": 0.3822,
2159
+ "step": 3070
2160
+ },
2161
+ {
2162
+ "epoch": 0.04804392587508579,
2163
+ "grad_norm": 3.1571285724639893,
2164
+ "learning_rate": 0.0001967981115201431,
2165
+ "loss": 0.4913,
2166
+ "step": 3080
2167
+ },
2168
+ {
2169
+ "epoch": 0.0481999126474075,
2170
+ "grad_norm": 6.932368755340576,
2171
+ "learning_rate": 0.0001967877124019883,
2172
+ "loss": 0.2643,
2173
+ "step": 3090
2174
+ },
2175
+ {
2176
+ "epoch": 0.048355899419729204,
2177
+ "grad_norm": 4.671253204345703,
2178
+ "learning_rate": 0.00019677731328383353,
2179
+ "loss": 0.3736,
2180
+ "step": 3100
2181
+ },
2182
+ {
2183
+ "epoch": 0.048511886192050914,
2184
+ "grad_norm": 4.238051891326904,
2185
+ "learning_rate": 0.00019676691416567875,
2186
+ "loss": 0.2905,
2187
+ "step": 3110
2188
+ },
2189
+ {
2190
+ "epoch": 0.048667872964372624,
2191
+ "grad_norm": 4.185165882110596,
2192
+ "learning_rate": 0.00019675651504752397,
2193
+ "loss": 0.5799,
2194
+ "step": 3120
2195
+ },
2196
+ {
2197
+ "epoch": 0.04882385973669433,
2198
+ "grad_norm": 5.000885486602783,
2199
+ "learning_rate": 0.0001967461159293692,
2200
+ "loss": 0.5353,
2201
+ "step": 3130
2202
+ },
2203
+ {
2204
+ "epoch": 0.04897984650901604,
2205
+ "grad_norm": 2.1454551219940186,
2206
+ "learning_rate": 0.0001967357168112144,
2207
+ "loss": 0.2612,
2208
+ "step": 3140
2209
+ },
2210
+ {
2211
+ "epoch": 0.04913583328133774,
2212
+ "grad_norm": 5.727521896362305,
2213
+ "learning_rate": 0.00019672531769305963,
2214
+ "loss": 0.3584,
2215
+ "step": 3150
2216
+ },
2217
+ {
2218
+ "epoch": 0.04929182005365945,
2219
+ "grad_norm": 2.14788556098938,
2220
+ "learning_rate": 0.00019671491857490485,
2221
+ "loss": 0.248,
2222
+ "step": 3160
2223
+ },
2224
+ {
2225
+ "epoch": 0.049447806825981155,
2226
+ "grad_norm": 5.505538463592529,
2227
+ "learning_rate": 0.00019670451945675007,
2228
+ "loss": 0.5042,
2229
+ "step": 3170
2230
+ },
2231
+ {
2232
+ "epoch": 0.049603793598302866,
2233
+ "grad_norm": 0.45676565170288086,
2234
+ "learning_rate": 0.0001966941203385953,
2235
+ "loss": 0.3119,
2236
+ "step": 3180
2237
+ },
2238
+ {
2239
+ "epoch": 0.04975978037062457,
2240
+ "grad_norm": 2.446408748626709,
2241
+ "learning_rate": 0.0001966837212204405,
2242
+ "loss": 0.1164,
2243
+ "step": 3190
2244
+ },
2245
+ {
2246
+ "epoch": 0.04991576714294628,
2247
+ "grad_norm": 0.8342123031616211,
2248
+ "learning_rate": 0.00019667332210228573,
2249
+ "loss": 0.3749,
2250
+ "step": 3200
2251
+ },
2252
+ {
2253
+ "epoch": 0.05007175391526798,
2254
+ "grad_norm": 3.0448403358459473,
2255
+ "learning_rate": 0.00019666292298413095,
2256
+ "loss": 0.4866,
2257
+ "step": 3210
2258
+ },
2259
+ {
2260
+ "epoch": 0.05022774068758969,
2261
+ "grad_norm": 0.2389092594385147,
2262
+ "learning_rate": 0.00019665252386597617,
2263
+ "loss": 0.4043,
2264
+ "step": 3220
2265
+ },
2266
+ {
2267
+ "epoch": 0.0503837274599114,
2268
+ "grad_norm": 5.156858921051025,
2269
+ "learning_rate": 0.00019664212474782139,
2270
+ "loss": 0.2329,
2271
+ "step": 3230
2272
+ },
2273
+ {
2274
+ "epoch": 0.05053971423223311,
2275
+ "grad_norm": 3.149811029434204,
2276
+ "learning_rate": 0.0001966317256296666,
2277
+ "loss": 0.3312,
2278
+ "step": 3240
2279
+ },
2280
+ {
2281
+ "epoch": 0.05069570100455481,
2282
+ "grad_norm": 0.45096033811569214,
2283
+ "learning_rate": 0.00019662132651151185,
2284
+ "loss": 0.3754,
2285
+ "step": 3250
2286
+ },
2287
+ {
2288
+ "epoch": 0.05085168777687652,
2289
+ "grad_norm": 0.986663818359375,
2290
+ "learning_rate": 0.00019661092739335704,
2291
+ "loss": 0.3235,
2292
+ "step": 3260
2293
+ },
2294
+ {
2295
+ "epoch": 0.05100767454919823,
2296
+ "grad_norm": 2.9946937561035156,
2297
+ "learning_rate": 0.0001966005282752023,
2298
+ "loss": 0.2073,
2299
+ "step": 3270
2300
+ },
2301
+ {
2302
+ "epoch": 0.051163661321519935,
2303
+ "grad_norm": 3.995633125305176,
2304
+ "learning_rate": 0.00019659012915704748,
2305
+ "loss": 0.3481,
2306
+ "step": 3280
2307
+ },
2308
+ {
2309
+ "epoch": 0.051319648093841645,
2310
+ "grad_norm": 1.7635934352874756,
2311
+ "learning_rate": 0.00019657973003889273,
2312
+ "loss": 0.1994,
2313
+ "step": 3290
2314
+ },
2315
+ {
2316
+ "epoch": 0.05147563486616335,
2317
+ "grad_norm": 4.386302471160889,
2318
+ "learning_rate": 0.00019656933092073792,
2319
+ "loss": 0.3384,
2320
+ "step": 3300
2321
+ },
2322
+ {
2323
+ "epoch": 0.05163162163848506,
2324
+ "grad_norm": 2.7013816833496094,
2325
+ "learning_rate": 0.00019655893180258317,
2326
+ "loss": 0.2823,
2327
+ "step": 3310
2328
+ },
2329
+ {
2330
+ "epoch": 0.05178760841080676,
2331
+ "grad_norm": 7.045024871826172,
2332
+ "learning_rate": 0.00019654853268442836,
2333
+ "loss": 0.2112,
2334
+ "step": 3320
2335
+ },
2336
+ {
2337
+ "epoch": 0.05194359518312847,
2338
+ "grad_norm": 2.587613105773926,
2339
+ "learning_rate": 0.0001965381335662736,
2340
+ "loss": 0.3959,
2341
+ "step": 3330
2342
+ },
2343
+ {
2344
+ "epoch": 0.052099581955450176,
2345
+ "grad_norm": 0.9898849129676819,
2346
+ "learning_rate": 0.0001965277344481188,
2347
+ "loss": 0.3295,
2348
+ "step": 3340
2349
+ },
2350
+ {
2351
+ "epoch": 0.05225556872777189,
2352
+ "grad_norm": 8.89372444152832,
2353
+ "learning_rate": 0.00019651733532996405,
2354
+ "loss": 0.3443,
2355
+ "step": 3350
2356
+ },
2357
+ {
2358
+ "epoch": 0.05241155550009359,
2359
+ "grad_norm": 3.0839977264404297,
2360
+ "learning_rate": 0.00019650693621180924,
2361
+ "loss": 0.4376,
2362
+ "step": 3360
2363
+ },
2364
+ {
2365
+ "epoch": 0.0525675422724153,
2366
+ "grad_norm": 5.862309455871582,
2367
+ "learning_rate": 0.0001964965370936545,
2368
+ "loss": 0.5446,
2369
+ "step": 3370
2370
+ },
2371
+ {
2372
+ "epoch": 0.052723529044737004,
2373
+ "grad_norm": 5.274170398712158,
2374
+ "learning_rate": 0.00019648613797549968,
2375
+ "loss": 0.3987,
2376
+ "step": 3380
2377
+ },
2378
+ {
2379
+ "epoch": 0.052879515817058714,
2380
+ "grad_norm": 4.794046878814697,
2381
+ "learning_rate": 0.00019647573885734493,
2382
+ "loss": 0.5267,
2383
+ "step": 3390
2384
+ },
2385
+ {
2386
+ "epoch": 0.05303550258938042,
2387
+ "grad_norm": 1.6140331029891968,
2388
+ "learning_rate": 0.00019646533973919012,
2389
+ "loss": 0.6051,
2390
+ "step": 3400
2391
+ },
2392
+ {
2393
+ "epoch": 0.05319148936170213,
2394
+ "grad_norm": 6.158959865570068,
2395
+ "learning_rate": 0.00019645494062103537,
2396
+ "loss": 0.3056,
2397
+ "step": 3410
2398
+ },
2399
+ {
2400
+ "epoch": 0.05334747613402383,
2401
+ "grad_norm": 1.6302224397659302,
2402
+ "learning_rate": 0.00019644454150288056,
2403
+ "loss": 0.3473,
2404
+ "step": 3420
2405
+ },
2406
+ {
2407
+ "epoch": 0.05350346290634554,
2408
+ "grad_norm": 4.881719589233398,
2409
+ "learning_rate": 0.0001964341423847258,
2410
+ "loss": 0.4032,
2411
+ "step": 3430
2412
+ },
2413
+ {
2414
+ "epoch": 0.05365944967866725,
2415
+ "grad_norm": 1.1205174922943115,
2416
+ "learning_rate": 0.000196423743266571,
2417
+ "loss": 0.4288,
2418
+ "step": 3440
2419
+ },
2420
+ {
2421
+ "epoch": 0.053815436450988956,
2422
+ "grad_norm": 6.742442607879639,
2423
+ "learning_rate": 0.00019641334414841624,
2424
+ "loss": 0.3299,
2425
+ "step": 3450
2426
+ },
2427
+ {
2428
+ "epoch": 0.053971423223310666,
2429
+ "grad_norm": 0.43702423572540283,
2430
+ "learning_rate": 0.00019640294503026144,
2431
+ "loss": 0.4711,
2432
+ "step": 3460
2433
+ },
2434
+ {
2435
+ "epoch": 0.05412740999563237,
2436
+ "grad_norm": 0.10137466341257095,
2437
+ "learning_rate": 0.00019639254591210668,
2438
+ "loss": 0.3355,
2439
+ "step": 3470
2440
+ },
2441
+ {
2442
+ "epoch": 0.05428339676795408,
2443
+ "grad_norm": 3.7362818717956543,
2444
+ "learning_rate": 0.00019638214679395188,
2445
+ "loss": 0.3051,
2446
+ "step": 3480
2447
+ },
2448
+ {
2449
+ "epoch": 0.05443938354027578,
2450
+ "grad_norm": 2.639213800430298,
2451
+ "learning_rate": 0.0001963717476757971,
2452
+ "loss": 0.3746,
2453
+ "step": 3490
2454
+ },
2455
+ {
2456
+ "epoch": 0.054595370312597494,
2457
+ "grad_norm": 0.7321776151657104,
2458
+ "learning_rate": 0.00019636134855764232,
2459
+ "loss": 0.4572,
2460
+ "step": 3500
2461
+ }
2462
+ ],
2463
+ "logging_steps": 10,
2464
+ "max_steps": 192324,
2465
+ "num_input_tokens_seen": 0,
2466
+ "num_train_epochs": 3,
2467
+ "save_steps": 500,
2468
+ "stateful_callbacks": {
2469
+ "TrainerControl": {
2470
+ "args": {
2471
+ "should_epoch_stop": false,
2472
+ "should_evaluate": false,
2473
+ "should_log": false,
2474
+ "should_save": true,
2475
+ "should_training_stop": false
2476
+ },
2477
+ "attributes": {}
2478
+ }
2479
+ },
2480
+ "total_flos": 2131352616960000.0,
2481
+ "train_batch_size": 1,
2482
+ "trial_name": null,
2483
+ "trial_params": null
2484
+ }
HVU_QA/t5-viet-qg-finetuned/checkpoint-3500/training_args.bin ADDED
Binary file (5.71 kB). View file