Sabbir772 commited on
Commit
51f0e19
·
verified ·
1 Parent(s): 7105d58

Upload 11 files

Browse files
banglat5_bn_sy/added_tokens.json ADDED
@@ -0,0 +1,104 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "<BN>": 32100,
3
+ "<SY>": 32101,
4
+ "<extra_id_0>": 32099,
5
+ "<extra_id_10>": 32089,
6
+ "<extra_id_11>": 32088,
7
+ "<extra_id_12>": 32087,
8
+ "<extra_id_13>": 32086,
9
+ "<extra_id_14>": 32085,
10
+ "<extra_id_15>": 32084,
11
+ "<extra_id_16>": 32083,
12
+ "<extra_id_17>": 32082,
13
+ "<extra_id_18>": 32081,
14
+ "<extra_id_19>": 32080,
15
+ "<extra_id_1>": 32098,
16
+ "<extra_id_20>": 32079,
17
+ "<extra_id_21>": 32078,
18
+ "<extra_id_22>": 32077,
19
+ "<extra_id_23>": 32076,
20
+ "<extra_id_24>": 32075,
21
+ "<extra_id_25>": 32074,
22
+ "<extra_id_26>": 32073,
23
+ "<extra_id_27>": 32072,
24
+ "<extra_id_28>": 32071,
25
+ "<extra_id_29>": 32070,
26
+ "<extra_id_2>": 32097,
27
+ "<extra_id_30>": 32069,
28
+ "<extra_id_31>": 32068,
29
+ "<extra_id_32>": 32067,
30
+ "<extra_id_33>": 32066,
31
+ "<extra_id_34>": 32065,
32
+ "<extra_id_35>": 32064,
33
+ "<extra_id_36>": 32063,
34
+ "<extra_id_37>": 32062,
35
+ "<extra_id_38>": 32061,
36
+ "<extra_id_39>": 32060,
37
+ "<extra_id_3>": 32096,
38
+ "<extra_id_40>": 32059,
39
+ "<extra_id_41>": 32058,
40
+ "<extra_id_42>": 32057,
41
+ "<extra_id_43>": 32056,
42
+ "<extra_id_44>": 32055,
43
+ "<extra_id_45>": 32054,
44
+ "<extra_id_46>": 32053,
45
+ "<extra_id_47>": 32052,
46
+ "<extra_id_48>": 32051,
47
+ "<extra_id_49>": 32050,
48
+ "<extra_id_4>": 32095,
49
+ "<extra_id_50>": 32049,
50
+ "<extra_id_51>": 32048,
51
+ "<extra_id_52>": 32047,
52
+ "<extra_id_53>": 32046,
53
+ "<extra_id_54>": 32045,
54
+ "<extra_id_55>": 32044,
55
+ "<extra_id_56>": 32043,
56
+ "<extra_id_57>": 32042,
57
+ "<extra_id_58>": 32041,
58
+ "<extra_id_59>": 32040,
59
+ "<extra_id_5>": 32094,
60
+ "<extra_id_60>": 32039,
61
+ "<extra_id_61>": 32038,
62
+ "<extra_id_62>": 32037,
63
+ "<extra_id_63>": 32036,
64
+ "<extra_id_64>": 32035,
65
+ "<extra_id_65>": 32034,
66
+ "<extra_id_66>": 32033,
67
+ "<extra_id_67>": 32032,
68
+ "<extra_id_68>": 32031,
69
+ "<extra_id_69>": 32030,
70
+ "<extra_id_6>": 32093,
71
+ "<extra_id_70>": 32029,
72
+ "<extra_id_71>": 32028,
73
+ "<extra_id_72>": 32027,
74
+ "<extra_id_73>": 32026,
75
+ "<extra_id_74>": 32025,
76
+ "<extra_id_75>": 32024,
77
+ "<extra_id_76>": 32023,
78
+ "<extra_id_77>": 32022,
79
+ "<extra_id_78>": 32021,
80
+ "<extra_id_79>": 32020,
81
+ "<extra_id_7>": 32092,
82
+ "<extra_id_80>": 32019,
83
+ "<extra_id_81>": 32018,
84
+ "<extra_id_82>": 32017,
85
+ "<extra_id_83>": 32016,
86
+ "<extra_id_84>": 32015,
87
+ "<extra_id_85>": 32014,
88
+ "<extra_id_86>": 32013,
89
+ "<extra_id_87>": 32012,
90
+ "<extra_id_88>": 32011,
91
+ "<extra_id_89>": 32010,
92
+ "<extra_id_8>": 32091,
93
+ "<extra_id_90>": 32009,
94
+ "<extra_id_91>": 32008,
95
+ "<extra_id_92>": 32007,
96
+ "<extra_id_93>": 32006,
97
+ "<extra_id_94>": 32005,
98
+ "<extra_id_95>": 32004,
99
+ "<extra_id_96>": 32003,
100
+ "<extra_id_97>": 32002,
101
+ "<extra_id_98>": 32001,
102
+ "<extra_id_99>": 32000,
103
+ "<extra_id_9>": 32090
104
+ }
banglat5_bn_sy/config.json ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "T5ForConditionalGeneration"
4
+ ],
5
+ "classifier_dropout": 0.0,
6
+ "d_ff": 2048,
7
+ "d_kv": 64,
8
+ "d_model": 768,
9
+ "decoder_start_token_id": 0,
10
+ "dense_act_fn": "gelu_new",
11
+ "dropout_rate": 0.1,
12
+ "eos_token_id": 1,
13
+ "feed_forward_proj": "gated-gelu",
14
+ "gradient_checkpointing": false,
15
+ "initializer_factor": 1.0,
16
+ "is_encoder_decoder": true,
17
+ "is_gated_act": true,
18
+ "layer_norm_epsilon": 1e-06,
19
+ "model_type": "t5",
20
+ "num_decoder_layers": 12,
21
+ "num_heads": 12,
22
+ "num_layers": 12,
23
+ "output_past": true,
24
+ "pad_token_id": 0,
25
+ "relative_attention_max_distance": 128,
26
+ "relative_attention_num_buckets": 32,
27
+ "tie_word_embeddings": false,
28
+ "torch_dtype": "float32",
29
+ "transformers_version": "4.53.2",
30
+ "use_cache": true,
31
+ "vocab_size": 32102
32
+ }
banglat5_bn_sy/generation_config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "decoder_start_token_id": 0,
4
+ "eos_token_id": 1,
5
+ "pad_token_id": 0,
6
+ "transformers_version": "4.53.2"
7
+ }
banglat5_bn_sy/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c0c08bdec294120524778324692cb170923e4ac4f359c3d017321abd76c9a6a7
3
+ size 990185320
banglat5_bn_sy/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e40c669cb61982c6b67d2b4f99a5c9e7aae11ccd91a27f1f3f3d382850ffa45e
3
+ size 14244
banglat5_bn_sy/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:28e90bb157421a7ddece7bade835f66189ad650ccd4d98491ab10f79351d0d29
3
+ size 1064
banglat5_bn_sy/special_tokens_map.json ADDED
@@ -0,0 +1,127 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<BN>",
4
+ "<SY>",
5
+ "<extra_id_0>",
6
+ "<extra_id_1>",
7
+ "<extra_id_2>",
8
+ "<extra_id_3>",
9
+ "<extra_id_4>",
10
+ "<extra_id_5>",
11
+ "<extra_id_6>",
12
+ "<extra_id_7>",
13
+ "<extra_id_8>",
14
+ "<extra_id_9>",
15
+ "<extra_id_10>",
16
+ "<extra_id_11>",
17
+ "<extra_id_12>",
18
+ "<extra_id_13>",
19
+ "<extra_id_14>",
20
+ "<extra_id_15>",
21
+ "<extra_id_16>",
22
+ "<extra_id_17>",
23
+ "<extra_id_18>",
24
+ "<extra_id_19>",
25
+ "<extra_id_20>",
26
+ "<extra_id_21>",
27
+ "<extra_id_22>",
28
+ "<extra_id_23>",
29
+ "<extra_id_24>",
30
+ "<extra_id_25>",
31
+ "<extra_id_26>",
32
+ "<extra_id_27>",
33
+ "<extra_id_28>",
34
+ "<extra_id_29>",
35
+ "<extra_id_30>",
36
+ "<extra_id_31>",
37
+ "<extra_id_32>",
38
+ "<extra_id_33>",
39
+ "<extra_id_34>",
40
+ "<extra_id_35>",
41
+ "<extra_id_36>",
42
+ "<extra_id_37>",
43
+ "<extra_id_38>",
44
+ "<extra_id_39>",
45
+ "<extra_id_40>",
46
+ "<extra_id_41>",
47
+ "<extra_id_42>",
48
+ "<extra_id_43>",
49
+ "<extra_id_44>",
50
+ "<extra_id_45>",
51
+ "<extra_id_46>",
52
+ "<extra_id_47>",
53
+ "<extra_id_48>",
54
+ "<extra_id_49>",
55
+ "<extra_id_50>",
56
+ "<extra_id_51>",
57
+ "<extra_id_52>",
58
+ "<extra_id_53>",
59
+ "<extra_id_54>",
60
+ "<extra_id_55>",
61
+ "<extra_id_56>",
62
+ "<extra_id_57>",
63
+ "<extra_id_58>",
64
+ "<extra_id_59>",
65
+ "<extra_id_60>",
66
+ "<extra_id_61>",
67
+ "<extra_id_62>",
68
+ "<extra_id_63>",
69
+ "<extra_id_64>",
70
+ "<extra_id_65>",
71
+ "<extra_id_66>",
72
+ "<extra_id_67>",
73
+ "<extra_id_68>",
74
+ "<extra_id_69>",
75
+ "<extra_id_70>",
76
+ "<extra_id_71>",
77
+ "<extra_id_72>",
78
+ "<extra_id_73>",
79
+ "<extra_id_74>",
80
+ "<extra_id_75>",
81
+ "<extra_id_76>",
82
+ "<extra_id_77>",
83
+ "<extra_id_78>",
84
+ "<extra_id_79>",
85
+ "<extra_id_80>",
86
+ "<extra_id_81>",
87
+ "<extra_id_82>",
88
+ "<extra_id_83>",
89
+ "<extra_id_84>",
90
+ "<extra_id_85>",
91
+ "<extra_id_86>",
92
+ "<extra_id_87>",
93
+ "<extra_id_88>",
94
+ "<extra_id_89>",
95
+ "<extra_id_90>",
96
+ "<extra_id_91>",
97
+ "<extra_id_92>",
98
+ "<extra_id_93>",
99
+ "<extra_id_94>",
100
+ "<extra_id_95>",
101
+ "<extra_id_96>",
102
+ "<extra_id_97>",
103
+ "<extra_id_98>",
104
+ "<extra_id_99>"
105
+ ],
106
+ "eos_token": {
107
+ "content": "</s>",
108
+ "lstrip": false,
109
+ "normalized": false,
110
+ "rstrip": false,
111
+ "single_word": false
112
+ },
113
+ "pad_token": {
114
+ "content": "<pad>",
115
+ "lstrip": false,
116
+ "normalized": false,
117
+ "rstrip": false,
118
+ "single_word": false
119
+ },
120
+ "unk_token": {
121
+ "content": "<unk>",
122
+ "lstrip": false,
123
+ "normalized": false,
124
+ "rstrip": false,
125
+ "single_word": false
126
+ }
127
+ }
banglat5_bn_sy/spiece.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7dcab96935a2a51b1461c84e44c952ea8a3640c8bc3e2c6ae7a21d855454ae27
3
+ size 1111492
banglat5_bn_sy/tokenizer_config.json ADDED
@@ -0,0 +1,959 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": true,
3
+ "added_tokens_decoder": {
4
+ "0": {
5
+ "content": "<pad>",
6
+ "lstrip": false,
7
+ "normalized": false,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "1": {
13
+ "content": "</s>",
14
+ "lstrip": false,
15
+ "normalized": false,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
+ },
20
+ "2": {
21
+ "content": "<unk>",
22
+ "lstrip": false,
23
+ "normalized": false,
24
+ "rstrip": false,
25
+ "single_word": false,
26
+ "special": true
27
+ },
28
+ "32000": {
29
+ "content": "<extra_id_99>",
30
+ "lstrip": true,
31
+ "normalized": false,
32
+ "rstrip": true,
33
+ "single_word": false,
34
+ "special": true
35
+ },
36
+ "32001": {
37
+ "content": "<extra_id_98>",
38
+ "lstrip": true,
39
+ "normalized": false,
40
+ "rstrip": true,
41
+ "single_word": false,
42
+ "special": true
43
+ },
44
+ "32002": {
45
+ "content": "<extra_id_97>",
46
+ "lstrip": true,
47
+ "normalized": false,
48
+ "rstrip": true,
49
+ "single_word": false,
50
+ "special": true
51
+ },
52
+ "32003": {
53
+ "content": "<extra_id_96>",
54
+ "lstrip": true,
55
+ "normalized": false,
56
+ "rstrip": true,
57
+ "single_word": false,
58
+ "special": true
59
+ },
60
+ "32004": {
61
+ "content": "<extra_id_95>",
62
+ "lstrip": true,
63
+ "normalized": false,
64
+ "rstrip": true,
65
+ "single_word": false,
66
+ "special": true
67
+ },
68
+ "32005": {
69
+ "content": "<extra_id_94>",
70
+ "lstrip": true,
71
+ "normalized": false,
72
+ "rstrip": true,
73
+ "single_word": false,
74
+ "special": true
75
+ },
76
+ "32006": {
77
+ "content": "<extra_id_93>",
78
+ "lstrip": true,
79
+ "normalized": false,
80
+ "rstrip": true,
81
+ "single_word": false,
82
+ "special": true
83
+ },
84
+ "32007": {
85
+ "content": "<extra_id_92>",
86
+ "lstrip": true,
87
+ "normalized": false,
88
+ "rstrip": true,
89
+ "single_word": false,
90
+ "special": true
91
+ },
92
+ "32008": {
93
+ "content": "<extra_id_91>",
94
+ "lstrip": true,
95
+ "normalized": false,
96
+ "rstrip": true,
97
+ "single_word": false,
98
+ "special": true
99
+ },
100
+ "32009": {
101
+ "content": "<extra_id_90>",
102
+ "lstrip": true,
103
+ "normalized": false,
104
+ "rstrip": true,
105
+ "single_word": false,
106
+ "special": true
107
+ },
108
+ "32010": {
109
+ "content": "<extra_id_89>",
110
+ "lstrip": true,
111
+ "normalized": false,
112
+ "rstrip": true,
113
+ "single_word": false,
114
+ "special": true
115
+ },
116
+ "32011": {
117
+ "content": "<extra_id_88>",
118
+ "lstrip": true,
119
+ "normalized": false,
120
+ "rstrip": true,
121
+ "single_word": false,
122
+ "special": true
123
+ },
124
+ "32012": {
125
+ "content": "<extra_id_87>",
126
+ "lstrip": true,
127
+ "normalized": false,
128
+ "rstrip": true,
129
+ "single_word": false,
130
+ "special": true
131
+ },
132
+ "32013": {
133
+ "content": "<extra_id_86>",
134
+ "lstrip": true,
135
+ "normalized": false,
136
+ "rstrip": true,
137
+ "single_word": false,
138
+ "special": true
139
+ },
140
+ "32014": {
141
+ "content": "<extra_id_85>",
142
+ "lstrip": true,
143
+ "normalized": false,
144
+ "rstrip": true,
145
+ "single_word": false,
146
+ "special": true
147
+ },
148
+ "32015": {
149
+ "content": "<extra_id_84>",
150
+ "lstrip": true,
151
+ "normalized": false,
152
+ "rstrip": true,
153
+ "single_word": false,
154
+ "special": true
155
+ },
156
+ "32016": {
157
+ "content": "<extra_id_83>",
158
+ "lstrip": true,
159
+ "normalized": false,
160
+ "rstrip": true,
161
+ "single_word": false,
162
+ "special": true
163
+ },
164
+ "32017": {
165
+ "content": "<extra_id_82>",
166
+ "lstrip": true,
167
+ "normalized": false,
168
+ "rstrip": true,
169
+ "single_word": false,
170
+ "special": true
171
+ },
172
+ "32018": {
173
+ "content": "<extra_id_81>",
174
+ "lstrip": true,
175
+ "normalized": false,
176
+ "rstrip": true,
177
+ "single_word": false,
178
+ "special": true
179
+ },
180
+ "32019": {
181
+ "content": "<extra_id_80>",
182
+ "lstrip": true,
183
+ "normalized": false,
184
+ "rstrip": true,
185
+ "single_word": false,
186
+ "special": true
187
+ },
188
+ "32020": {
189
+ "content": "<extra_id_79>",
190
+ "lstrip": true,
191
+ "normalized": false,
192
+ "rstrip": true,
193
+ "single_word": false,
194
+ "special": true
195
+ },
196
+ "32021": {
197
+ "content": "<extra_id_78>",
198
+ "lstrip": true,
199
+ "normalized": false,
200
+ "rstrip": true,
201
+ "single_word": false,
202
+ "special": true
203
+ },
204
+ "32022": {
205
+ "content": "<extra_id_77>",
206
+ "lstrip": true,
207
+ "normalized": false,
208
+ "rstrip": true,
209
+ "single_word": false,
210
+ "special": true
211
+ },
212
+ "32023": {
213
+ "content": "<extra_id_76>",
214
+ "lstrip": true,
215
+ "normalized": false,
216
+ "rstrip": true,
217
+ "single_word": false,
218
+ "special": true
219
+ },
220
+ "32024": {
221
+ "content": "<extra_id_75>",
222
+ "lstrip": true,
223
+ "normalized": false,
224
+ "rstrip": true,
225
+ "single_word": false,
226
+ "special": true
227
+ },
228
+ "32025": {
229
+ "content": "<extra_id_74>",
230
+ "lstrip": true,
231
+ "normalized": false,
232
+ "rstrip": true,
233
+ "single_word": false,
234
+ "special": true
235
+ },
236
+ "32026": {
237
+ "content": "<extra_id_73>",
238
+ "lstrip": true,
239
+ "normalized": false,
240
+ "rstrip": true,
241
+ "single_word": false,
242
+ "special": true
243
+ },
244
+ "32027": {
245
+ "content": "<extra_id_72>",
246
+ "lstrip": true,
247
+ "normalized": false,
248
+ "rstrip": true,
249
+ "single_word": false,
250
+ "special": true
251
+ },
252
+ "32028": {
253
+ "content": "<extra_id_71>",
254
+ "lstrip": true,
255
+ "normalized": false,
256
+ "rstrip": true,
257
+ "single_word": false,
258
+ "special": true
259
+ },
260
+ "32029": {
261
+ "content": "<extra_id_70>",
262
+ "lstrip": true,
263
+ "normalized": false,
264
+ "rstrip": true,
265
+ "single_word": false,
266
+ "special": true
267
+ },
268
+ "32030": {
269
+ "content": "<extra_id_69>",
270
+ "lstrip": true,
271
+ "normalized": false,
272
+ "rstrip": true,
273
+ "single_word": false,
274
+ "special": true
275
+ },
276
+ "32031": {
277
+ "content": "<extra_id_68>",
278
+ "lstrip": true,
279
+ "normalized": false,
280
+ "rstrip": true,
281
+ "single_word": false,
282
+ "special": true
283
+ },
284
+ "32032": {
285
+ "content": "<extra_id_67>",
286
+ "lstrip": true,
287
+ "normalized": false,
288
+ "rstrip": true,
289
+ "single_word": false,
290
+ "special": true
291
+ },
292
+ "32033": {
293
+ "content": "<extra_id_66>",
294
+ "lstrip": true,
295
+ "normalized": false,
296
+ "rstrip": true,
297
+ "single_word": false,
298
+ "special": true
299
+ },
300
+ "32034": {
301
+ "content": "<extra_id_65>",
302
+ "lstrip": true,
303
+ "normalized": false,
304
+ "rstrip": true,
305
+ "single_word": false,
306
+ "special": true
307
+ },
308
+ "32035": {
309
+ "content": "<extra_id_64>",
310
+ "lstrip": true,
311
+ "normalized": false,
312
+ "rstrip": true,
313
+ "single_word": false,
314
+ "special": true
315
+ },
316
+ "32036": {
317
+ "content": "<extra_id_63>",
318
+ "lstrip": true,
319
+ "normalized": false,
320
+ "rstrip": true,
321
+ "single_word": false,
322
+ "special": true
323
+ },
324
+ "32037": {
325
+ "content": "<extra_id_62>",
326
+ "lstrip": true,
327
+ "normalized": false,
328
+ "rstrip": true,
329
+ "single_word": false,
330
+ "special": true
331
+ },
332
+ "32038": {
333
+ "content": "<extra_id_61>",
334
+ "lstrip": true,
335
+ "normalized": false,
336
+ "rstrip": true,
337
+ "single_word": false,
338
+ "special": true
339
+ },
340
+ "32039": {
341
+ "content": "<extra_id_60>",
342
+ "lstrip": true,
343
+ "normalized": false,
344
+ "rstrip": true,
345
+ "single_word": false,
346
+ "special": true
347
+ },
348
+ "32040": {
349
+ "content": "<extra_id_59>",
350
+ "lstrip": true,
351
+ "normalized": false,
352
+ "rstrip": true,
353
+ "single_word": false,
354
+ "special": true
355
+ },
356
+ "32041": {
357
+ "content": "<extra_id_58>",
358
+ "lstrip": true,
359
+ "normalized": false,
360
+ "rstrip": true,
361
+ "single_word": false,
362
+ "special": true
363
+ },
364
+ "32042": {
365
+ "content": "<extra_id_57>",
366
+ "lstrip": true,
367
+ "normalized": false,
368
+ "rstrip": true,
369
+ "single_word": false,
370
+ "special": true
371
+ },
372
+ "32043": {
373
+ "content": "<extra_id_56>",
374
+ "lstrip": true,
375
+ "normalized": false,
376
+ "rstrip": true,
377
+ "single_word": false,
378
+ "special": true
379
+ },
380
+ "32044": {
381
+ "content": "<extra_id_55>",
382
+ "lstrip": true,
383
+ "normalized": false,
384
+ "rstrip": true,
385
+ "single_word": false,
386
+ "special": true
387
+ },
388
+ "32045": {
389
+ "content": "<extra_id_54>",
390
+ "lstrip": true,
391
+ "normalized": false,
392
+ "rstrip": true,
393
+ "single_word": false,
394
+ "special": true
395
+ },
396
+ "32046": {
397
+ "content": "<extra_id_53>",
398
+ "lstrip": true,
399
+ "normalized": false,
400
+ "rstrip": true,
401
+ "single_word": false,
402
+ "special": true
403
+ },
404
+ "32047": {
405
+ "content": "<extra_id_52>",
406
+ "lstrip": true,
407
+ "normalized": false,
408
+ "rstrip": true,
409
+ "single_word": false,
410
+ "special": true
411
+ },
412
+ "32048": {
413
+ "content": "<extra_id_51>",
414
+ "lstrip": true,
415
+ "normalized": false,
416
+ "rstrip": true,
417
+ "single_word": false,
418
+ "special": true
419
+ },
420
+ "32049": {
421
+ "content": "<extra_id_50>",
422
+ "lstrip": true,
423
+ "normalized": false,
424
+ "rstrip": true,
425
+ "single_word": false,
426
+ "special": true
427
+ },
428
+ "32050": {
429
+ "content": "<extra_id_49>",
430
+ "lstrip": true,
431
+ "normalized": false,
432
+ "rstrip": true,
433
+ "single_word": false,
434
+ "special": true
435
+ },
436
+ "32051": {
437
+ "content": "<extra_id_48>",
438
+ "lstrip": true,
439
+ "normalized": false,
440
+ "rstrip": true,
441
+ "single_word": false,
442
+ "special": true
443
+ },
444
+ "32052": {
445
+ "content": "<extra_id_47>",
446
+ "lstrip": true,
447
+ "normalized": false,
448
+ "rstrip": true,
449
+ "single_word": false,
450
+ "special": true
451
+ },
452
+ "32053": {
453
+ "content": "<extra_id_46>",
454
+ "lstrip": true,
455
+ "normalized": false,
456
+ "rstrip": true,
457
+ "single_word": false,
458
+ "special": true
459
+ },
460
+ "32054": {
461
+ "content": "<extra_id_45>",
462
+ "lstrip": true,
463
+ "normalized": false,
464
+ "rstrip": true,
465
+ "single_word": false,
466
+ "special": true
467
+ },
468
+ "32055": {
469
+ "content": "<extra_id_44>",
470
+ "lstrip": true,
471
+ "normalized": false,
472
+ "rstrip": true,
473
+ "single_word": false,
474
+ "special": true
475
+ },
476
+ "32056": {
477
+ "content": "<extra_id_43>",
478
+ "lstrip": true,
479
+ "normalized": false,
480
+ "rstrip": true,
481
+ "single_word": false,
482
+ "special": true
483
+ },
484
+ "32057": {
485
+ "content": "<extra_id_42>",
486
+ "lstrip": true,
487
+ "normalized": false,
488
+ "rstrip": true,
489
+ "single_word": false,
490
+ "special": true
491
+ },
492
+ "32058": {
493
+ "content": "<extra_id_41>",
494
+ "lstrip": true,
495
+ "normalized": false,
496
+ "rstrip": true,
497
+ "single_word": false,
498
+ "special": true
499
+ },
500
+ "32059": {
501
+ "content": "<extra_id_40>",
502
+ "lstrip": true,
503
+ "normalized": false,
504
+ "rstrip": true,
505
+ "single_word": false,
506
+ "special": true
507
+ },
508
+ "32060": {
509
+ "content": "<extra_id_39>",
510
+ "lstrip": true,
511
+ "normalized": false,
512
+ "rstrip": true,
513
+ "single_word": false,
514
+ "special": true
515
+ },
516
+ "32061": {
517
+ "content": "<extra_id_38>",
518
+ "lstrip": true,
519
+ "normalized": false,
520
+ "rstrip": true,
521
+ "single_word": false,
522
+ "special": true
523
+ },
524
+ "32062": {
525
+ "content": "<extra_id_37>",
526
+ "lstrip": true,
527
+ "normalized": false,
528
+ "rstrip": true,
529
+ "single_word": false,
530
+ "special": true
531
+ },
532
+ "32063": {
533
+ "content": "<extra_id_36>",
534
+ "lstrip": true,
535
+ "normalized": false,
536
+ "rstrip": true,
537
+ "single_word": false,
538
+ "special": true
539
+ },
540
+ "32064": {
541
+ "content": "<extra_id_35>",
542
+ "lstrip": true,
543
+ "normalized": false,
544
+ "rstrip": true,
545
+ "single_word": false,
546
+ "special": true
547
+ },
548
+ "32065": {
549
+ "content": "<extra_id_34>",
550
+ "lstrip": true,
551
+ "normalized": false,
552
+ "rstrip": true,
553
+ "single_word": false,
554
+ "special": true
555
+ },
556
+ "32066": {
557
+ "content": "<extra_id_33>",
558
+ "lstrip": true,
559
+ "normalized": false,
560
+ "rstrip": true,
561
+ "single_word": false,
562
+ "special": true
563
+ },
564
+ "32067": {
565
+ "content": "<extra_id_32>",
566
+ "lstrip": true,
567
+ "normalized": false,
568
+ "rstrip": true,
569
+ "single_word": false,
570
+ "special": true
571
+ },
572
+ "32068": {
573
+ "content": "<extra_id_31>",
574
+ "lstrip": true,
575
+ "normalized": false,
576
+ "rstrip": true,
577
+ "single_word": false,
578
+ "special": true
579
+ },
580
+ "32069": {
581
+ "content": "<extra_id_30>",
582
+ "lstrip": true,
583
+ "normalized": false,
584
+ "rstrip": true,
585
+ "single_word": false,
586
+ "special": true
587
+ },
588
+ "32070": {
589
+ "content": "<extra_id_29>",
590
+ "lstrip": true,
591
+ "normalized": false,
592
+ "rstrip": true,
593
+ "single_word": false,
594
+ "special": true
595
+ },
596
+ "32071": {
597
+ "content": "<extra_id_28>",
598
+ "lstrip": true,
599
+ "normalized": false,
600
+ "rstrip": true,
601
+ "single_word": false,
602
+ "special": true
603
+ },
604
+ "32072": {
605
+ "content": "<extra_id_27>",
606
+ "lstrip": true,
607
+ "normalized": false,
608
+ "rstrip": true,
609
+ "single_word": false,
610
+ "special": true
611
+ },
612
+ "32073": {
613
+ "content": "<extra_id_26>",
614
+ "lstrip": true,
615
+ "normalized": false,
616
+ "rstrip": true,
617
+ "single_word": false,
618
+ "special": true
619
+ },
620
+ "32074": {
621
+ "content": "<extra_id_25>",
622
+ "lstrip": true,
623
+ "normalized": false,
624
+ "rstrip": true,
625
+ "single_word": false,
626
+ "special": true
627
+ },
628
+ "32075": {
629
+ "content": "<extra_id_24>",
630
+ "lstrip": true,
631
+ "normalized": false,
632
+ "rstrip": true,
633
+ "single_word": false,
634
+ "special": true
635
+ },
636
+ "32076": {
637
+ "content": "<extra_id_23>",
638
+ "lstrip": true,
639
+ "normalized": false,
640
+ "rstrip": true,
641
+ "single_word": false,
642
+ "special": true
643
+ },
644
+ "32077": {
645
+ "content": "<extra_id_22>",
646
+ "lstrip": true,
647
+ "normalized": false,
648
+ "rstrip": true,
649
+ "single_word": false,
650
+ "special": true
651
+ },
652
+ "32078": {
653
+ "content": "<extra_id_21>",
654
+ "lstrip": true,
655
+ "normalized": false,
656
+ "rstrip": true,
657
+ "single_word": false,
658
+ "special": true
659
+ },
660
+ "32079": {
661
+ "content": "<extra_id_20>",
662
+ "lstrip": true,
663
+ "normalized": false,
664
+ "rstrip": true,
665
+ "single_word": false,
666
+ "special": true
667
+ },
668
+ "32080": {
669
+ "content": "<extra_id_19>",
670
+ "lstrip": true,
671
+ "normalized": false,
672
+ "rstrip": true,
673
+ "single_word": false,
674
+ "special": true
675
+ },
676
+ "32081": {
677
+ "content": "<extra_id_18>",
678
+ "lstrip": true,
679
+ "normalized": false,
680
+ "rstrip": true,
681
+ "single_word": false,
682
+ "special": true
683
+ },
684
+ "32082": {
685
+ "content": "<extra_id_17>",
686
+ "lstrip": true,
687
+ "normalized": false,
688
+ "rstrip": true,
689
+ "single_word": false,
690
+ "special": true
691
+ },
692
+ "32083": {
693
+ "content": "<extra_id_16>",
694
+ "lstrip": true,
695
+ "normalized": false,
696
+ "rstrip": true,
697
+ "single_word": false,
698
+ "special": true
699
+ },
700
+ "32084": {
701
+ "content": "<extra_id_15>",
702
+ "lstrip": true,
703
+ "normalized": false,
704
+ "rstrip": true,
705
+ "single_word": false,
706
+ "special": true
707
+ },
708
+ "32085": {
709
+ "content": "<extra_id_14>",
710
+ "lstrip": true,
711
+ "normalized": false,
712
+ "rstrip": true,
713
+ "single_word": false,
714
+ "special": true
715
+ },
716
+ "32086": {
717
+ "content": "<extra_id_13>",
718
+ "lstrip": true,
719
+ "normalized": false,
720
+ "rstrip": true,
721
+ "single_word": false,
722
+ "special": true
723
+ },
724
+ "32087": {
725
+ "content": "<extra_id_12>",
726
+ "lstrip": true,
727
+ "normalized": false,
728
+ "rstrip": true,
729
+ "single_word": false,
730
+ "special": true
731
+ },
732
+ "32088": {
733
+ "content": "<extra_id_11>",
734
+ "lstrip": true,
735
+ "normalized": false,
736
+ "rstrip": true,
737
+ "single_word": false,
738
+ "special": true
739
+ },
740
+ "32089": {
741
+ "content": "<extra_id_10>",
742
+ "lstrip": true,
743
+ "normalized": false,
744
+ "rstrip": true,
745
+ "single_word": false,
746
+ "special": true
747
+ },
748
+ "32090": {
749
+ "content": "<extra_id_9>",
750
+ "lstrip": true,
751
+ "normalized": false,
752
+ "rstrip": true,
753
+ "single_word": false,
754
+ "special": true
755
+ },
756
+ "32091": {
757
+ "content": "<extra_id_8>",
758
+ "lstrip": true,
759
+ "normalized": false,
760
+ "rstrip": true,
761
+ "single_word": false,
762
+ "special": true
763
+ },
764
+ "32092": {
765
+ "content": "<extra_id_7>",
766
+ "lstrip": true,
767
+ "normalized": false,
768
+ "rstrip": true,
769
+ "single_word": false,
770
+ "special": true
771
+ },
772
+ "32093": {
773
+ "content": "<extra_id_6>",
774
+ "lstrip": true,
775
+ "normalized": false,
776
+ "rstrip": true,
777
+ "single_word": false,
778
+ "special": true
779
+ },
780
+ "32094": {
781
+ "content": "<extra_id_5>",
782
+ "lstrip": true,
783
+ "normalized": false,
784
+ "rstrip": true,
785
+ "single_word": false,
786
+ "special": true
787
+ },
788
+ "32095": {
789
+ "content": "<extra_id_4>",
790
+ "lstrip": true,
791
+ "normalized": false,
792
+ "rstrip": true,
793
+ "single_word": false,
794
+ "special": true
795
+ },
796
+ "32096": {
797
+ "content": "<extra_id_3>",
798
+ "lstrip": true,
799
+ "normalized": false,
800
+ "rstrip": true,
801
+ "single_word": false,
802
+ "special": true
803
+ },
804
+ "32097": {
805
+ "content": "<extra_id_2>",
806
+ "lstrip": true,
807
+ "normalized": false,
808
+ "rstrip": true,
809
+ "single_word": false,
810
+ "special": true
811
+ },
812
+ "32098": {
813
+ "content": "<extra_id_1>",
814
+ "lstrip": true,
815
+ "normalized": false,
816
+ "rstrip": true,
817
+ "single_word": false,
818
+ "special": true
819
+ },
820
+ "32099": {
821
+ "content": "<extra_id_0>",
822
+ "lstrip": true,
823
+ "normalized": false,
824
+ "rstrip": true,
825
+ "single_word": false,
826
+ "special": true
827
+ },
828
+ "32100": {
829
+ "content": "<BN>",
830
+ "lstrip": false,
831
+ "normalized": false,
832
+ "rstrip": false,
833
+ "single_word": false,
834
+ "special": true
835
+ },
836
+ "32101": {
837
+ "content": "<SY>",
838
+ "lstrip": false,
839
+ "normalized": false,
840
+ "rstrip": false,
841
+ "single_word": false,
842
+ "special": true
843
+ }
844
+ },
845
+ "additional_special_tokens": [
846
+ "<BN>",
847
+ "<SY>",
848
+ "<extra_id_0>",
849
+ "<extra_id_1>",
850
+ "<extra_id_2>",
851
+ "<extra_id_3>",
852
+ "<extra_id_4>",
853
+ "<extra_id_5>",
854
+ "<extra_id_6>",
855
+ "<extra_id_7>",
856
+ "<extra_id_8>",
857
+ "<extra_id_9>",
858
+ "<extra_id_10>",
859
+ "<extra_id_11>",
860
+ "<extra_id_12>",
861
+ "<extra_id_13>",
862
+ "<extra_id_14>",
863
+ "<extra_id_15>",
864
+ "<extra_id_16>",
865
+ "<extra_id_17>",
866
+ "<extra_id_18>",
867
+ "<extra_id_19>",
868
+ "<extra_id_20>",
869
+ "<extra_id_21>",
870
+ "<extra_id_22>",
871
+ "<extra_id_23>",
872
+ "<extra_id_24>",
873
+ "<extra_id_25>",
874
+ "<extra_id_26>",
875
+ "<extra_id_27>",
876
+ "<extra_id_28>",
877
+ "<extra_id_29>",
878
+ "<extra_id_30>",
879
+ "<extra_id_31>",
880
+ "<extra_id_32>",
881
+ "<extra_id_33>",
882
+ "<extra_id_34>",
883
+ "<extra_id_35>",
884
+ "<extra_id_36>",
885
+ "<extra_id_37>",
886
+ "<extra_id_38>",
887
+ "<extra_id_39>",
888
+ "<extra_id_40>",
889
+ "<extra_id_41>",
890
+ "<extra_id_42>",
891
+ "<extra_id_43>",
892
+ "<extra_id_44>",
893
+ "<extra_id_45>",
894
+ "<extra_id_46>",
895
+ "<extra_id_47>",
896
+ "<extra_id_48>",
897
+ "<extra_id_49>",
898
+ "<extra_id_50>",
899
+ "<extra_id_51>",
900
+ "<extra_id_52>",
901
+ "<extra_id_53>",
902
+ "<extra_id_54>",
903
+ "<extra_id_55>",
904
+ "<extra_id_56>",
905
+ "<extra_id_57>",
906
+ "<extra_id_58>",
907
+ "<extra_id_59>",
908
+ "<extra_id_60>",
909
+ "<extra_id_61>",
910
+ "<extra_id_62>",
911
+ "<extra_id_63>",
912
+ "<extra_id_64>",
913
+ "<extra_id_65>",
914
+ "<extra_id_66>",
915
+ "<extra_id_67>",
916
+ "<extra_id_68>",
917
+ "<extra_id_69>",
918
+ "<extra_id_70>",
919
+ "<extra_id_71>",
920
+ "<extra_id_72>",
921
+ "<extra_id_73>",
922
+ "<extra_id_74>",
923
+ "<extra_id_75>",
924
+ "<extra_id_76>",
925
+ "<extra_id_77>",
926
+ "<extra_id_78>",
927
+ "<extra_id_79>",
928
+ "<extra_id_80>",
929
+ "<extra_id_81>",
930
+ "<extra_id_82>",
931
+ "<extra_id_83>",
932
+ "<extra_id_84>",
933
+ "<extra_id_85>",
934
+ "<extra_id_86>",
935
+ "<extra_id_87>",
936
+ "<extra_id_88>",
937
+ "<extra_id_89>",
938
+ "<extra_id_90>",
939
+ "<extra_id_91>",
940
+ "<extra_id_92>",
941
+ "<extra_id_93>",
942
+ "<extra_id_94>",
943
+ "<extra_id_95>",
944
+ "<extra_id_96>",
945
+ "<extra_id_97>",
946
+ "<extra_id_98>",
947
+ "<extra_id_99>"
948
+ ],
949
+ "clean_up_tokenization_spaces": false,
950
+ "eos_token": "</s>",
951
+ "extra_ids": 100,
952
+ "extra_special_tokens": {},
953
+ "legacy": true,
954
+ "model_max_length": 512,
955
+ "pad_token": "<pad>",
956
+ "sp_model_kwargs": {},
957
+ "tokenizer_class": "T5Tokenizer",
958
+ "unk_token": "<unk>"
959
+ }
banglat5_bn_sy/trainer_state.json ADDED
@@ -0,0 +1,1700 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 14.0,
6
+ "eval_steps": 500,
7
+ "global_step": 22274,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.06285355122564425,
14
+ "grad_norm": 608.9674682617188,
15
+ "learning_rate": 4.844437460716531e-05,
16
+ "loss": 14.2524,
17
+ "step": 100
18
+ },
19
+ {
20
+ "epoch": 0.1257071024512885,
21
+ "grad_norm": 34.65327453613281,
22
+ "learning_rate": 4.6873035826524205e-05,
23
+ "loss": 10.3562,
24
+ "step": 200
25
+ },
26
+ {
27
+ "epoch": 0.18856065367693275,
28
+ "grad_norm": 21.24808120727539,
29
+ "learning_rate": 4.5301697045883096e-05,
30
+ "loss": 7.8551,
31
+ "step": 300
32
+ },
33
+ {
34
+ "epoch": 0.251414204902577,
35
+ "grad_norm": 17.404918670654297,
36
+ "learning_rate": 4.373035826524199e-05,
37
+ "loss": 6.6346,
38
+ "step": 400
39
+ },
40
+ {
41
+ "epoch": 0.3142677561282212,
42
+ "grad_norm": 12.713433265686035,
43
+ "learning_rate": 4.2159019484600884e-05,
44
+ "loss": 5.9755,
45
+ "step": 500
46
+ },
47
+ {
48
+ "epoch": 0.3771213073538655,
49
+ "grad_norm": 10.050477981567383,
50
+ "learning_rate": 4.0587680703959775e-05,
51
+ "loss": 5.5595,
52
+ "step": 600
53
+ },
54
+ {
55
+ "epoch": 0.43997485857950974,
56
+ "grad_norm": 13.709216117858887,
57
+ "learning_rate": 3.9016341923318666e-05,
58
+ "loss": 5.2853,
59
+ "step": 700
60
+ },
61
+ {
62
+ "epoch": 0.502828409805154,
63
+ "grad_norm": 9.112940788269043,
64
+ "learning_rate": 3.744500314267756e-05,
65
+ "loss": 5.1417,
66
+ "step": 800
67
+ },
68
+ {
69
+ "epoch": 0.5656819610307983,
70
+ "grad_norm": 8.267425537109375,
71
+ "learning_rate": 3.587366436203646e-05,
72
+ "loss": 4.9615,
73
+ "step": 900
74
+ },
75
+ {
76
+ "epoch": 0.6285355122564424,
77
+ "grad_norm": 9.709076881408691,
78
+ "learning_rate": 3.430232558139535e-05,
79
+ "loss": 4.6907,
80
+ "step": 1000
81
+ },
82
+ {
83
+ "epoch": 0.6913890634820867,
84
+ "grad_norm": 845.80859375,
85
+ "learning_rate": 3.273098680075424e-05,
86
+ "loss": 4.5456,
87
+ "step": 1100
88
+ },
89
+ {
90
+ "epoch": 0.754242614707731,
91
+ "grad_norm": 5.943735599517822,
92
+ "learning_rate": 3.115964802011313e-05,
93
+ "loss": 4.4291,
94
+ "step": 1200
95
+ },
96
+ {
97
+ "epoch": 0.8170961659333752,
98
+ "grad_norm": 5.8759989738464355,
99
+ "learning_rate": 2.9588309239472034e-05,
100
+ "loss": 4.3252,
101
+ "step": 1300
102
+ },
103
+ {
104
+ "epoch": 0.8799497171590195,
105
+ "grad_norm": 14.995753288269043,
106
+ "learning_rate": 2.8016970458830928e-05,
107
+ "loss": 4.2586,
108
+ "step": 1400
109
+ },
110
+ {
111
+ "epoch": 0.9428032683846638,
112
+ "grad_norm": 23.3351993560791,
113
+ "learning_rate": 2.644563167818982e-05,
114
+ "loss": 4.1372,
115
+ "step": 1500
116
+ },
117
+ {
118
+ "epoch": 1.0,
119
+ "eval_loss": 3.215750217437744,
120
+ "eval_runtime": 19.7611,
121
+ "eval_samples_per_second": 48.479,
122
+ "eval_steps_per_second": 6.073,
123
+ "step": 1591
124
+ },
125
+ {
126
+ "epoch": 1.005656819610308,
127
+ "grad_norm": 8.584565162658691,
128
+ "learning_rate": 2.4874292897548713e-05,
129
+ "loss": 4.0272,
130
+ "step": 1600
131
+ },
132
+ {
133
+ "epoch": 1.0685103708359522,
134
+ "grad_norm": 6.45043420791626,
135
+ "learning_rate": 2.3302954116907607e-05,
136
+ "loss": 3.9602,
137
+ "step": 1700
138
+ },
139
+ {
140
+ "epoch": 1.1313639220615965,
141
+ "grad_norm": 6.03476095199585,
142
+ "learning_rate": 2.17316153362665e-05,
143
+ "loss": 3.9052,
144
+ "step": 1800
145
+ },
146
+ {
147
+ "epoch": 1.1942174732872408,
148
+ "grad_norm": 5.746309280395508,
149
+ "learning_rate": 2.0160276555625392e-05,
150
+ "loss": 3.9282,
151
+ "step": 1900
152
+ },
153
+ {
154
+ "epoch": 1.2570710245128849,
155
+ "grad_norm": 8.062549591064453,
156
+ "learning_rate": 1.858893777498429e-05,
157
+ "loss": 3.8096,
158
+ "step": 2000
159
+ },
160
+ {
161
+ "epoch": 1.3199245757385292,
162
+ "grad_norm": 8.58310317993164,
163
+ "learning_rate": 1.701759899434318e-05,
164
+ "loss": 3.803,
165
+ "step": 2100
166
+ },
167
+ {
168
+ "epoch": 1.3827781269641735,
169
+ "grad_norm": 7.599905490875244,
170
+ "learning_rate": 1.5446260213702074e-05,
171
+ "loss": 3.8381,
172
+ "step": 2200
173
+ },
174
+ {
175
+ "epoch": 1.4456316781898177,
176
+ "grad_norm": 22.772512435913086,
177
+ "learning_rate": 1.3874921433060969e-05,
178
+ "loss": 3.6456,
179
+ "step": 2300
180
+ },
181
+ {
182
+ "epoch": 1.508485229415462,
183
+ "grad_norm": 6.949570178985596,
184
+ "learning_rate": 1.2303582652419863e-05,
185
+ "loss": 3.7442,
186
+ "step": 2400
187
+ },
188
+ {
189
+ "epoch": 1.5713387806411063,
190
+ "grad_norm": 5.7536821365356445,
191
+ "learning_rate": 1.0732243871778757e-05,
192
+ "loss": 3.691,
193
+ "step": 2500
194
+ },
195
+ {
196
+ "epoch": 1.6341923318667506,
197
+ "grad_norm": 55.64060974121094,
198
+ "learning_rate": 9.160905091137651e-06,
199
+ "loss": 3.7461,
200
+ "step": 2600
201
+ },
202
+ {
203
+ "epoch": 1.6970458830923947,
204
+ "grad_norm": 6.573077201843262,
205
+ "learning_rate": 7.589566310496543e-06,
206
+ "loss": 3.6186,
207
+ "step": 2700
208
+ },
209
+ {
210
+ "epoch": 1.759899434318039,
211
+ "grad_norm": 8.615326881408691,
212
+ "learning_rate": 6.018227529855437e-06,
213
+ "loss": 3.6546,
214
+ "step": 2800
215
+ },
216
+ {
217
+ "epoch": 1.8227529855436833,
218
+ "grad_norm": 6.359428405761719,
219
+ "learning_rate": 4.446888749214331e-06,
220
+ "loss": 3.5724,
221
+ "step": 2900
222
+ },
223
+ {
224
+ "epoch": 1.8856065367693273,
225
+ "grad_norm": 5.5190582275390625,
226
+ "learning_rate": 2.8755499685732243e-06,
227
+ "loss": 3.6164,
228
+ "step": 3000
229
+ },
230
+ {
231
+ "epoch": 1.9484600879949716,
232
+ "grad_norm": 5.9382004737854,
233
+ "learning_rate": 1.3042111879321182e-06,
234
+ "loss": 3.52,
235
+ "step": 3100
236
+ },
237
+ {
238
+ "epoch": 2.0,
239
+ "eval_loss": 2.803544521331787,
240
+ "eval_runtime": 19.8643,
241
+ "eval_samples_per_second": 48.227,
242
+ "eval_steps_per_second": 6.041,
243
+ "step": 3182
244
+ },
245
+ {
246
+ "epoch": 2.011313639220616,
247
+ "grad_norm": 10.074417114257812,
248
+ "learning_rate": 3.9946574481458206e-05,
249
+ "loss": 3.5087,
250
+ "step": 3200
251
+ },
252
+ {
253
+ "epoch": 2.07416719044626,
254
+ "grad_norm": 6.9990434646606445,
255
+ "learning_rate": 3.963230672532998e-05,
256
+ "loss": 3.5746,
257
+ "step": 3300
258
+ },
259
+ {
260
+ "epoch": 2.1370207416719045,
261
+ "grad_norm": 6.968172073364258,
262
+ "learning_rate": 3.931803896920176e-05,
263
+ "loss": 3.6324,
264
+ "step": 3400
265
+ },
266
+ {
267
+ "epoch": 2.1998742928975488,
268
+ "grad_norm": 179.99803161621094,
269
+ "learning_rate": 3.9003771213073545e-05,
270
+ "loss": 3.4072,
271
+ "step": 3500
272
+ },
273
+ {
274
+ "epoch": 2.262727844123193,
275
+ "grad_norm": 59.86805725097656,
276
+ "learning_rate": 3.868950345694532e-05,
277
+ "loss": 3.391,
278
+ "step": 3600
279
+ },
280
+ {
281
+ "epoch": 2.3255813953488373,
282
+ "grad_norm": 7.445355415344238,
283
+ "learning_rate": 3.83752357008171e-05,
284
+ "loss": 3.2032,
285
+ "step": 3700
286
+ },
287
+ {
288
+ "epoch": 2.3884349465744816,
289
+ "grad_norm": 5.553746700286865,
290
+ "learning_rate": 3.806096794468888e-05,
291
+ "loss": 3.3644,
292
+ "step": 3800
293
+ },
294
+ {
295
+ "epoch": 2.4512884978001255,
296
+ "grad_norm": 6.544325351715088,
297
+ "learning_rate": 3.7746700188560656e-05,
298
+ "loss": 3.1666,
299
+ "step": 3900
300
+ },
301
+ {
302
+ "epoch": 2.5141420490257698,
303
+ "grad_norm": 7.863962650299072,
304
+ "learning_rate": 3.7432432432432436e-05,
305
+ "loss": 3.1982,
306
+ "step": 4000
307
+ },
308
+ {
309
+ "epoch": 2.576995600251414,
310
+ "grad_norm": 10.573624610900879,
311
+ "learning_rate": 3.7118164676304215e-05,
312
+ "loss": 3.1336,
313
+ "step": 4100
314
+ },
315
+ {
316
+ "epoch": 2.6398491514770583,
317
+ "grad_norm": 8.506134986877441,
318
+ "learning_rate": 3.680389692017599e-05,
319
+ "loss": 3.0191,
320
+ "step": 4200
321
+ },
322
+ {
323
+ "epoch": 2.7027027027027026,
324
+ "grad_norm": 7.1274518966674805,
325
+ "learning_rate": 3.6489629164047774e-05,
326
+ "loss": 3.003,
327
+ "step": 4300
328
+ },
329
+ {
330
+ "epoch": 2.765556253928347,
331
+ "grad_norm": 5.121671199798584,
332
+ "learning_rate": 3.617536140791955e-05,
333
+ "loss": 3.085,
334
+ "step": 4400
335
+ },
336
+ {
337
+ "epoch": 2.828409805153991,
338
+ "grad_norm": 6.66685152053833,
339
+ "learning_rate": 3.5861093651791327e-05,
340
+ "loss": 3.0205,
341
+ "step": 4500
342
+ },
343
+ {
344
+ "epoch": 2.8912633563796355,
345
+ "grad_norm": 8.410430908203125,
346
+ "learning_rate": 3.5546825895663106e-05,
347
+ "loss": 2.9611,
348
+ "step": 4600
349
+ },
350
+ {
351
+ "epoch": 2.95411690760528,
352
+ "grad_norm": 6.266846179962158,
353
+ "learning_rate": 3.5232558139534886e-05,
354
+ "loss": 2.9299,
355
+ "step": 4700
356
+ },
357
+ {
358
+ "epoch": 3.0,
359
+ "eval_loss": 2.3084471225738525,
360
+ "eval_runtime": 20.0337,
361
+ "eval_samples_per_second": 47.819,
362
+ "eval_steps_per_second": 5.99,
363
+ "step": 4773
364
+ },
365
+ {
366
+ "epoch": 3.016970458830924,
367
+ "grad_norm": 6.011202335357666,
368
+ "learning_rate": 3.4918290383406665e-05,
369
+ "loss": 2.886,
370
+ "step": 4800
371
+ },
372
+ {
373
+ "epoch": 3.0798240100565684,
374
+ "grad_norm": 7.204225063323975,
375
+ "learning_rate": 3.4604022627278445e-05,
376
+ "loss": 2.8579,
377
+ "step": 4900
378
+ },
379
+ {
380
+ "epoch": 3.1426775612822127,
381
+ "grad_norm": 10.316048622131348,
382
+ "learning_rate": 3.428975487115022e-05,
383
+ "loss": 2.8155,
384
+ "step": 5000
385
+ },
386
+ {
387
+ "epoch": 3.2055311125078565,
388
+ "grad_norm": 6.55385684967041,
389
+ "learning_rate": 3.3975487115022e-05,
390
+ "loss": 2.8938,
391
+ "step": 5100
392
+ },
393
+ {
394
+ "epoch": 3.268384663733501,
395
+ "grad_norm": 6.081694602966309,
396
+ "learning_rate": 3.366121935889378e-05,
397
+ "loss": 2.7344,
398
+ "step": 5200
399
+ },
400
+ {
401
+ "epoch": 3.331238214959145,
402
+ "grad_norm": 8.186753273010254,
403
+ "learning_rate": 3.3346951602765556e-05,
404
+ "loss": 2.7899,
405
+ "step": 5300
406
+ },
407
+ {
408
+ "epoch": 3.3940917661847894,
409
+ "grad_norm": 7.425989627838135,
410
+ "learning_rate": 3.3032683846637335e-05,
411
+ "loss": 2.7317,
412
+ "step": 5400
413
+ },
414
+ {
415
+ "epoch": 3.4569453174104336,
416
+ "grad_norm": 5.459439277648926,
417
+ "learning_rate": 3.2718416090509115e-05,
418
+ "loss": 2.6456,
419
+ "step": 5500
420
+ },
421
+ {
422
+ "epoch": 3.519798868636078,
423
+ "grad_norm": 5.077919006347656,
424
+ "learning_rate": 3.2404148334380894e-05,
425
+ "loss": 2.6816,
426
+ "step": 5600
427
+ },
428
+ {
429
+ "epoch": 3.5826524198617222,
430
+ "grad_norm": 5.81939172744751,
431
+ "learning_rate": 3.2089880578252674e-05,
432
+ "loss": 2.64,
433
+ "step": 5700
434
+ },
435
+ {
436
+ "epoch": 3.6455059710873665,
437
+ "grad_norm": 39.74727249145508,
438
+ "learning_rate": 3.177561282212445e-05,
439
+ "loss": 2.6725,
440
+ "step": 5800
441
+ },
442
+ {
443
+ "epoch": 3.708359522313011,
444
+ "grad_norm": 5.927642345428467,
445
+ "learning_rate": 3.1461345065996226e-05,
446
+ "loss": 2.5395,
447
+ "step": 5900
448
+ },
449
+ {
450
+ "epoch": 3.771213073538655,
451
+ "grad_norm": 5.984442710876465,
452
+ "learning_rate": 3.114707730986801e-05,
453
+ "loss": 2.6297,
454
+ "step": 6000
455
+ },
456
+ {
457
+ "epoch": 3.834066624764299,
458
+ "grad_norm": 5.258358478546143,
459
+ "learning_rate": 3.083280955373979e-05,
460
+ "loss": 2.6291,
461
+ "step": 6100
462
+ },
463
+ {
464
+ "epoch": 3.8969201759899432,
465
+ "grad_norm": 5.7379937171936035,
466
+ "learning_rate": 3.0518541797611565e-05,
467
+ "loss": 2.6116,
468
+ "step": 6200
469
+ },
470
+ {
471
+ "epoch": 3.9597737272155875,
472
+ "grad_norm": 5.038835048675537,
473
+ "learning_rate": 3.0204274041483344e-05,
474
+ "loss": 2.6695,
475
+ "step": 6300
476
+ },
477
+ {
478
+ "epoch": 4.0,
479
+ "eval_loss": 2.0932769775390625,
480
+ "eval_runtime": 20.0417,
481
+ "eval_samples_per_second": 47.8,
482
+ "eval_steps_per_second": 5.988,
483
+ "step": 6364
484
+ },
485
+ {
486
+ "epoch": 4.022627278441232,
487
+ "grad_norm": 7.459395885467529,
488
+ "learning_rate": 2.9890006285355127e-05,
489
+ "loss": 2.6404,
490
+ "step": 6400
491
+ },
492
+ {
493
+ "epoch": 4.085480829666876,
494
+ "grad_norm": 6.721461296081543,
495
+ "learning_rate": 2.9575738529226903e-05,
496
+ "loss": 2.4614,
497
+ "step": 6500
498
+ },
499
+ {
500
+ "epoch": 4.14833438089252,
501
+ "grad_norm": 6.69769287109375,
502
+ "learning_rate": 2.9261470773098683e-05,
503
+ "loss": 2.457,
504
+ "step": 6600
505
+ },
506
+ {
507
+ "epoch": 4.211187932118165,
508
+ "grad_norm": 5.306356906890869,
509
+ "learning_rate": 2.894720301697046e-05,
510
+ "loss": 2.513,
511
+ "step": 6700
512
+ },
513
+ {
514
+ "epoch": 4.274041483343809,
515
+ "grad_norm": 5.425265312194824,
516
+ "learning_rate": 2.8632935260842235e-05,
517
+ "loss": 2.5467,
518
+ "step": 6800
519
+ },
520
+ {
521
+ "epoch": 4.336895034569453,
522
+ "grad_norm": 4.722207546234131,
523
+ "learning_rate": 2.8318667504714018e-05,
524
+ "loss": 2.3467,
525
+ "step": 6900
526
+ },
527
+ {
528
+ "epoch": 4.3997485857950975,
529
+ "grad_norm": 4.346086502075195,
530
+ "learning_rate": 2.8004399748585797e-05,
531
+ "loss": 2.5098,
532
+ "step": 7000
533
+ },
534
+ {
535
+ "epoch": 4.462602137020742,
536
+ "grad_norm": 7.4684319496154785,
537
+ "learning_rate": 2.7690131992457573e-05,
538
+ "loss": 2.4396,
539
+ "step": 7100
540
+ },
541
+ {
542
+ "epoch": 4.525455688246386,
543
+ "grad_norm": 5.709039688110352,
544
+ "learning_rate": 2.7375864236329353e-05,
545
+ "loss": 2.4688,
546
+ "step": 7200
547
+ },
548
+ {
549
+ "epoch": 4.58830923947203,
550
+ "grad_norm": 4.952858924865723,
551
+ "learning_rate": 2.7061596480201136e-05,
552
+ "loss": 2.3643,
553
+ "step": 7300
554
+ },
555
+ {
556
+ "epoch": 4.651162790697675,
557
+ "grad_norm": 6.68017578125,
558
+ "learning_rate": 2.6747328724072912e-05,
559
+ "loss": 2.4242,
560
+ "step": 7400
561
+ },
562
+ {
563
+ "epoch": 4.714016341923319,
564
+ "grad_norm": 3.584669828414917,
565
+ "learning_rate": 2.6433060967944688e-05,
566
+ "loss": 2.4552,
567
+ "step": 7500
568
+ },
569
+ {
570
+ "epoch": 4.776869893148963,
571
+ "grad_norm": 5.264488220214844,
572
+ "learning_rate": 2.6118793211816468e-05,
573
+ "loss": 2.4232,
574
+ "step": 7600
575
+ },
576
+ {
577
+ "epoch": 4.8397234443746076,
578
+ "grad_norm": 4.609414100646973,
579
+ "learning_rate": 2.580452545568825e-05,
580
+ "loss": 2.4418,
581
+ "step": 7700
582
+ },
583
+ {
584
+ "epoch": 4.902576995600251,
585
+ "grad_norm": 4.986881256103516,
586
+ "learning_rate": 2.5490257699560027e-05,
587
+ "loss": 2.4065,
588
+ "step": 7800
589
+ },
590
+ {
591
+ "epoch": 4.965430546825896,
592
+ "grad_norm": 4.9718098640441895,
593
+ "learning_rate": 2.5175989943431806e-05,
594
+ "loss": 2.4589,
595
+ "step": 7900
596
+ },
597
+ {
598
+ "epoch": 5.0,
599
+ "eval_loss": 1.984979271888733,
600
+ "eval_runtime": 20.0353,
601
+ "eval_samples_per_second": 47.816,
602
+ "eval_steps_per_second": 5.989,
603
+ "step": 7955
604
+ },
605
+ {
606
+ "epoch": 5.0282840980515395,
607
+ "grad_norm": 5.2526750564575195,
608
+ "learning_rate": 2.4861722187303586e-05,
609
+ "loss": 2.2708,
610
+ "step": 8000
611
+ },
612
+ {
613
+ "epoch": 5.091137649277184,
614
+ "grad_norm": 5.312747001647949,
615
+ "learning_rate": 2.454745443117536e-05,
616
+ "loss": 2.3068,
617
+ "step": 8100
618
+ },
619
+ {
620
+ "epoch": 5.153991200502828,
621
+ "grad_norm": 7.204046726226807,
622
+ "learning_rate": 2.423318667504714e-05,
623
+ "loss": 2.3729,
624
+ "step": 8200
625
+ },
626
+ {
627
+ "epoch": 5.216844751728472,
628
+ "grad_norm": 4.8044753074646,
629
+ "learning_rate": 2.391891891891892e-05,
630
+ "loss": 2.3501,
631
+ "step": 8300
632
+ },
633
+ {
634
+ "epoch": 5.279698302954117,
635
+ "grad_norm": 6.9473185539245605,
636
+ "learning_rate": 2.3604651162790697e-05,
637
+ "loss": 2.3398,
638
+ "step": 8400
639
+ },
640
+ {
641
+ "epoch": 5.342551854179761,
642
+ "grad_norm": 4.014726161956787,
643
+ "learning_rate": 2.3290383406662476e-05,
644
+ "loss": 2.2938,
645
+ "step": 8500
646
+ },
647
+ {
648
+ "epoch": 5.405405405405405,
649
+ "grad_norm": 6.722488880157471,
650
+ "learning_rate": 2.2976115650534256e-05,
651
+ "loss": 2.2354,
652
+ "step": 8600
653
+ },
654
+ {
655
+ "epoch": 5.4682589566310495,
656
+ "grad_norm": 5.856524467468262,
657
+ "learning_rate": 2.2661847894406035e-05,
658
+ "loss": 2.2757,
659
+ "step": 8700
660
+ },
661
+ {
662
+ "epoch": 5.531112507856694,
663
+ "grad_norm": 4.9930644035339355,
664
+ "learning_rate": 2.234758013827781e-05,
665
+ "loss": 2.2586,
666
+ "step": 8800
667
+ },
668
+ {
669
+ "epoch": 5.593966059082338,
670
+ "grad_norm": 5.49005126953125,
671
+ "learning_rate": 2.2033312382149594e-05,
672
+ "loss": 2.3155,
673
+ "step": 8900
674
+ },
675
+ {
676
+ "epoch": 5.656819610307982,
677
+ "grad_norm": 8.850517272949219,
678
+ "learning_rate": 2.171904462602137e-05,
679
+ "loss": 2.2841,
680
+ "step": 9000
681
+ },
682
+ {
683
+ "epoch": 5.719673161533627,
684
+ "grad_norm": 5.094405651092529,
685
+ "learning_rate": 2.140477686989315e-05,
686
+ "loss": 2.3147,
687
+ "step": 9100
688
+ },
689
+ {
690
+ "epoch": 5.782526712759271,
691
+ "grad_norm": 4.709909439086914,
692
+ "learning_rate": 2.109050911376493e-05,
693
+ "loss": 2.1584,
694
+ "step": 9200
695
+ },
696
+ {
697
+ "epoch": 5.845380263984915,
698
+ "grad_norm": 4.1693525314331055,
699
+ "learning_rate": 2.077624135763671e-05,
700
+ "loss": 2.2396,
701
+ "step": 9300
702
+ },
703
+ {
704
+ "epoch": 5.90823381521056,
705
+ "grad_norm": 6.800940036773682,
706
+ "learning_rate": 2.0461973601508485e-05,
707
+ "loss": 2.301,
708
+ "step": 9400
709
+ },
710
+ {
711
+ "epoch": 5.971087366436204,
712
+ "grad_norm": 7.419278144836426,
713
+ "learning_rate": 2.0147705845380265e-05,
714
+ "loss": 2.3142,
715
+ "step": 9500
716
+ },
717
+ {
718
+ "epoch": 6.0,
719
+ "eval_loss": 1.905881643295288,
720
+ "eval_runtime": 20.0332,
721
+ "eval_samples_per_second": 47.821,
722
+ "eval_steps_per_second": 5.99,
723
+ "step": 9546
724
+ },
725
+ {
726
+ "epoch": 6.033940917661848,
727
+ "grad_norm": 4.217894077301025,
728
+ "learning_rate": 1.9833438089252044e-05,
729
+ "loss": 2.1013,
730
+ "step": 9600
731
+ },
732
+ {
733
+ "epoch": 6.096794468887492,
734
+ "grad_norm": 5.345584869384766,
735
+ "learning_rate": 1.9519170333123824e-05,
736
+ "loss": 2.2714,
737
+ "step": 9700
738
+ },
739
+ {
740
+ "epoch": 6.159648020113137,
741
+ "grad_norm": 5.364700794219971,
742
+ "learning_rate": 1.92049025769956e-05,
743
+ "loss": 2.2381,
744
+ "step": 9800
745
+ },
746
+ {
747
+ "epoch": 6.222501571338781,
748
+ "grad_norm": 4.380568504333496,
749
+ "learning_rate": 1.8890634820867383e-05,
750
+ "loss": 2.1527,
751
+ "step": 9900
752
+ },
753
+ {
754
+ "epoch": 6.285355122564425,
755
+ "grad_norm": 6.300790309906006,
756
+ "learning_rate": 1.857636706473916e-05,
757
+ "loss": 2.1771,
758
+ "step": 10000
759
+ },
760
+ {
761
+ "epoch": 6.348208673790069,
762
+ "grad_norm": 5.757110118865967,
763
+ "learning_rate": 1.8262099308610938e-05,
764
+ "loss": 2.1695,
765
+ "step": 10100
766
+ },
767
+ {
768
+ "epoch": 6.411062225015713,
769
+ "grad_norm": 4.908361434936523,
770
+ "learning_rate": 1.7947831552482718e-05,
771
+ "loss": 2.1056,
772
+ "step": 10200
773
+ },
774
+ {
775
+ "epoch": 6.473915776241357,
776
+ "grad_norm": 5.048102378845215,
777
+ "learning_rate": 1.7633563796354494e-05,
778
+ "loss": 2.2112,
779
+ "step": 10300
780
+ },
781
+ {
782
+ "epoch": 6.536769327467002,
783
+ "grad_norm": 8.040143013000488,
784
+ "learning_rate": 1.7319296040226273e-05,
785
+ "loss": 2.0298,
786
+ "step": 10400
787
+ },
788
+ {
789
+ "epoch": 6.599622878692646,
790
+ "grad_norm": 5.15581750869751,
791
+ "learning_rate": 1.7005028284098053e-05,
792
+ "loss": 2.1224,
793
+ "step": 10500
794
+ },
795
+ {
796
+ "epoch": 6.66247642991829,
797
+ "grad_norm": 4.935842514038086,
798
+ "learning_rate": 1.6690760527969832e-05,
799
+ "loss": 2.0772,
800
+ "step": 10600
801
+ },
802
+ {
803
+ "epoch": 6.725329981143934,
804
+ "grad_norm": 5.487718105316162,
805
+ "learning_rate": 1.637649277184161e-05,
806
+ "loss": 2.2552,
807
+ "step": 10700
808
+ },
809
+ {
810
+ "epoch": 6.788183532369579,
811
+ "grad_norm": 5.713748455047607,
812
+ "learning_rate": 1.6062225015713388e-05,
813
+ "loss": 2.1358,
814
+ "step": 10800
815
+ },
816
+ {
817
+ "epoch": 6.851037083595223,
818
+ "grad_norm": 4.882757186889648,
819
+ "learning_rate": 1.5747957259585168e-05,
820
+ "loss": 2.1613,
821
+ "step": 10900
822
+ },
823
+ {
824
+ "epoch": 6.913890634820867,
825
+ "grad_norm": 5.634950637817383,
826
+ "learning_rate": 1.5433689503456947e-05,
827
+ "loss": 2.2567,
828
+ "step": 11000
829
+ },
830
+ {
831
+ "epoch": 6.976744186046512,
832
+ "grad_norm": 5.634829044342041,
833
+ "learning_rate": 1.5119421747328725e-05,
834
+ "loss": 2.1283,
835
+ "step": 11100
836
+ },
837
+ {
838
+ "epoch": 7.0,
839
+ "eval_loss": 1.84635591506958,
840
+ "eval_runtime": 20.0367,
841
+ "eval_samples_per_second": 47.812,
842
+ "eval_steps_per_second": 5.989,
843
+ "step": 11137
844
+ },
845
+ {
846
+ "epoch": 7.039597737272156,
847
+ "grad_norm": 5.635861873626709,
848
+ "learning_rate": 1.4805153991200504e-05,
849
+ "loss": 2.0938,
850
+ "step": 11200
851
+ },
852
+ {
853
+ "epoch": 7.1024512884978,
854
+ "grad_norm": 5.214977741241455,
855
+ "learning_rate": 1.4490886235072282e-05,
856
+ "loss": 2.062,
857
+ "step": 11300
858
+ },
859
+ {
860
+ "epoch": 7.1653048397234445,
861
+ "grad_norm": 7.498839855194092,
862
+ "learning_rate": 1.4176618478944062e-05,
863
+ "loss": 2.1292,
864
+ "step": 11400
865
+ },
866
+ {
867
+ "epoch": 7.228158390949089,
868
+ "grad_norm": 5.83459997177124,
869
+ "learning_rate": 1.386235072281584e-05,
870
+ "loss": 2.0796,
871
+ "step": 11500
872
+ },
873
+ {
874
+ "epoch": 7.291011942174733,
875
+ "grad_norm": 3.8935282230377197,
876
+ "learning_rate": 1.3548082966687619e-05,
877
+ "loss": 2.1414,
878
+ "step": 11600
879
+ },
880
+ {
881
+ "epoch": 7.353865493400377,
882
+ "grad_norm": 5.774020671844482,
883
+ "learning_rate": 1.3233815210559397e-05,
884
+ "loss": 2.145,
885
+ "step": 11700
886
+ },
887
+ {
888
+ "epoch": 7.416719044626022,
889
+ "grad_norm": 128.24192810058594,
890
+ "learning_rate": 1.2919547454431178e-05,
891
+ "loss": 2.0242,
892
+ "step": 11800
893
+ },
894
+ {
895
+ "epoch": 7.479572595851666,
896
+ "grad_norm": 4.4846367835998535,
897
+ "learning_rate": 1.2605279698302954e-05,
898
+ "loss": 2.0936,
899
+ "step": 11900
900
+ },
901
+ {
902
+ "epoch": 7.54242614707731,
903
+ "grad_norm": 5.091222763061523,
904
+ "learning_rate": 1.2291011942174734e-05,
905
+ "loss": 2.1988,
906
+ "step": 12000
907
+ },
908
+ {
909
+ "epoch": 7.6052796983029545,
910
+ "grad_norm": 3.3482093811035156,
911
+ "learning_rate": 1.1976744186046513e-05,
912
+ "loss": 2.1323,
913
+ "step": 12100
914
+ },
915
+ {
916
+ "epoch": 7.668133249528598,
917
+ "grad_norm": 5.329409599304199,
918
+ "learning_rate": 1.1662476429918291e-05,
919
+ "loss": 2.0587,
920
+ "step": 12200
921
+ },
922
+ {
923
+ "epoch": 7.730986800754243,
924
+ "grad_norm": 7.584386348724365,
925
+ "learning_rate": 1.134820867379007e-05,
926
+ "loss": 2.1341,
927
+ "step": 12300
928
+ },
929
+ {
930
+ "epoch": 7.7938403519798864,
931
+ "grad_norm": 5.996345520019531,
932
+ "learning_rate": 1.1033940917661848e-05,
933
+ "loss": 2.1108,
934
+ "step": 12400
935
+ },
936
+ {
937
+ "epoch": 7.856693903205531,
938
+ "grad_norm": 6.1731648445129395,
939
+ "learning_rate": 1.0719673161533628e-05,
940
+ "loss": 2.1218,
941
+ "step": 12500
942
+ },
943
+ {
944
+ "epoch": 7.919547454431175,
945
+ "grad_norm": 5.414481163024902,
946
+ "learning_rate": 1.0405405405405407e-05,
947
+ "loss": 2.028,
948
+ "step": 12600
949
+ },
950
+ {
951
+ "epoch": 7.982401005656819,
952
+ "grad_norm": 7.198294639587402,
953
+ "learning_rate": 1.0091137649277185e-05,
954
+ "loss": 2.0489,
955
+ "step": 12700
956
+ },
957
+ {
958
+ "epoch": 8.0,
959
+ "eval_loss": 1.8111430406570435,
960
+ "eval_runtime": 20.0666,
961
+ "eval_samples_per_second": 47.741,
962
+ "eval_steps_per_second": 5.98,
963
+ "step": 12728
964
+ },
965
+ {
966
+ "epoch": 8.045254556882464,
967
+ "grad_norm": 6.677022933959961,
968
+ "learning_rate": 9.776869893148963e-06,
969
+ "loss": 2.0814,
970
+ "step": 12800
971
+ },
972
+ {
973
+ "epoch": 8.108108108108109,
974
+ "grad_norm": 5.1916728019714355,
975
+ "learning_rate": 9.46260213702074e-06,
976
+ "loss": 2.119,
977
+ "step": 12900
978
+ },
979
+ {
980
+ "epoch": 8.170961659333752,
981
+ "grad_norm": 6.04162073135376,
982
+ "learning_rate": 9.14833438089252e-06,
983
+ "loss": 2.0058,
984
+ "step": 13000
985
+ },
986
+ {
987
+ "epoch": 8.233815210559397,
988
+ "grad_norm": 4.764267444610596,
989
+ "learning_rate": 8.8340666247643e-06,
990
+ "loss": 2.0113,
991
+ "step": 13100
992
+ },
993
+ {
994
+ "epoch": 8.29666876178504,
995
+ "grad_norm": 5.77971887588501,
996
+ "learning_rate": 8.519798868636078e-06,
997
+ "loss": 2.0392,
998
+ "step": 13200
999
+ },
1000
+ {
1001
+ "epoch": 8.359522313010686,
1002
+ "grad_norm": 5.698218822479248,
1003
+ "learning_rate": 8.205531112507857e-06,
1004
+ "loss": 2.107,
1005
+ "step": 13300
1006
+ },
1007
+ {
1008
+ "epoch": 8.42237586423633,
1009
+ "grad_norm": 5.236012935638428,
1010
+ "learning_rate": 7.891263356379635e-06,
1011
+ "loss": 2.0829,
1012
+ "step": 13400
1013
+ },
1014
+ {
1015
+ "epoch": 8.485229415461973,
1016
+ "grad_norm": 4.379955291748047,
1017
+ "learning_rate": 7.576995600251414e-06,
1018
+ "loss": 1.9321,
1019
+ "step": 13500
1020
+ },
1021
+ {
1022
+ "epoch": 8.548082966687618,
1023
+ "grad_norm": 6.034859657287598,
1024
+ "learning_rate": 7.262727844123193e-06,
1025
+ "loss": 2.1013,
1026
+ "step": 13600
1027
+ },
1028
+ {
1029
+ "epoch": 8.610936517913261,
1030
+ "grad_norm": 5.320705413818359,
1031
+ "learning_rate": 6.948460087994972e-06,
1032
+ "loss": 2.0543,
1033
+ "step": 13700
1034
+ },
1035
+ {
1036
+ "epoch": 8.673790069138906,
1037
+ "grad_norm": 5.735895156860352,
1038
+ "learning_rate": 6.634192331866751e-06,
1039
+ "loss": 2.0594,
1040
+ "step": 13800
1041
+ },
1042
+ {
1043
+ "epoch": 8.73664362036455,
1044
+ "grad_norm": 4.845800876617432,
1045
+ "learning_rate": 6.31992457573853e-06,
1046
+ "loss": 1.9402,
1047
+ "step": 13900
1048
+ },
1049
+ {
1050
+ "epoch": 8.799497171590195,
1051
+ "grad_norm": 4.628382682800293,
1052
+ "learning_rate": 6.0056568196103085e-06,
1053
+ "loss": 1.9937,
1054
+ "step": 14000
1055
+ },
1056
+ {
1057
+ "epoch": 8.862350722815838,
1058
+ "grad_norm": 4.747410774230957,
1059
+ "learning_rate": 5.691389063482086e-06,
1060
+ "loss": 2.0654,
1061
+ "step": 14100
1062
+ },
1063
+ {
1064
+ "epoch": 8.925204274041484,
1065
+ "grad_norm": 4.694166660308838,
1066
+ "learning_rate": 5.377121307353866e-06,
1067
+ "loss": 2.0523,
1068
+ "step": 14200
1069
+ },
1070
+ {
1071
+ "epoch": 8.988057825267127,
1072
+ "grad_norm": 6.711084365844727,
1073
+ "learning_rate": 5.0628535512256445e-06,
1074
+ "loss": 1.9856,
1075
+ "step": 14300
1076
+ },
1077
+ {
1078
+ "epoch": 9.0,
1079
+ "eval_loss": 1.7920939922332764,
1080
+ "eval_runtime": 20.0378,
1081
+ "eval_samples_per_second": 47.81,
1082
+ "eval_steps_per_second": 5.989,
1083
+ "step": 14319
1084
+ },
1085
+ {
1086
+ "epoch": 9.050911376492772,
1087
+ "grad_norm": 6.053162097930908,
1088
+ "learning_rate": 4.748585795097423e-06,
1089
+ "loss": 2.0392,
1090
+ "step": 14400
1091
+ },
1092
+ {
1093
+ "epoch": 9.113764927718416,
1094
+ "grad_norm": 4.806529521942139,
1095
+ "learning_rate": 4.434318038969202e-06,
1096
+ "loss": 2.0308,
1097
+ "step": 14500
1098
+ },
1099
+ {
1100
+ "epoch": 9.17661847894406,
1101
+ "grad_norm": 4.725819110870361,
1102
+ "learning_rate": 4.1200502828409805e-06,
1103
+ "loss": 2.0441,
1104
+ "step": 14600
1105
+ },
1106
+ {
1107
+ "epoch": 9.239472030169704,
1108
+ "grad_norm": 4.637420177459717,
1109
+ "learning_rate": 3.8057825267127596e-06,
1110
+ "loss": 2.0061,
1111
+ "step": 14700
1112
+ },
1113
+ {
1114
+ "epoch": 9.30232558139535,
1115
+ "grad_norm": 6.441665172576904,
1116
+ "learning_rate": 3.4915147705845382e-06,
1117
+ "loss": 2.1299,
1118
+ "step": 14800
1119
+ },
1120
+ {
1121
+ "epoch": 9.365179132620993,
1122
+ "grad_norm": 3.506943941116333,
1123
+ "learning_rate": 3.1772470144563173e-06,
1124
+ "loss": 1.9443,
1125
+ "step": 14900
1126
+ },
1127
+ {
1128
+ "epoch": 9.428032683846638,
1129
+ "grad_norm": 8.454822540283203,
1130
+ "learning_rate": 2.8629792583280956e-06,
1131
+ "loss": 2.0327,
1132
+ "step": 15000
1133
+ },
1134
+ {
1135
+ "epoch": 9.490886235072281,
1136
+ "grad_norm": 5.021187782287598,
1137
+ "learning_rate": 2.5487115021998746e-06,
1138
+ "loss": 1.9839,
1139
+ "step": 15100
1140
+ },
1141
+ {
1142
+ "epoch": 9.553739786297927,
1143
+ "grad_norm": 6.3962016105651855,
1144
+ "learning_rate": 2.234443746071653e-06,
1145
+ "loss": 2.0604,
1146
+ "step": 15200
1147
+ },
1148
+ {
1149
+ "epoch": 9.61659333752357,
1150
+ "grad_norm": 5.531436443328857,
1151
+ "learning_rate": 1.920175989943432e-06,
1152
+ "loss": 2.0168,
1153
+ "step": 15300
1154
+ },
1155
+ {
1156
+ "epoch": 9.679446888749215,
1157
+ "grad_norm": 4.300695896148682,
1158
+ "learning_rate": 1.6059082338152106e-06,
1159
+ "loss": 1.9994,
1160
+ "step": 15400
1161
+ },
1162
+ {
1163
+ "epoch": 9.742300439974859,
1164
+ "grad_norm": 3.102018356323242,
1165
+ "learning_rate": 1.2916404776869893e-06,
1166
+ "loss": 2.0441,
1167
+ "step": 15500
1168
+ },
1169
+ {
1170
+ "epoch": 9.805153991200502,
1171
+ "grad_norm": 4.91919469833374,
1172
+ "learning_rate": 9.773727215587681e-07,
1173
+ "loss": 1.9584,
1174
+ "step": 15600
1175
+ },
1176
+ {
1177
+ "epoch": 9.868007542426147,
1178
+ "grad_norm": 4.21737813949585,
1179
+ "learning_rate": 6.631049654305469e-07,
1180
+ "loss": 2.0019,
1181
+ "step": 15700
1182
+ },
1183
+ {
1184
+ "epoch": 9.930861093651792,
1185
+ "grad_norm": 4.098769187927246,
1186
+ "learning_rate": 3.4883720930232557e-07,
1187
+ "loss": 2.0121,
1188
+ "step": 15800
1189
+ },
1190
+ {
1191
+ "epoch": 9.993714644877436,
1192
+ "grad_norm": 4.722096920013428,
1193
+ "learning_rate": 3.456945317410434e-08,
1194
+ "loss": 2.0196,
1195
+ "step": 15900
1196
+ },
1197
+ {
1198
+ "epoch": 10.0,
1199
+ "eval_loss": 1.787421464920044,
1200
+ "eval_runtime": 20.0243,
1201
+ "eval_samples_per_second": 47.842,
1202
+ "eval_steps_per_second": 5.993,
1203
+ "step": 15910
1204
+ },
1205
+ {
1206
+ "epoch": 10.056568196103079,
1207
+ "grad_norm": 3.8331987857818604,
1208
+ "learning_rate": 2.4860150848522942e-05,
1209
+ "loss": 2.0388,
1210
+ "step": 16000
1211
+ },
1212
+ {
1213
+ "epoch": 10.119421747328724,
1214
+ "grad_norm": 3.9292027950286865,
1215
+ "learning_rate": 2.4703016970458832e-05,
1216
+ "loss": 2.0913,
1217
+ "step": 16100
1218
+ },
1219
+ {
1220
+ "epoch": 10.182275298554368,
1221
+ "grad_norm": 5.124855995178223,
1222
+ "learning_rate": 2.454588309239472e-05,
1223
+ "loss": 2.0452,
1224
+ "step": 16200
1225
+ },
1226
+ {
1227
+ "epoch": 10.245128849780013,
1228
+ "grad_norm": 5.743933200836182,
1229
+ "learning_rate": 2.438874921433061e-05,
1230
+ "loss": 2.016,
1231
+ "step": 16300
1232
+ },
1233
+ {
1234
+ "epoch": 10.307982401005656,
1235
+ "grad_norm": 6.4510931968688965,
1236
+ "learning_rate": 2.42316153362665e-05,
1237
+ "loss": 1.9785,
1238
+ "step": 16400
1239
+ },
1240
+ {
1241
+ "epoch": 10.370835952231301,
1242
+ "grad_norm": 6.550465106964111,
1243
+ "learning_rate": 2.4074481458202387e-05,
1244
+ "loss": 1.9912,
1245
+ "step": 16500
1246
+ },
1247
+ {
1248
+ "epoch": 10.433689503456945,
1249
+ "grad_norm": 5.37285852432251,
1250
+ "learning_rate": 2.391734758013828e-05,
1251
+ "loss": 2.0549,
1252
+ "step": 16600
1253
+ },
1254
+ {
1255
+ "epoch": 10.49654305468259,
1256
+ "grad_norm": 5.4893412590026855,
1257
+ "learning_rate": 2.376021370207417e-05,
1258
+ "loss": 1.9434,
1259
+ "step": 16700
1260
+ },
1261
+ {
1262
+ "epoch": 10.559396605908233,
1263
+ "grad_norm": 4.316259384155273,
1264
+ "learning_rate": 2.3603079824010057e-05,
1265
+ "loss": 1.8413,
1266
+ "step": 16800
1267
+ },
1268
+ {
1269
+ "epoch": 10.622250157133879,
1270
+ "grad_norm": 3.4342756271362305,
1271
+ "learning_rate": 2.3445945945945946e-05,
1272
+ "loss": 1.9312,
1273
+ "step": 16900
1274
+ },
1275
+ {
1276
+ "epoch": 10.685103708359522,
1277
+ "grad_norm": 5.680815696716309,
1278
+ "learning_rate": 2.3288812067881836e-05,
1279
+ "loss": 1.9678,
1280
+ "step": 17000
1281
+ },
1282
+ {
1283
+ "epoch": 10.747957259585167,
1284
+ "grad_norm": 6.04569149017334,
1285
+ "learning_rate": 2.3131678189817726e-05,
1286
+ "loss": 2.0329,
1287
+ "step": 17100
1288
+ },
1289
+ {
1290
+ "epoch": 10.81081081081081,
1291
+ "grad_norm": 9.336991310119629,
1292
+ "learning_rate": 2.2974544311753616e-05,
1293
+ "loss": 1.9575,
1294
+ "step": 17200
1295
+ },
1296
+ {
1297
+ "epoch": 10.873664362036456,
1298
+ "grad_norm": 3.826447010040283,
1299
+ "learning_rate": 2.2817410433689505e-05,
1300
+ "loss": 1.9692,
1301
+ "step": 17300
1302
+ },
1303
+ {
1304
+ "epoch": 10.936517913262099,
1305
+ "grad_norm": 4.134801387786865,
1306
+ "learning_rate": 2.2660276555625392e-05,
1307
+ "loss": 2.0406,
1308
+ "step": 17400
1309
+ },
1310
+ {
1311
+ "epoch": 10.999371464487744,
1312
+ "grad_norm": 5.291431903839111,
1313
+ "learning_rate": 2.2503142677561285e-05,
1314
+ "loss": 1.9631,
1315
+ "step": 17500
1316
+ },
1317
+ {
1318
+ "epoch": 11.0,
1319
+ "eval_loss": 1.7517410516738892,
1320
+ "eval_runtime": 21.6572,
1321
+ "eval_samples_per_second": 44.235,
1322
+ "eval_steps_per_second": 5.541,
1323
+ "step": 17501
1324
+ },
1325
+ {
1326
+ "epoch": 11.062225015713388,
1327
+ "grad_norm": 4.9575066566467285,
1328
+ "learning_rate": 2.234600879949717e-05,
1329
+ "loss": 1.9381,
1330
+ "step": 17600
1331
+ },
1332
+ {
1333
+ "epoch": 11.125078566939033,
1334
+ "grad_norm": 12.871175765991211,
1335
+ "learning_rate": 2.218887492143306e-05,
1336
+ "loss": 1.8867,
1337
+ "step": 17700
1338
+ },
1339
+ {
1340
+ "epoch": 11.187932118164676,
1341
+ "grad_norm": 4.3662519454956055,
1342
+ "learning_rate": 2.203174104336895e-05,
1343
+ "loss": 1.9713,
1344
+ "step": 17800
1345
+ },
1346
+ {
1347
+ "epoch": 11.250785669390321,
1348
+ "grad_norm": 5.662289619445801,
1349
+ "learning_rate": 2.187460716530484e-05,
1350
+ "loss": 1.9188,
1351
+ "step": 17900
1352
+ },
1353
+ {
1354
+ "epoch": 11.313639220615965,
1355
+ "grad_norm": 7.633818626403809,
1356
+ "learning_rate": 2.171747328724073e-05,
1357
+ "loss": 1.9142,
1358
+ "step": 18000
1359
+ },
1360
+ {
1361
+ "epoch": 11.376492771841608,
1362
+ "grad_norm": 4.940028667449951,
1363
+ "learning_rate": 2.156033940917662e-05,
1364
+ "loss": 1.8697,
1365
+ "step": 18100
1366
+ },
1367
+ {
1368
+ "epoch": 11.439346323067253,
1369
+ "grad_norm": 5.070211410522461,
1370
+ "learning_rate": 2.1403205531112506e-05,
1371
+ "loss": 1.9624,
1372
+ "step": 18200
1373
+ },
1374
+ {
1375
+ "epoch": 11.502199874292897,
1376
+ "grad_norm": 7.409548282623291,
1377
+ "learning_rate": 2.12460716530484e-05,
1378
+ "loss": 1.9283,
1379
+ "step": 18300
1380
+ },
1381
+ {
1382
+ "epoch": 11.565053425518542,
1383
+ "grad_norm": 6.541192531585693,
1384
+ "learning_rate": 2.108893777498429e-05,
1385
+ "loss": 1.9357,
1386
+ "step": 18400
1387
+ },
1388
+ {
1389
+ "epoch": 11.627906976744185,
1390
+ "grad_norm": 5.941864967346191,
1391
+ "learning_rate": 2.0931803896920176e-05,
1392
+ "loss": 1.869,
1393
+ "step": 18500
1394
+ },
1395
+ {
1396
+ "epoch": 11.69076052796983,
1397
+ "grad_norm": 9.418646812438965,
1398
+ "learning_rate": 2.0774670018856065e-05,
1399
+ "loss": 1.8518,
1400
+ "step": 18600
1401
+ },
1402
+ {
1403
+ "epoch": 11.753614079195474,
1404
+ "grad_norm": 5.367152690887451,
1405
+ "learning_rate": 2.061753614079196e-05,
1406
+ "loss": 1.8945,
1407
+ "step": 18700
1408
+ },
1409
+ {
1410
+ "epoch": 11.81646763042112,
1411
+ "grad_norm": 5.896432399749756,
1412
+ "learning_rate": 2.0460402262727845e-05,
1413
+ "loss": 1.8569,
1414
+ "step": 18800
1415
+ },
1416
+ {
1417
+ "epoch": 11.879321181646763,
1418
+ "grad_norm": 6.137564182281494,
1419
+ "learning_rate": 2.0303268384663735e-05,
1420
+ "loss": 1.9179,
1421
+ "step": 18900
1422
+ },
1423
+ {
1424
+ "epoch": 11.942174732872408,
1425
+ "grad_norm": 4.5933918952941895,
1426
+ "learning_rate": 2.0146134506599625e-05,
1427
+ "loss": 1.8941,
1428
+ "step": 19000
1429
+ },
1430
+ {
1431
+ "epoch": 12.0,
1432
+ "eval_loss": 1.7062737941741943,
1433
+ "eval_runtime": 21.7167,
1434
+ "eval_samples_per_second": 44.114,
1435
+ "eval_steps_per_second": 5.526,
1436
+ "step": 19092
1437
+ },
1438
+ {
1439
+ "epoch": 12.005028284098051,
1440
+ "grad_norm": 5.298050880432129,
1441
+ "learning_rate": 1.998900062853551e-05,
1442
+ "loss": 1.8681,
1443
+ "step": 19100
1444
+ },
1445
+ {
1446
+ "epoch": 12.067881835323696,
1447
+ "grad_norm": 7.001854419708252,
1448
+ "learning_rate": 1.9831866750471404e-05,
1449
+ "loss": 1.8377,
1450
+ "step": 19200
1451
+ },
1452
+ {
1453
+ "epoch": 12.13073538654934,
1454
+ "grad_norm": 4.692386150360107,
1455
+ "learning_rate": 1.9674732872407294e-05,
1456
+ "loss": 1.8279,
1457
+ "step": 19300
1458
+ },
1459
+ {
1460
+ "epoch": 12.193588937774985,
1461
+ "grad_norm": 6.864208221435547,
1462
+ "learning_rate": 1.951759899434318e-05,
1463
+ "loss": 1.8855,
1464
+ "step": 19400
1465
+ },
1466
+ {
1467
+ "epoch": 12.256442489000628,
1468
+ "grad_norm": 3.883880853652954,
1469
+ "learning_rate": 1.936046511627907e-05,
1470
+ "loss": 1.84,
1471
+ "step": 19500
1472
+ },
1473
+ {
1474
+ "epoch": 12.319296040226273,
1475
+ "grad_norm": 5.302524566650391,
1476
+ "learning_rate": 1.920333123821496e-05,
1477
+ "loss": 1.8791,
1478
+ "step": 19600
1479
+ },
1480
+ {
1481
+ "epoch": 12.382149591451917,
1482
+ "grad_norm": 6.854051113128662,
1483
+ "learning_rate": 1.904619736015085e-05,
1484
+ "loss": 1.9189,
1485
+ "step": 19700
1486
+ },
1487
+ {
1488
+ "epoch": 12.445003142677562,
1489
+ "grad_norm": 4.728283405303955,
1490
+ "learning_rate": 1.888906348208674e-05,
1491
+ "loss": 1.8903,
1492
+ "step": 19800
1493
+ },
1494
+ {
1495
+ "epoch": 12.507856693903205,
1496
+ "grad_norm": 4.314347267150879,
1497
+ "learning_rate": 1.8731929604022626e-05,
1498
+ "loss": 1.8615,
1499
+ "step": 19900
1500
+ },
1501
+ {
1502
+ "epoch": 12.57071024512885,
1503
+ "grad_norm": 3.873619318008423,
1504
+ "learning_rate": 1.857479572595852e-05,
1505
+ "loss": 1.8232,
1506
+ "step": 20000
1507
+ },
1508
+ {
1509
+ "epoch": 12.633563796354494,
1510
+ "grad_norm": 6.445096969604492,
1511
+ "learning_rate": 1.841766184789441e-05,
1512
+ "loss": 1.7764,
1513
+ "step": 20100
1514
+ },
1515
+ {
1516
+ "epoch": 12.696417347580137,
1517
+ "grad_norm": 4.258322715759277,
1518
+ "learning_rate": 1.8260527969830295e-05,
1519
+ "loss": 1.869,
1520
+ "step": 20200
1521
+ },
1522
+ {
1523
+ "epoch": 12.759270898805783,
1524
+ "grad_norm": 7.782538414001465,
1525
+ "learning_rate": 1.8103394091766185e-05,
1526
+ "loss": 1.7986,
1527
+ "step": 20300
1528
+ },
1529
+ {
1530
+ "epoch": 12.822124450031426,
1531
+ "grad_norm": 7.189488887786865,
1532
+ "learning_rate": 1.7946260213702078e-05,
1533
+ "loss": 1.8448,
1534
+ "step": 20400
1535
+ },
1536
+ {
1537
+ "epoch": 12.884978001257071,
1538
+ "grad_norm": 5.59601354598999,
1539
+ "learning_rate": 1.7789126335637964e-05,
1540
+ "loss": 1.7924,
1541
+ "step": 20500
1542
+ },
1543
+ {
1544
+ "epoch": 12.947831552482715,
1545
+ "grad_norm": 4.675200939178467,
1546
+ "learning_rate": 1.7631992457573854e-05,
1547
+ "loss": 1.8212,
1548
+ "step": 20600
1549
+ },
1550
+ {
1551
+ "epoch": 13.0,
1552
+ "eval_loss": 1.6696668863296509,
1553
+ "eval_runtime": 21.645,
1554
+ "eval_samples_per_second": 44.26,
1555
+ "eval_steps_per_second": 5.544,
1556
+ "step": 20683
1557
+ },
1558
+ {
1559
+ "epoch": 13.01068510370836,
1560
+ "grad_norm": 3.3650217056274414,
1561
+ "learning_rate": 1.7474858579509744e-05,
1562
+ "loss": 1.6872,
1563
+ "step": 20700
1564
+ },
1565
+ {
1566
+ "epoch": 13.073538654934003,
1567
+ "grad_norm": 6.4758219718933105,
1568
+ "learning_rate": 1.731772470144563e-05,
1569
+ "loss": 1.8029,
1570
+ "step": 20800
1571
+ },
1572
+ {
1573
+ "epoch": 13.136392206159648,
1574
+ "grad_norm": 4.500367641448975,
1575
+ "learning_rate": 1.7160590823381523e-05,
1576
+ "loss": 1.8655,
1577
+ "step": 20900
1578
+ },
1579
+ {
1580
+ "epoch": 13.199245757385292,
1581
+ "grad_norm": 5.369949817657471,
1582
+ "learning_rate": 1.7003456945317413e-05,
1583
+ "loss": 1.821,
1584
+ "step": 21000
1585
+ },
1586
+ {
1587
+ "epoch": 13.262099308610937,
1588
+ "grad_norm": 4.84245491027832,
1589
+ "learning_rate": 1.68463230672533e-05,
1590
+ "loss": 1.7454,
1591
+ "step": 21100
1592
+ },
1593
+ {
1594
+ "epoch": 13.32495285983658,
1595
+ "grad_norm": 4.510051727294922,
1596
+ "learning_rate": 1.668918918918919e-05,
1597
+ "loss": 1.8378,
1598
+ "step": 21200
1599
+ },
1600
+ {
1601
+ "epoch": 13.387806411062225,
1602
+ "grad_norm": 5.163560390472412,
1603
+ "learning_rate": 1.653205531112508e-05,
1604
+ "loss": 1.7985,
1605
+ "step": 21300
1606
+ },
1607
+ {
1608
+ "epoch": 13.450659962287869,
1609
+ "grad_norm": 4.454617023468018,
1610
+ "learning_rate": 1.637492143306097e-05,
1611
+ "loss": 1.8177,
1612
+ "step": 21400
1613
+ },
1614
+ {
1615
+ "epoch": 13.513513513513514,
1616
+ "grad_norm": 3.672908067703247,
1617
+ "learning_rate": 1.6217787554996858e-05,
1618
+ "loss": 1.6908,
1619
+ "step": 21500
1620
+ },
1621
+ {
1622
+ "epoch": 13.576367064739157,
1623
+ "grad_norm": 4.549923419952393,
1624
+ "learning_rate": 1.6060653676932748e-05,
1625
+ "loss": 1.7603,
1626
+ "step": 21600
1627
+ },
1628
+ {
1629
+ "epoch": 13.639220615964803,
1630
+ "grad_norm": 5.733989715576172,
1631
+ "learning_rate": 1.5903519798868638e-05,
1632
+ "loss": 1.7689,
1633
+ "step": 21700
1634
+ },
1635
+ {
1636
+ "epoch": 13.702074167190446,
1637
+ "grad_norm": 4.507519245147705,
1638
+ "learning_rate": 1.5746385920804527e-05,
1639
+ "loss": 1.7984,
1640
+ "step": 21800
1641
+ },
1642
+ {
1643
+ "epoch": 13.764927718416091,
1644
+ "grad_norm": 4.713226795196533,
1645
+ "learning_rate": 1.5589252042740414e-05,
1646
+ "loss": 1.8011,
1647
+ "step": 21900
1648
+ },
1649
+ {
1650
+ "epoch": 13.827781269641735,
1651
+ "grad_norm": 4.300686359405518,
1652
+ "learning_rate": 1.5432118164676304e-05,
1653
+ "loss": 1.7743,
1654
+ "step": 22000
1655
+ },
1656
+ {
1657
+ "epoch": 13.89063482086738,
1658
+ "grad_norm": 4.702789306640625,
1659
+ "learning_rate": 1.5274984286612197e-05,
1660
+ "loss": 1.6903,
1661
+ "step": 22100
1662
+ },
1663
+ {
1664
+ "epoch": 13.953488372093023,
1665
+ "grad_norm": 6.481640815734863,
1666
+ "learning_rate": 1.5117850408548085e-05,
1667
+ "loss": 1.822,
1668
+ "step": 22200
1669
+ },
1670
+ {
1671
+ "epoch": 14.0,
1672
+ "eval_loss": 1.648952603340149,
1673
+ "eval_runtime": 21.6512,
1674
+ "eval_samples_per_second": 44.247,
1675
+ "eval_steps_per_second": 5.542,
1676
+ "step": 22274
1677
+ }
1678
+ ],
1679
+ "logging_steps": 100,
1680
+ "max_steps": 31820,
1681
+ "num_input_tokens_seen": 0,
1682
+ "num_train_epochs": 20,
1683
+ "save_steps": 500,
1684
+ "stateful_callbacks": {
1685
+ "TrainerControl": {
1686
+ "args": {
1687
+ "should_epoch_stop": false,
1688
+ "should_evaluate": false,
1689
+ "should_log": false,
1690
+ "should_save": true,
1691
+ "should_training_stop": false
1692
+ },
1693
+ "attributes": {}
1694
+ }
1695
+ },
1696
+ "total_flos": 3.0487459270754304e+16,
1697
+ "train_batch_size": 8,
1698
+ "trial_name": null,
1699
+ "trial_params": null
1700
+ }
banglat5_bn_sy/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:882cea5e45defc4426c6734e0e3442ceb7d58854298cd5766c8467e7c71160b8
3
+ size 5496