MihaiPopa-1 commited on
Commit
c8758c1
·
verified ·
1 Parent(s): 22cafdb

Upload folder using huggingface_hub

Browse files
checkpoint-1953/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7400cbdb1cc8249dd839f5ffe98f4556fcac45d32f2ac7bbc1f240c78e22c427
3
+ size 46781032
checkpoint-1953/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a6a4d27c2d4004eb84399da08b05324bb8afe325670480729c209b6e1720f778
3
+ size 88853790
checkpoint-1953/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7b55f8eecdb2366444fb635d338c6c2e8aeed3c8770b339a705baf1af3af0965
3
+ size 14709
checkpoint-1953/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:93042e4d76683626f6f0e5dff3455a21baea86355546c98695345d1062b9caa9
3
+ size 1465
checkpoint-1953/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-1953/tokenizer_config.json ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": true,
3
+ "backend": "tokenizers",
4
+ "bos_token": "[CLS]",
5
+ "cls_token": "[CLS]",
6
+ "do_lower_case": true,
7
+ "eos_token": "[SEP]",
8
+ "is_local": false,
9
+ "keep_accents": false,
10
+ "mask_token": "[MASK]",
11
+ "model_max_length": 512,
12
+ "pad_token": "<pad>",
13
+ "sep_token": "[SEP]",
14
+ "tokenizer_class": "AlbertTokenizer",
15
+ "trim_offsets": true,
16
+ "unk_token": "<unk>"
17
+ }
checkpoint-1953/trainer_state.json ADDED
@@ -0,0 +1,66 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": 1953,
3
+ "best_metric": 0.07590867578983307,
4
+ "best_model_checkpoint": "./results_albert_punctuation_casing/checkpoint-1953",
5
+ "epoch": 1.0,
6
+ "eval_steps": 500,
7
+ "global_step": 1953,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.2560163850486431,
14
+ "grad_norm": 7.659163475036621,
15
+ "learning_rate": 1.8296637651476362e-05,
16
+ "loss": 0.17497076416015625,
17
+ "step": 500
18
+ },
19
+ {
20
+ "epoch": 0.5120327700972862,
21
+ "grad_norm": 9.602242469787598,
22
+ "learning_rate": 1.6589861751152075e-05,
23
+ "loss": 0.09840100860595703,
24
+ "step": 1000
25
+ },
26
+ {
27
+ "epoch": 0.7680491551459293,
28
+ "grad_norm": 2.2302889823913574,
29
+ "learning_rate": 1.4883085850827787e-05,
30
+ "loss": 0.08661653137207032,
31
+ "step": 1500
32
+ },
33
+ {
34
+ "epoch": 1.0,
35
+ "eval_casing_accuracy": 0.5982128829536528,
36
+ "eval_loss": 0.07590867578983307,
37
+ "eval_overall_accuracy": 0.597888845247447,
38
+ "eval_punctuation_accuracy": 0.5975648075412412,
39
+ "eval_runtime": 11.014,
40
+ "eval_samples_per_second": 709.096,
41
+ "eval_steps_per_second": 44.398,
42
+ "step": 1953
43
+ }
44
+ ],
45
+ "logging_steps": 500,
46
+ "max_steps": 5859,
47
+ "num_input_tokens_seen": 0,
48
+ "num_train_epochs": 3,
49
+ "save_steps": 500,
50
+ "stateful_callbacks": {
51
+ "TrainerControl": {
52
+ "args": {
53
+ "should_epoch_stop": false,
54
+ "should_evaluate": false,
55
+ "should_log": false,
56
+ "should_save": true,
57
+ "should_training_stop": false
58
+ },
59
+ "attributes": {}
60
+ }
61
+ },
62
+ "total_flos": 0.0,
63
+ "train_batch_size": 16,
64
+ "trial_name": null,
65
+ "trial_params": null
66
+ }
checkpoint-1953/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:564d1f2f5aef116f928ab806870d35d30703b30b96f0b6e8ba67019dc262c956
3
+ size 5265
checkpoint-3906/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:29710bfae86de683cfaab56f0be1872477993b2bb922a0bcb8316d9185f496ec
3
+ size 46781032
checkpoint-3906/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:70446afa570eddb4c5b4440a5fa3832f1c97bc25956520405bf398009e906dc6
3
+ size 88853790
checkpoint-3906/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:10142afb48383537512bbda75d1a3659f03009b79fddbe13d6a83040da090afc
3
+ size 14709
checkpoint-3906/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ef01b065e0b9ef65b0bb44e0c2cfff6ee797121d76aaf2deccd9c953a0a973ca
3
+ size 1465
checkpoint-3906/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-3906/tokenizer_config.json ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": true,
3
+ "backend": "tokenizers",
4
+ "bos_token": "[CLS]",
5
+ "cls_token": "[CLS]",
6
+ "do_lower_case": true,
7
+ "eos_token": "[SEP]",
8
+ "is_local": false,
9
+ "keep_accents": false,
10
+ "mask_token": "[MASK]",
11
+ "model_max_length": 512,
12
+ "pad_token": "<pad>",
13
+ "sep_token": "[SEP]",
14
+ "tokenizer_class": "AlbertTokenizer",
15
+ "trim_offsets": true,
16
+ "unk_token": "<unk>"
17
+ }
checkpoint-3906/trainer_state.json ADDED
@@ -0,0 +1,105 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": 3906,
3
+ "best_metric": 0.0746106207370758,
4
+ "best_model_checkpoint": "./results_albert_punctuation_casing/checkpoint-3906",
5
+ "epoch": 2.0,
6
+ "eval_steps": 500,
7
+ "global_step": 3906,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.2560163850486431,
14
+ "grad_norm": 7.659163475036621,
15
+ "learning_rate": 1.8296637651476362e-05,
16
+ "loss": 0.17497076416015625,
17
+ "step": 500
18
+ },
19
+ {
20
+ "epoch": 0.5120327700972862,
21
+ "grad_norm": 9.602242469787598,
22
+ "learning_rate": 1.6589861751152075e-05,
23
+ "loss": 0.09840100860595703,
24
+ "step": 1000
25
+ },
26
+ {
27
+ "epoch": 0.7680491551459293,
28
+ "grad_norm": 2.2302889823913574,
29
+ "learning_rate": 1.4883085850827787e-05,
30
+ "loss": 0.08661653137207032,
31
+ "step": 1500
32
+ },
33
+ {
34
+ "epoch": 1.0,
35
+ "eval_casing_accuracy": 0.5982128829536528,
36
+ "eval_loss": 0.07590867578983307,
37
+ "eval_overall_accuracy": 0.597888845247447,
38
+ "eval_punctuation_accuracy": 0.5975648075412412,
39
+ "eval_runtime": 11.014,
40
+ "eval_samples_per_second": 709.096,
41
+ "eval_steps_per_second": 44.398,
42
+ "step": 1953
43
+ },
44
+ {
45
+ "epoch": 1.0240655401945724,
46
+ "grad_norm": 0.46477240324020386,
47
+ "learning_rate": 1.3176309950503501e-05,
48
+ "loss": 0.08205690765380859,
49
+ "step": 2000
50
+ },
51
+ {
52
+ "epoch": 1.2800819252432156,
53
+ "grad_norm": 6.8253865242004395,
54
+ "learning_rate": 1.1469534050179212e-05,
55
+ "loss": 0.0625931053161621,
56
+ "step": 2500
57
+ },
58
+ {
59
+ "epoch": 1.5360983102918588,
60
+ "grad_norm": 3.1819510459899902,
61
+ "learning_rate": 9.762758149854925e-06,
62
+ "loss": 0.06253623962402344,
63
+ "step": 3000
64
+ },
65
+ {
66
+ "epoch": 1.7921146953405018,
67
+ "grad_norm": 4.698487281799316,
68
+ "learning_rate": 8.055982249530637e-06,
69
+ "loss": 0.053002174377441404,
70
+ "step": 3500
71
+ },
72
+ {
73
+ "epoch": 2.0,
74
+ "eval_casing_accuracy": 0.5975058915946583,
75
+ "eval_loss": 0.0746106207370758,
76
+ "eval_overall_accuracy": 0.5965730557737627,
77
+ "eval_punctuation_accuracy": 0.5956402199528672,
78
+ "eval_runtime": 11.2558,
79
+ "eval_samples_per_second": 693.865,
80
+ "eval_steps_per_second": 43.444,
81
+ "step": 3906
82
+ }
83
+ ],
84
+ "logging_steps": 500,
85
+ "max_steps": 5859,
86
+ "num_input_tokens_seen": 0,
87
+ "num_train_epochs": 3,
88
+ "save_steps": 500,
89
+ "stateful_callbacks": {
90
+ "TrainerControl": {
91
+ "args": {
92
+ "should_epoch_stop": false,
93
+ "should_evaluate": false,
94
+ "should_log": false,
95
+ "should_save": true,
96
+ "should_training_stop": false
97
+ },
98
+ "attributes": {}
99
+ }
100
+ },
101
+ "total_flos": 0.0,
102
+ "train_batch_size": 16,
103
+ "trial_name": null,
104
+ "trial_params": null
105
+ }
checkpoint-3906/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:564d1f2f5aef116f928ab806870d35d30703b30b96f0b6e8ba67019dc262c956
3
+ size 5265
checkpoint-5859/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6b0593de29e7a6c178e88a9363450ae261bb62f4a2270ded7bb19bc2611f14a9
3
+ size 46781032
checkpoint-5859/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e8cba2752c31dd57eceb7b761ead10a5d9615e3a78a549ffbd7a0f54e5190eb9
3
+ size 88853790
checkpoint-5859/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bc889aeff6d9d6990d071a5d590b5ed43cf56675e6edb53189a2c60098beb746
3
+ size 14709
checkpoint-5859/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:82bc495f0e9d473971fac867a4529e9a1ecda97719ace06ceef20966f1cc19ea
3
+ size 1465
checkpoint-5859/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-5859/tokenizer_config.json ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": true,
3
+ "backend": "tokenizers",
4
+ "bos_token": "[CLS]",
5
+ "cls_token": "[CLS]",
6
+ "do_lower_case": true,
7
+ "eos_token": "[SEP]",
8
+ "is_local": false,
9
+ "keep_accents": false,
10
+ "mask_token": "[MASK]",
11
+ "model_max_length": 512,
12
+ "pad_token": "<pad>",
13
+ "sep_token": "[SEP]",
14
+ "tokenizer_class": "AlbertTokenizer",
15
+ "trim_offsets": true,
16
+ "unk_token": "<unk>"
17
+ }
checkpoint-5859/trainer_state.json ADDED
@@ -0,0 +1,144 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": 3906,
3
+ "best_metric": 0.0746106207370758,
4
+ "best_model_checkpoint": "./results_albert_punctuation_casing/checkpoint-3906",
5
+ "epoch": 3.0,
6
+ "eval_steps": 500,
7
+ "global_step": 5859,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.2560163850486431,
14
+ "grad_norm": 7.659163475036621,
15
+ "learning_rate": 1.8296637651476362e-05,
16
+ "loss": 0.17497076416015625,
17
+ "step": 500
18
+ },
19
+ {
20
+ "epoch": 0.5120327700972862,
21
+ "grad_norm": 9.602242469787598,
22
+ "learning_rate": 1.6589861751152075e-05,
23
+ "loss": 0.09840100860595703,
24
+ "step": 1000
25
+ },
26
+ {
27
+ "epoch": 0.7680491551459293,
28
+ "grad_norm": 2.2302889823913574,
29
+ "learning_rate": 1.4883085850827787e-05,
30
+ "loss": 0.08661653137207032,
31
+ "step": 1500
32
+ },
33
+ {
34
+ "epoch": 1.0,
35
+ "eval_casing_accuracy": 0.5982128829536528,
36
+ "eval_loss": 0.07590867578983307,
37
+ "eval_overall_accuracy": 0.597888845247447,
38
+ "eval_punctuation_accuracy": 0.5975648075412412,
39
+ "eval_runtime": 11.014,
40
+ "eval_samples_per_second": 709.096,
41
+ "eval_steps_per_second": 44.398,
42
+ "step": 1953
43
+ },
44
+ {
45
+ "epoch": 1.0240655401945724,
46
+ "grad_norm": 0.46477240324020386,
47
+ "learning_rate": 1.3176309950503501e-05,
48
+ "loss": 0.08205690765380859,
49
+ "step": 2000
50
+ },
51
+ {
52
+ "epoch": 1.2800819252432156,
53
+ "grad_norm": 6.8253865242004395,
54
+ "learning_rate": 1.1469534050179212e-05,
55
+ "loss": 0.0625931053161621,
56
+ "step": 2500
57
+ },
58
+ {
59
+ "epoch": 1.5360983102918588,
60
+ "grad_norm": 3.1819510459899902,
61
+ "learning_rate": 9.762758149854925e-06,
62
+ "loss": 0.06253623962402344,
63
+ "step": 3000
64
+ },
65
+ {
66
+ "epoch": 1.7921146953405018,
67
+ "grad_norm": 4.698487281799316,
68
+ "learning_rate": 8.055982249530637e-06,
69
+ "loss": 0.053002174377441404,
70
+ "step": 3500
71
+ },
72
+ {
73
+ "epoch": 2.0,
74
+ "eval_casing_accuracy": 0.5975058915946583,
75
+ "eval_loss": 0.0746106207370758,
76
+ "eval_overall_accuracy": 0.5965730557737627,
77
+ "eval_punctuation_accuracy": 0.5956402199528672,
78
+ "eval_runtime": 11.2558,
79
+ "eval_samples_per_second": 693.865,
80
+ "eval_steps_per_second": 43.444,
81
+ "step": 3906
82
+ },
83
+ {
84
+ "epoch": 2.048131080389145,
85
+ "grad_norm": 0.29059869050979614,
86
+ "learning_rate": 6.349206349206349e-06,
87
+ "loss": 0.057237117767333985,
88
+ "step": 4000
89
+ },
90
+ {
91
+ "epoch": 2.3041474654377883,
92
+ "grad_norm": 1.9593513011932373,
93
+ "learning_rate": 4.642430448882062e-06,
94
+ "loss": 0.03903835678100586,
95
+ "step": 4500
96
+ },
97
+ {
98
+ "epoch": 2.5601638504864312,
99
+ "grad_norm": 4.75918436050415,
100
+ "learning_rate": 2.9356545485577746e-06,
101
+ "loss": 0.03857281112670898,
102
+ "step": 5000
103
+ },
104
+ {
105
+ "epoch": 2.8161802355350742,
106
+ "grad_norm": 1.5033795833587646,
107
+ "learning_rate": 1.228878648233487e-06,
108
+ "loss": 0.03791716384887695,
109
+ "step": 5500
110
+ },
111
+ {
112
+ "epoch": 3.0,
113
+ "eval_casing_accuracy": 0.5970542026708563,
114
+ "eval_loss": 0.07589453458786011,
115
+ "eval_overall_accuracy": 0.5967596229379419,
116
+ "eval_punctuation_accuracy": 0.5964650432050275,
117
+ "eval_runtime": 11.3344,
118
+ "eval_samples_per_second": 689.056,
119
+ "eval_steps_per_second": 43.143,
120
+ "step": 5859
121
+ }
122
+ ],
123
+ "logging_steps": 500,
124
+ "max_steps": 5859,
125
+ "num_input_tokens_seen": 0,
126
+ "num_train_epochs": 3,
127
+ "save_steps": 500,
128
+ "stateful_callbacks": {
129
+ "TrainerControl": {
130
+ "args": {
131
+ "should_epoch_stop": false,
132
+ "should_evaluate": false,
133
+ "should_log": false,
134
+ "should_save": true,
135
+ "should_training_stop": true
136
+ },
137
+ "attributes": {}
138
+ }
139
+ },
140
+ "total_flos": 0.0,
141
+ "train_batch_size": 16,
142
+ "trial_name": null,
143
+ "trial_params": null
144
+ }
checkpoint-5859/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:564d1f2f5aef116f928ab806870d35d30703b30b96f0b6e8ba67019dc262c956
3
+ size 5265
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6b0593de29e7a6c178e88a9363450ae261bb62f4a2270ded7bb19bc2611f14a9
3
+ size 46781032
optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e8cba2752c31dd57eceb7b761ead10a5d9615e3a78a549ffbd7a0f54e5190eb9
3
+ size 88853790
rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bc889aeff6d9d6990d071a5d590b5ed43cf56675e6edb53189a2c60098beb746
3
+ size 14709
runs/Mar28_13-58-32_6078593f7de5/events.out.tfevents.1774706312.6078593f7de5.7992.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6e7a21151d221884ff4261a4a96a55175e7fc0b3f896c5f23eed46a5926d41c1
3
+ size 7650
scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:82bc495f0e9d473971fac867a4529e9a1ecda97719ace06ceef20966f1cc19ea
3
+ size 1465
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": true,
3
+ "backend": "tokenizers",
4
+ "bos_token": "[CLS]",
5
+ "cls_token": "[CLS]",
6
+ "do_lower_case": true,
7
+ "eos_token": "[SEP]",
8
+ "is_local": false,
9
+ "keep_accents": false,
10
+ "mask_token": "[MASK]",
11
+ "model_max_length": 512,
12
+ "pad_token": "<pad>",
13
+ "sep_token": "[SEP]",
14
+ "tokenizer_class": "AlbertTokenizer",
15
+ "trim_offsets": true,
16
+ "unk_token": "<unk>"
17
+ }
trainer_state.json ADDED
@@ -0,0 +1,144 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": 3906,
3
+ "best_metric": 0.0746106207370758,
4
+ "best_model_checkpoint": "./results_albert_punctuation_casing/checkpoint-3906",
5
+ "epoch": 3.0,
6
+ "eval_steps": 500,
7
+ "global_step": 5859,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.2560163850486431,
14
+ "grad_norm": 7.659163475036621,
15
+ "learning_rate": 1.8296637651476362e-05,
16
+ "loss": 0.17497076416015625,
17
+ "step": 500
18
+ },
19
+ {
20
+ "epoch": 0.5120327700972862,
21
+ "grad_norm": 9.602242469787598,
22
+ "learning_rate": 1.6589861751152075e-05,
23
+ "loss": 0.09840100860595703,
24
+ "step": 1000
25
+ },
26
+ {
27
+ "epoch": 0.7680491551459293,
28
+ "grad_norm": 2.2302889823913574,
29
+ "learning_rate": 1.4883085850827787e-05,
30
+ "loss": 0.08661653137207032,
31
+ "step": 1500
32
+ },
33
+ {
34
+ "epoch": 1.0,
35
+ "eval_casing_accuracy": 0.5982128829536528,
36
+ "eval_loss": 0.07590867578983307,
37
+ "eval_overall_accuracy": 0.597888845247447,
38
+ "eval_punctuation_accuracy": 0.5975648075412412,
39
+ "eval_runtime": 11.014,
40
+ "eval_samples_per_second": 709.096,
41
+ "eval_steps_per_second": 44.398,
42
+ "step": 1953
43
+ },
44
+ {
45
+ "epoch": 1.0240655401945724,
46
+ "grad_norm": 0.46477240324020386,
47
+ "learning_rate": 1.3176309950503501e-05,
48
+ "loss": 0.08205690765380859,
49
+ "step": 2000
50
+ },
51
+ {
52
+ "epoch": 1.2800819252432156,
53
+ "grad_norm": 6.8253865242004395,
54
+ "learning_rate": 1.1469534050179212e-05,
55
+ "loss": 0.0625931053161621,
56
+ "step": 2500
57
+ },
58
+ {
59
+ "epoch": 1.5360983102918588,
60
+ "grad_norm": 3.1819510459899902,
61
+ "learning_rate": 9.762758149854925e-06,
62
+ "loss": 0.06253623962402344,
63
+ "step": 3000
64
+ },
65
+ {
66
+ "epoch": 1.7921146953405018,
67
+ "grad_norm": 4.698487281799316,
68
+ "learning_rate": 8.055982249530637e-06,
69
+ "loss": 0.053002174377441404,
70
+ "step": 3500
71
+ },
72
+ {
73
+ "epoch": 2.0,
74
+ "eval_casing_accuracy": 0.5975058915946583,
75
+ "eval_loss": 0.0746106207370758,
76
+ "eval_overall_accuracy": 0.5965730557737627,
77
+ "eval_punctuation_accuracy": 0.5956402199528672,
78
+ "eval_runtime": 11.2558,
79
+ "eval_samples_per_second": 693.865,
80
+ "eval_steps_per_second": 43.444,
81
+ "step": 3906
82
+ },
83
+ {
84
+ "epoch": 2.048131080389145,
85
+ "grad_norm": 0.29059869050979614,
86
+ "learning_rate": 6.349206349206349e-06,
87
+ "loss": 0.057237117767333985,
88
+ "step": 4000
89
+ },
90
+ {
91
+ "epoch": 2.3041474654377883,
92
+ "grad_norm": 1.9593513011932373,
93
+ "learning_rate": 4.642430448882062e-06,
94
+ "loss": 0.03903835678100586,
95
+ "step": 4500
96
+ },
97
+ {
98
+ "epoch": 2.5601638504864312,
99
+ "grad_norm": 4.75918436050415,
100
+ "learning_rate": 2.9356545485577746e-06,
101
+ "loss": 0.03857281112670898,
102
+ "step": 5000
103
+ },
104
+ {
105
+ "epoch": 2.8161802355350742,
106
+ "grad_norm": 1.5033795833587646,
107
+ "learning_rate": 1.228878648233487e-06,
108
+ "loss": 0.03791716384887695,
109
+ "step": 5500
110
+ },
111
+ {
112
+ "epoch": 3.0,
113
+ "eval_casing_accuracy": 0.5970542026708563,
114
+ "eval_loss": 0.07589453458786011,
115
+ "eval_overall_accuracy": 0.5967596229379419,
116
+ "eval_punctuation_accuracy": 0.5964650432050275,
117
+ "eval_runtime": 11.3344,
118
+ "eval_samples_per_second": 689.056,
119
+ "eval_steps_per_second": 43.143,
120
+ "step": 5859
121
+ }
122
+ ],
123
+ "logging_steps": 500,
124
+ "max_steps": 5859,
125
+ "num_input_tokens_seen": 0,
126
+ "num_train_epochs": 3,
127
+ "save_steps": 500,
128
+ "stateful_callbacks": {
129
+ "TrainerControl": {
130
+ "args": {
131
+ "should_epoch_stop": false,
132
+ "should_evaluate": false,
133
+ "should_log": false,
134
+ "should_save": true,
135
+ "should_training_stop": true
136
+ },
137
+ "attributes": {}
138
+ }
139
+ },
140
+ "total_flos": 0.0,
141
+ "train_batch_size": 16,
142
+ "trial_name": null,
143
+ "trial_params": null
144
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:564d1f2f5aef116f928ab806870d35d30703b30b96f0b6e8ba67019dc262c956
3
+ size 5265