abarbosa commited on
Commit
d00a232
·
verified ·
1 Parent(s): 608101d

Pushing fine-tuned model to Hugging Face Hub

Browse files
README.md ADDED
@@ -0,0 +1,48 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ ---
3
+ language:
4
+ - pt
5
+ - en
6
+ tags:
7
+ - aes
8
+ datasets:
9
+ - kamel-usp/aes_enem_dataset
10
+ base_model: google-bert/bert-base-multilingual-cased
11
+ metrics:
12
+ - accuracy
13
+ - qwk
14
+ library_name: transformers
15
+ model-index:
16
+ - name: bert-base-multilingual-cased-encoder_classification-C5-essay_only
17
+ results:
18
+ - task:
19
+ type: text-classification
20
+ name: Automated Essay Score
21
+ dataset:
22
+ name: Automated Essay Score ENEM Dataset
23
+ type: kamel-usp/aes_enem_dataset
24
+ config: JBCS2025
25
+ split: test
26
+ metrics:
27
+ - name: Macro F1
28
+ type: f1
29
+ value: 0.0780476833108412
30
+ - name: QWK
31
+ type: qwk
32
+ value: 0.0042337002540219
33
+ - name: Weighted Macro F1
34
+ type: f1
35
+ value: 0.0895186684660368
36
+ ---
37
+ # Model ID: bert-base-multilingual-cased-encoder_classification-C5-essay_only
38
+ ## Results
39
+ | | test_data |
40
+ |:-----------------|------------:|
41
+ | eval_accuracy | 0.166667 |
42
+ | eval_RMSE | 95.3407 |
43
+ | eval_QWK | 0.0042337 |
44
+ | eval_Macro_F1 | 0.0780477 |
45
+ | eval_Weighted_F1 | 0.0895187 |
46
+ | eval_Micro_F1 | 0.166667 |
47
+ | eval_HDIV | 0.369565 |
48
+
config.json ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "BertForSequenceClassification"
4
+ ],
5
+ "attention_probs_dropout_prob": 0.1,
6
+ "classifier_dropout": null,
7
+ "directionality": "bidi",
8
+ "hidden_act": "gelu",
9
+ "hidden_dropout_prob": 0.1,
10
+ "hidden_size": 768,
11
+ "id2label": {
12
+ "0": 0,
13
+ "1": 40,
14
+ "2": 80,
15
+ "3": 120,
16
+ "4": 160,
17
+ "5": 200
18
+ },
19
+ "initializer_range": 0.02,
20
+ "intermediate_size": 3072,
21
+ "label2id": {
22
+ "0": 0,
23
+ "40": 1,
24
+ "80": 2,
25
+ "120": 3,
26
+ "160": 4,
27
+ "200": 5
28
+ },
29
+ "layer_norm_eps": 1e-12,
30
+ "max_position_embeddings": 512,
31
+ "model_type": "bert",
32
+ "num_attention_heads": 12,
33
+ "num_hidden_layers": 12,
34
+ "pad_token_id": 0,
35
+ "pooler_fc_size": 768,
36
+ "pooler_num_attention_heads": 12,
37
+ "pooler_num_fc_layers": 3,
38
+ "pooler_size_per_head": 128,
39
+ "pooler_type": "first_token_transform",
40
+ "position_embedding_type": "absolute",
41
+ "torch_dtype": "float32",
42
+ "transformers_version": "4.53.1",
43
+ "type_vocab_size": 2,
44
+ "use_cache": true,
45
+ "vocab_size": 119547
46
+ }
emissions.csv ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ timestamp,project_name,run_id,experiment_id,duration,emissions,emissions_rate,cpu_power,gpu_power,ram_power,cpu_energy,gpu_energy,ram_energy,energy_consumed,country_name,country_iso_code,region,cloud_provider,cloud_region,os,python_version,codecarbon_version,cpu_count,cpu_model,gpu_count,gpu_model,longitude,latitude,ram_total_size,tracking_mode,on_cloud,pue
2
+ 2025-07-09T17:19:26,jbcs2025,3fa00142-2406-457d-b59c-cd29d149447f,bert-base-multilingual-cased-encoder_classification-C5-essay_only,63.13267161301337,0.001186007782525365,1.878595903235776e-05,45.78333333333334,111.88458537532756,58.0,0.0007518834245264855,0.0031850239369060063,0.0009928166112728328,0.004929723972705325,Romania,ROU,gorj county,,,Linux-5.15.0-143-generic-x86_64-with-glibc2.35,3.12.11,3.0.2,36,Intel(R) Xeon(R) Gold 6248R CPU @ 3.00GHz,1,1 x NVIDIA RTX A6000,23.2904,45.0489,393.6063117980957,machine,N,1.0
evaluation_results.csv ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ eval_loss,eval_model_preparation_time,eval_accuracy,eval_RMSE,eval_QWK,eval_HDIV,eval_Macro_F1,eval_Micro_F1,eval_Weighted_F1,eval_TP_0,eval_TN_0,eval_FP_0,eval_FN_0,eval_TP_1,eval_TN_1,eval_FP_1,eval_FN_1,eval_TP_2,eval_TN_2,eval_FP_2,eval_FN_2,eval_TP_3,eval_TN_3,eval_FP_3,eval_FN_3,eval_TP_4,eval_TN_4,eval_FP_4,eval_FN_4,eval_TP_5,eval_TN_5,eval_FP_5,eval_FN_5,eval_runtime,eval_samples_per_second,eval_steps_per_second,epoch,reference,timestamp,id
2
+ 1.8017789125442505,0.0021,0.11363636363636363,100.36297759994605,0.026102706158107614,0.43181818181818177,0.03968253968253968,0.11363636363636363,0.04329004329004328,0,115,0,17,0,96,0,36,0,112,0,20,0,99,0,33,15,21,87,9,0,100,30,2,0.6139,215.012,14.66,-1,validation_before_training,2025-07-09 17:18:31,bert-base-multilingual-cased-encoder_classification-C5-essay_only
3
+ 1.7867826223373413,0.0021,0.20454545454545456,84.63826772587187,0.05513298774284192,0.2803030303030303,0.09654954099398544,0.20454545454545456,0.11889550778439668,11,62,53,6,16,44,52,20,0,112,0,20,0,99,0,33,0,108,0,24,0,130,0,2,0.3875,340.632,23.225,6.0,validation_after_training,2025-07-09 17:18:31,bert-base-multilingual-cased-encoder_classification-C5-essay_only
4
+ 1.7879478931427002,0.0021,0.16666666666666666,95.34073144497984,0.0042337002540219215,0.3695652173913043,0.0780476833108412,0.16666666666666666,0.08951866846603689,15,39,77,7,8,68,38,24,0,114,0,24,0,113,0,25,0,106,0,32,0,135,0,3,0.4106,336.133,21.922,6.0,test_results,2025-07-09 17:18:31,bert-base-multilingual-cased-encoder_classification-C5-essay_only
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:802d00810c3dcf85133e95125cda27d24537b51a3b9297a2de290aa9479194ac
3
+ size 711455760
run_experiment.log ADDED
@@ -0,0 +1,324 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2025-07-09 17:18:20,006][__main__][INFO] - cache_dir: /tmp/
2
+ dataset:
3
+ name: kamel-usp/aes_enem_dataset
4
+ split: JBCS2025
5
+ training_params:
6
+ seed: 42
7
+ num_train_epochs: 20
8
+ logging_steps: 100
9
+ metric_for_best_model: QWK
10
+ bf16: true
11
+ bootstrap:
12
+ enabled: true
13
+ n_bootstrap: 10000
14
+ bootstrap_seed: 42
15
+ metrics:
16
+ - QWK
17
+ - Macro_F1
18
+ - Weighted_F1
19
+ post_training_results:
20
+ model_path: /workspace/jbcs2025/outputs/2025-03-24/20-42-59
21
+ experiments:
22
+ model:
23
+ name: google-bert/bert-base-multilingual-cased
24
+ type: encoder_classification
25
+ num_labels: 6
26
+ output_dir: ./results/
27
+ logging_dir: ./logs/
28
+ best_model_dir: ./results/best_model
29
+ tokenizer:
30
+ name: google-bert/bert-base-multilingual-cased
31
+ dataset:
32
+ grade_index: 4
33
+ use_full_context: false
34
+ training_params:
35
+ weight_decay: 0.01
36
+ warmup_ratio: 0.1
37
+ learning_rate: 5.0e-05
38
+ train_batch_size: 16
39
+ eval_batch_size: 16
40
+ gradient_accumulation_steps: 1
41
+ gradient_checkpointing: false
42
+
43
+ [2025-07-09 17:18:23,847][__main__][INFO] - GPU 0: NVIDIA RTX A6000 | TDP ≈ 300 W
44
+ [2025-07-09 17:18:23,847][__main__][INFO] - Starting the Fine Tuning training process.
45
+ [2025-07-09 17:18:28,407][transformers.configuration_utils][INFO] - loading configuration file config.json from cache at /tmp/models--google-bert--bert-base-multilingual-cased/snapshots/3f076fdb1ab68d5b2880cb87a0886f315b8146f8/config.json
46
+ [2025-07-09 17:18:28,408][transformers.configuration_utils][INFO] - Model config BertConfig {
47
+ "architectures": [
48
+ "BertForMaskedLM"
49
+ ],
50
+ "attention_probs_dropout_prob": 0.1,
51
+ "classifier_dropout": null,
52
+ "directionality": "bidi",
53
+ "hidden_act": "gelu",
54
+ "hidden_dropout_prob": 0.1,
55
+ "hidden_size": 768,
56
+ "initializer_range": 0.02,
57
+ "intermediate_size": 3072,
58
+ "layer_norm_eps": 1e-12,
59
+ "max_position_embeddings": 512,
60
+ "model_type": "bert",
61
+ "num_attention_heads": 12,
62
+ "num_hidden_layers": 12,
63
+ "pad_token_id": 0,
64
+ "pooler_fc_size": 768,
65
+ "pooler_num_attention_heads": 12,
66
+ "pooler_num_fc_layers": 3,
67
+ "pooler_size_per_head": 128,
68
+ "pooler_type": "first_token_transform",
69
+ "position_embedding_type": "absolute",
70
+ "transformers_version": "4.53.1",
71
+ "type_vocab_size": 2,
72
+ "use_cache": true,
73
+ "vocab_size": 119547
74
+ }
75
+
76
+ [2025-07-09 17:18:28,606][transformers.configuration_utils][INFO] - loading configuration file config.json from cache at /tmp/models--google-bert--bert-base-multilingual-cased/snapshots/3f076fdb1ab68d5b2880cb87a0886f315b8146f8/config.json
77
+ [2025-07-09 17:18:28,607][transformers.configuration_utils][INFO] - Model config BertConfig {
78
+ "architectures": [
79
+ "BertForMaskedLM"
80
+ ],
81
+ "attention_probs_dropout_prob": 0.1,
82
+ "classifier_dropout": null,
83
+ "directionality": "bidi",
84
+ "hidden_act": "gelu",
85
+ "hidden_dropout_prob": 0.1,
86
+ "hidden_size": 768,
87
+ "initializer_range": 0.02,
88
+ "intermediate_size": 3072,
89
+ "layer_norm_eps": 1e-12,
90
+ "max_position_embeddings": 512,
91
+ "model_type": "bert",
92
+ "num_attention_heads": 12,
93
+ "num_hidden_layers": 12,
94
+ "pad_token_id": 0,
95
+ "pooler_fc_size": 768,
96
+ "pooler_num_attention_heads": 12,
97
+ "pooler_num_fc_layers": 3,
98
+ "pooler_size_per_head": 128,
99
+ "pooler_type": "first_token_transform",
100
+ "position_embedding_type": "absolute",
101
+ "transformers_version": "4.53.1",
102
+ "type_vocab_size": 2,
103
+ "use_cache": true,
104
+ "vocab_size": 119547
105
+ }
106
+
107
+ [2025-07-09 17:18:28,793][transformers.tokenization_utils_base][INFO] - loading file vocab.txt from cache at /tmp/models--google-bert--bert-base-multilingual-cased/snapshots/3f076fdb1ab68d5b2880cb87a0886f315b8146f8/vocab.txt
108
+ [2025-07-09 17:18:28,794][transformers.tokenization_utils_base][INFO] - loading file tokenizer.json from cache at /tmp/models--google-bert--bert-base-multilingual-cased/snapshots/3f076fdb1ab68d5b2880cb87a0886f315b8146f8/tokenizer.json
109
+ [2025-07-09 17:18:28,794][transformers.tokenization_utils_base][INFO] - loading file added_tokens.json from cache at None
110
+ [2025-07-09 17:18:28,794][transformers.tokenization_utils_base][INFO] - loading file special_tokens_map.json from cache at None
111
+ [2025-07-09 17:18:28,794][transformers.tokenization_utils_base][INFO] - loading file tokenizer_config.json from cache at /tmp/models--google-bert--bert-base-multilingual-cased/snapshots/3f076fdb1ab68d5b2880cb87a0886f315b8146f8/tokenizer_config.json
112
+ [2025-07-09 17:18:28,794][transformers.tokenization_utils_base][INFO] - loading file chat_template.jinja from cache at None
113
+ [2025-07-09 17:18:28,794][transformers.configuration_utils][INFO] - loading configuration file config.json from cache at /tmp/models--google-bert--bert-base-multilingual-cased/snapshots/3f076fdb1ab68d5b2880cb87a0886f315b8146f8/config.json
114
+ [2025-07-09 17:18:28,794][transformers.configuration_utils][INFO] - Model config BertConfig {
115
+ "architectures": [
116
+ "BertForMaskedLM"
117
+ ],
118
+ "attention_probs_dropout_prob": 0.1,
119
+ "classifier_dropout": null,
120
+ "directionality": "bidi",
121
+ "hidden_act": "gelu",
122
+ "hidden_dropout_prob": 0.1,
123
+ "hidden_size": 768,
124
+ "initializer_range": 0.02,
125
+ "intermediate_size": 3072,
126
+ "layer_norm_eps": 1e-12,
127
+ "max_position_embeddings": 512,
128
+ "model_type": "bert",
129
+ "num_attention_heads": 12,
130
+ "num_hidden_layers": 12,
131
+ "pad_token_id": 0,
132
+ "pooler_fc_size": 768,
133
+ "pooler_num_attention_heads": 12,
134
+ "pooler_num_fc_layers": 3,
135
+ "pooler_size_per_head": 128,
136
+ "pooler_type": "first_token_transform",
137
+ "position_embedding_type": "absolute",
138
+ "transformers_version": "4.53.1",
139
+ "type_vocab_size": 2,
140
+ "use_cache": true,
141
+ "vocab_size": 119547
142
+ }
143
+
144
+ [2025-07-09 17:18:29,028][__main__][INFO] - Tokenizer function parameters- Padding:longest; Truncation: True; Use Full Context: False
145
+ [2025-07-09 17:18:29,705][__main__][INFO] -
146
+ Token statistics for 'train' split:
147
+ [2025-07-09 17:18:29,705][__main__][INFO] - Total examples: 500
148
+ [2025-07-09 17:18:29,705][__main__][INFO] - Min tokens: 512
149
+ [2025-07-09 17:18:29,705][__main__][INFO] - Max tokens: 512
150
+ [2025-07-09 17:18:29,705][__main__][INFO] - Avg tokens: 512.00
151
+ [2025-07-09 17:18:29,705][__main__][INFO] - Std tokens: 0.00
152
+ [2025-07-09 17:18:29,801][__main__][INFO] -
153
+ Token statistics for 'validation' split:
154
+ [2025-07-09 17:18:29,801][__main__][INFO] - Total examples: 132
155
+ [2025-07-09 17:18:29,801][__main__][INFO] - Min tokens: 512
156
+ [2025-07-09 17:18:29,801][__main__][INFO] - Max tokens: 512
157
+ [2025-07-09 17:18:29,801][__main__][INFO] - Avg tokens: 512.00
158
+ [2025-07-09 17:18:29,801][__main__][INFO] - Std tokens: 0.00
159
+ [2025-07-09 17:18:29,900][__main__][INFO] -
160
+ Token statistics for 'test' split:
161
+ [2025-07-09 17:18:29,900][__main__][INFO] - Total examples: 138
162
+ [2025-07-09 17:18:29,900][__main__][INFO] - Min tokens: 512
163
+ [2025-07-09 17:18:29,900][__main__][INFO] - Max tokens: 512
164
+ [2025-07-09 17:18:29,900][__main__][INFO] - Avg tokens: 512.00
165
+ [2025-07-09 17:18:29,900][__main__][INFO] - Std tokens: 0.00
166
+ [2025-07-09 17:18:29,900][__main__][INFO] - If token statistics are the same (max, avg, min) keep in mind that this is due to batched tokenization and padding.
167
+ [2025-07-09 17:18:29,900][__main__][INFO] - Model max length: 512. If it is the same as stats, then there is a high chance that sequences are being truncated.
168
+ [2025-07-09 17:18:30,107][transformers.configuration_utils][INFO] - loading configuration file config.json from cache at /tmp/models--google-bert--bert-base-multilingual-cased/snapshots/3f076fdb1ab68d5b2880cb87a0886f315b8146f8/config.json
169
+ [2025-07-09 17:18:30,108][transformers.configuration_utils][INFO] - Model config BertConfig {
170
+ "architectures": [
171
+ "BertForMaskedLM"
172
+ ],
173
+ "attention_probs_dropout_prob": 0.1,
174
+ "classifier_dropout": null,
175
+ "directionality": "bidi",
176
+ "hidden_act": "gelu",
177
+ "hidden_dropout_prob": 0.1,
178
+ "hidden_size": 768,
179
+ "id2label": {
180
+ "0": 0,
181
+ "1": 40,
182
+ "2": 80,
183
+ "3": 120,
184
+ "4": 160,
185
+ "5": 200
186
+ },
187
+ "initializer_range": 0.02,
188
+ "intermediate_size": 3072,
189
+ "label2id": {
190
+ "0": 0,
191
+ "40": 1,
192
+ "80": 2,
193
+ "120": 3,
194
+ "160": 4,
195
+ "200": 5
196
+ },
197
+ "layer_norm_eps": 1e-12,
198
+ "max_position_embeddings": 512,
199
+ "model_type": "bert",
200
+ "num_attention_heads": 12,
201
+ "num_hidden_layers": 12,
202
+ "pad_token_id": 0,
203
+ "pooler_fc_size": 768,
204
+ "pooler_num_attention_heads": 12,
205
+ "pooler_num_fc_layers": 3,
206
+ "pooler_size_per_head": 128,
207
+ "pooler_type": "first_token_transform",
208
+ "position_embedding_type": "absolute",
209
+ "transformers_version": "4.53.1",
210
+ "type_vocab_size": 2,
211
+ "use_cache": true,
212
+ "vocab_size": 119547
213
+ }
214
+
215
+ [2025-07-09 17:18:30,282][transformers.modeling_utils][INFO] - loading weights file model.safetensors from cache at /tmp/models--google-bert--bert-base-multilingual-cased/snapshots/3f076fdb1ab68d5b2880cb87a0886f315b8146f8/model.safetensors
216
+ [2025-07-09 17:18:30,289][transformers.modeling_utils][INFO] - Since the `torch_dtype` attribute can't be found in model's config object, will use torch_dtype={torch_dtype} as derived from model's weights
217
+ [2025-07-09 17:18:30,289][transformers.modeling_utils][INFO] - Instantiating BertForSequenceClassification model under default dtype torch.float32.
218
+ [2025-07-09 17:18:30,475][transformers.modeling_utils][INFO] - A pretrained model of type `BertForSequenceClassification` contains parameters that have been renamed internally (a few are listed below but more are present in the model):
219
+ * `cls.predictions.transform.LayerNorm.beta` -> `cls.predictions.transform.LayerNorm.bias`
220
+ * `cls.predictions.transform.LayerNorm.gamma` -> `cls.predictions.transform.LayerNorm.weight`
221
+ If you are using a model from the Hub, consider submitting a PR to adjust these weights and help future users.
222
+ [2025-07-09 17:18:31,287][transformers.modeling_utils][INFO] - Some weights of the model checkpoint at google-bert/bert-base-multilingual-cased were not used when initializing BertForSequenceClassification: ['cls.predictions.bias', 'cls.predictions.transform.LayerNorm.bias', 'cls.predictions.transform.LayerNorm.weight', 'cls.predictions.transform.dense.bias', 'cls.predictions.transform.dense.weight', 'cls.seq_relationship.bias', 'cls.seq_relationship.weight']
223
+ - This IS expected if you are initializing BertForSequenceClassification from the checkpoint of a model trained on another task or with another architecture (e.g. initializing a BertForSequenceClassification model from a BertForPreTraining model).
224
+ - This IS NOT expected if you are initializing BertForSequenceClassification from the checkpoint of a model that you expect to be exactly identical (initializing a BertForSequenceClassification model from a BertForSequenceClassification model).
225
+ [2025-07-09 17:18:31,287][transformers.modeling_utils][WARNING] - Some weights of BertForSequenceClassification were not initialized from the model checkpoint at google-bert/bert-base-multilingual-cased and are newly initialized: ['classifier.bias', 'classifier.weight']
226
+ You should probably TRAIN this model on a down-stream task to be able to use it for predictions and inference.
227
+ [2025-07-09 17:18:31,301][transformers.training_args][INFO] - PyTorch: setting up devices
228
+ [2025-07-09 17:18:31,322][__main__][INFO] - Total steps: 620. Number of warmup steps: 62
229
+ [2025-07-09 17:18:31,330][transformers.trainer][INFO] - You have loaded a model on multiple GPUs. `is_model_parallel` attribute will be force-set to `True` to avoid any unexpected behavior such as device placement mismatching.
230
+ [2025-07-09 17:18:31,358][transformers.trainer][INFO] - Using auto half precision backend
231
+ [2025-07-09 17:18:31,360][transformers.trainer][INFO] - The following columns in the Evaluation set don't have a corresponding argument in `BertForSequenceClassification.forward` and have been ignored: id_prompt, reference, prompt, supporting_text, essay_text, id, grades, essay_year. If id_prompt, reference, prompt, supporting_text, essay_text, id, grades, essay_year are not expected by `BertForSequenceClassification.forward`, you can safely ignore this message.
232
+ [2025-07-09 17:18:31,365][transformers.trainer][INFO] -
233
+ ***** Running Evaluation *****
234
+ [2025-07-09 17:18:31,365][transformers.trainer][INFO] - Num examples = 132
235
+ [2025-07-09 17:18:31,365][transformers.trainer][INFO] - Batch size = 16
236
+ [2025-07-09 17:18:32,236][transformers.trainer][INFO] - The following columns in the Training set don't have a corresponding argument in `BertForSequenceClassification.forward` and have been ignored: id_prompt, reference, prompt, supporting_text, essay_text, id, grades, essay_year. If id_prompt, reference, prompt, supporting_text, essay_text, id, grades, essay_year are not expected by `BertForSequenceClassification.forward`, you can safely ignore this message.
237
+ [2025-07-09 17:18:32,245][transformers.trainer][INFO] - ***** Running training *****
238
+ [2025-07-09 17:18:32,245][transformers.trainer][INFO] - Num examples = 500
239
+ [2025-07-09 17:18:32,245][transformers.trainer][INFO] - Num Epochs = 20
240
+ [2025-07-09 17:18:32,245][transformers.trainer][INFO] - Instantaneous batch size per device = 16
241
+ [2025-07-09 17:18:32,245][transformers.trainer][INFO] - Total train batch size (w. parallel, distributed & accumulation) = 16
242
+ [2025-07-09 17:18:32,245][transformers.trainer][INFO] - Gradient Accumulation steps = 1
243
+ [2025-07-09 17:18:32,245][transformers.trainer][INFO] - Total optimization steps = 640
244
+ [2025-07-09 17:18:32,245][transformers.trainer][INFO] - Number of trainable parameters = 177,858,054
245
+ [2025-07-09 17:18:37,175][transformers.trainer][INFO] - The following columns in the Evaluation set don't have a corresponding argument in `BertForSequenceClassification.forward` and have been ignored: id_prompt, reference, prompt, supporting_text, essay_text, id, grades, essay_year. If id_prompt, reference, prompt, supporting_text, essay_text, id, grades, essay_year are not expected by `BertForSequenceClassification.forward`, you can safely ignore this message.
246
+ [2025-07-09 17:18:37,178][transformers.trainer][INFO] -
247
+ ***** Running Evaluation *****
248
+ [2025-07-09 17:18:37,178][transformers.trainer][INFO] - Num examples = 132
249
+ [2025-07-09 17:18:37,178][transformers.trainer][INFO] - Batch size = 16
250
+ [2025-07-09 17:18:37,562][transformers.trainer][INFO] - Saving model checkpoint to /workspace/jbcs2025/outputs/2025-07-09/17-18-19/results/checkpoint-32
251
+ [2025-07-09 17:18:37,564][transformers.configuration_utils][INFO] - Configuration saved in /workspace/jbcs2025/outputs/2025-07-09/17-18-19/results/checkpoint-32/config.json
252
+ [2025-07-09 17:18:38,812][transformers.modeling_utils][INFO] - Model weights saved in /workspace/jbcs2025/outputs/2025-07-09/17-18-19/results/checkpoint-32/model.safetensors
253
+ [2025-07-09 17:18:45,673][transformers.trainer][INFO] - The following columns in the Evaluation set don't have a corresponding argument in `BertForSequenceClassification.forward` and have been ignored: id_prompt, reference, prompt, supporting_text, essay_text, id, grades, essay_year. If id_prompt, reference, prompt, supporting_text, essay_text, id, grades, essay_year are not expected by `BertForSequenceClassification.forward`, you can safely ignore this message.
254
+ [2025-07-09 17:18:45,676][transformers.trainer][INFO] -
255
+ ***** Running Evaluation *****
256
+ [2025-07-09 17:18:45,676][transformers.trainer][INFO] - Num examples = 132
257
+ [2025-07-09 17:18:45,676][transformers.trainer][INFO] - Batch size = 16
258
+ [2025-07-09 17:18:46,063][transformers.trainer][INFO] - Saving model checkpoint to /workspace/jbcs2025/outputs/2025-07-09/17-18-19/results/checkpoint-64
259
+ [2025-07-09 17:18:46,064][transformers.configuration_utils][INFO] - Configuration saved in /workspace/jbcs2025/outputs/2025-07-09/17-18-19/results/checkpoint-64/config.json
260
+ [2025-07-09 17:18:47,352][transformers.modeling_utils][INFO] - Model weights saved in /workspace/jbcs2025/outputs/2025-07-09/17-18-19/results/checkpoint-64/model.safetensors
261
+ [2025-07-09 17:18:54,206][transformers.trainer][INFO] - The following columns in the Evaluation set don't have a corresponding argument in `BertForSequenceClassification.forward` and have been ignored: id_prompt, reference, prompt, supporting_text, essay_text, id, grades, essay_year. If id_prompt, reference, prompt, supporting_text, essay_text, id, grades, essay_year are not expected by `BertForSequenceClassification.forward`, you can safely ignore this message.
262
+ [2025-07-09 17:18:54,209][transformers.trainer][INFO] -
263
+ ***** Running Evaluation *****
264
+ [2025-07-09 17:18:54,209][transformers.trainer][INFO] - Num examples = 132
265
+ [2025-07-09 17:18:54,209][transformers.trainer][INFO] - Batch size = 16
266
+ [2025-07-09 17:18:54,635][transformers.trainer][INFO] - Saving model checkpoint to /workspace/jbcs2025/outputs/2025-07-09/17-18-19/results/checkpoint-96
267
+ [2025-07-09 17:18:54,636][transformers.configuration_utils][INFO] - Configuration saved in /workspace/jbcs2025/outputs/2025-07-09/17-18-19/results/checkpoint-96/config.json
268
+ [2025-07-09 17:18:55,771][transformers.modeling_utils][INFO] - Model weights saved in /workspace/jbcs2025/outputs/2025-07-09/17-18-19/results/checkpoint-96/model.safetensors
269
+ [2025-07-09 17:18:57,427][transformers.trainer][INFO] - Deleting older checkpoint [/workspace/jbcs2025/outputs/2025-07-09/17-18-19/results/checkpoint-64] due to args.save_total_limit
270
+ [2025-07-09 17:19:02,405][transformers.trainer][INFO] - The following columns in the Evaluation set don't have a corresponding argument in `BertForSequenceClassification.forward` and have been ignored: id_prompt, reference, prompt, supporting_text, essay_text, id, grades, essay_year. If id_prompt, reference, prompt, supporting_text, essay_text, id, grades, essay_year are not expected by `BertForSequenceClassification.forward`, you can safely ignore this message.
271
+ [2025-07-09 17:19:02,407][transformers.trainer][INFO] -
272
+ ***** Running Evaluation *****
273
+ [2025-07-09 17:19:02,408][transformers.trainer][INFO] - Num examples = 132
274
+ [2025-07-09 17:19:02,408][transformers.trainer][INFO] - Batch size = 16
275
+ [2025-07-09 17:19:02,795][transformers.trainer][INFO] - Saving model checkpoint to /workspace/jbcs2025/outputs/2025-07-09/17-18-19/results/checkpoint-128
276
+ [2025-07-09 17:19:02,797][transformers.configuration_utils][INFO] - Configuration saved in /workspace/jbcs2025/outputs/2025-07-09/17-18-19/results/checkpoint-128/config.json
277
+ [2025-07-09 17:19:03,916][transformers.modeling_utils][INFO] - Model weights saved in /workspace/jbcs2025/outputs/2025-07-09/17-18-19/results/checkpoint-128/model.safetensors
278
+ [2025-07-09 17:19:05,535][transformers.trainer][INFO] - Deleting older checkpoint [/workspace/jbcs2025/outputs/2025-07-09/17-18-19/results/checkpoint-96] due to args.save_total_limit
279
+ [2025-07-09 17:19:10,434][transformers.trainer][INFO] - The following columns in the Evaluation set don't have a corresponding argument in `BertForSequenceClassification.forward` and have been ignored: id_prompt, reference, prompt, supporting_text, essay_text, id, grades, essay_year. If id_prompt, reference, prompt, supporting_text, essay_text, id, grades, essay_year are not expected by `BertForSequenceClassification.forward`, you can safely ignore this message.
280
+ [2025-07-09 17:19:10,437][transformers.trainer][INFO] -
281
+ ***** Running Evaluation *****
282
+ [2025-07-09 17:19:10,437][transformers.trainer][INFO] - Num examples = 132
283
+ [2025-07-09 17:19:10,437][transformers.trainer][INFO] - Batch size = 16
284
+ [2025-07-09 17:19:10,843][transformers.trainer][INFO] - Saving model checkpoint to /workspace/jbcs2025/outputs/2025-07-09/17-18-19/results/checkpoint-160
285
+ [2025-07-09 17:19:10,845][transformers.configuration_utils][INFO] - Configuration saved in /workspace/jbcs2025/outputs/2025-07-09/17-18-19/results/checkpoint-160/config.json
286
+ [2025-07-09 17:19:12,031][transformers.modeling_utils][INFO] - Model weights saved in /workspace/jbcs2025/outputs/2025-07-09/17-18-19/results/checkpoint-160/model.safetensors
287
+ [2025-07-09 17:19:13,744][transformers.trainer][INFO] - Deleting older checkpoint [/workspace/jbcs2025/outputs/2025-07-09/17-18-19/results/checkpoint-128] due to args.save_total_limit
288
+ [2025-07-09 17:19:18,745][transformers.trainer][INFO] - The following columns in the Evaluation set don't have a corresponding argument in `BertForSequenceClassification.forward` and have been ignored: id_prompt, reference, prompt, supporting_text, essay_text, id, grades, essay_year. If id_prompt, reference, prompt, supporting_text, essay_text, id, grades, essay_year are not expected by `BertForSequenceClassification.forward`, you can safely ignore this message.
289
+ [2025-07-09 17:19:18,749][transformers.trainer][INFO] -
290
+ ***** Running Evaluation *****
291
+ [2025-07-09 17:19:18,749][transformers.trainer][INFO] - Num examples = 132
292
+ [2025-07-09 17:19:18,749][transformers.trainer][INFO] - Batch size = 16
293
+ [2025-07-09 17:19:19,164][transformers.trainer][INFO] - Saving model checkpoint to /workspace/jbcs2025/outputs/2025-07-09/17-18-19/results/checkpoint-192
294
+ [2025-07-09 17:19:19,165][transformers.configuration_utils][INFO] - Configuration saved in /workspace/jbcs2025/outputs/2025-07-09/17-18-19/results/checkpoint-192/config.json
295
+ [2025-07-09 17:19:20,821][transformers.modeling_utils][INFO] - Model weights saved in /workspace/jbcs2025/outputs/2025-07-09/17-18-19/results/checkpoint-192/model.safetensors
296
+ [2025-07-09 17:19:22,309][transformers.trainer][INFO] - Deleting older checkpoint [/workspace/jbcs2025/outputs/2025-07-09/17-18-19/results/checkpoint-160] due to args.save_total_limit
297
+ [2025-07-09 17:19:22,420][transformers.trainer][INFO] -
298
+
299
+ Training completed. Do not forget to share your model on huggingface.co/models =)
300
+
301
+
302
+ [2025-07-09 17:19:22,421][transformers.trainer][INFO] - Loading best model from /workspace/jbcs2025/outputs/2025-07-09/17-18-19/results/checkpoint-32 (score: 0.05513298774284192).
303
+ [2025-07-09 17:19:23,340][transformers.trainer][INFO] - Deleting older checkpoint [/workspace/jbcs2025/outputs/2025-07-09/17-18-19/results/checkpoint-192] due to args.save_total_limit
304
+ [2025-07-09 17:19:23,475][transformers.trainer][INFO] - The following columns in the Evaluation set don't have a corresponding argument in `BertForSequenceClassification.forward` and have been ignored: id_prompt, reference, prompt, supporting_text, essay_text, id, grades, essay_year. If id_prompt, reference, prompt, supporting_text, essay_text, id, grades, essay_year are not expected by `BertForSequenceClassification.forward`, you can safely ignore this message.
305
+ [2025-07-09 17:19:23,478][transformers.trainer][INFO] -
306
+ ***** Running Evaluation *****
307
+ [2025-07-09 17:19:23,479][transformers.trainer][INFO] - Num examples = 132
308
+ [2025-07-09 17:19:23,479][transformers.trainer][INFO] - Batch size = 16
309
+ [2025-07-09 17:19:23,874][__main__][INFO] - Training completed successfully.
310
+ [2025-07-09 17:19:23,874][__main__][INFO] - Running on Test
311
+ [2025-07-09 17:19:23,874][transformers.trainer][INFO] - The following columns in the Evaluation set don't have a corresponding argument in `BertForSequenceClassification.forward` and have been ignored: id_prompt, reference, prompt, supporting_text, essay_text, id, grades, essay_year. If id_prompt, reference, prompt, supporting_text, essay_text, id, grades, essay_year are not expected by `BertForSequenceClassification.forward`, you can safely ignore this message.
312
+ [2025-07-09 17:19:23,877][transformers.trainer][INFO] -
313
+ ***** Running Evaluation *****
314
+ [2025-07-09 17:19:23,877][transformers.trainer][INFO] - Num examples = 138
315
+ [2025-07-09 17:19:23,877][transformers.trainer][INFO] - Batch size = 16
316
+ [2025-07-09 17:19:24,292][__main__][INFO] - Test metrics: {'eval_loss': 1.7879478931427002, 'eval_model_preparation_time': 0.0021, 'eval_accuracy': 0.16666666666666666, 'eval_RMSE': 95.34073144497984, 'eval_QWK': 0.0042337002540219215, 'eval_HDIV': 0.3695652173913043, 'eval_Macro_F1': 0.0780476833108412, 'eval_Micro_F1': 0.16666666666666666, 'eval_Weighted_F1': 0.08951866846603689, 'eval_TP_0': 15, 'eval_TN_0': 39, 'eval_FP_0': 77, 'eval_FN_0': 7, 'eval_TP_1': 8, 'eval_TN_1': 68, 'eval_FP_1': 38, 'eval_FN_1': 24, 'eval_TP_2': 0, 'eval_TN_2': 114, 'eval_FP_2': 0, 'eval_FN_2': 24, 'eval_TP_3': 0, 'eval_TN_3': 113, 'eval_FP_3': 0, 'eval_FN_3': 25, 'eval_TP_4': 0, 'eval_TN_4': 106, 'eval_FP_4': 0, 'eval_FN_4': 32, 'eval_TP_5': 0, 'eval_TN_5': 135, 'eval_FP_5': 0, 'eval_FN_5': 3, 'eval_runtime': 0.4106, 'eval_samples_per_second': 336.133, 'eval_steps_per_second': 21.922, 'epoch': 6.0}
317
+ [2025-07-09 17:19:24,293][transformers.trainer][INFO] - Saving model checkpoint to ./results/best_model
318
+ [2025-07-09 17:19:24,295][transformers.configuration_utils][INFO] - Configuration saved in ./results/best_model/config.json
319
+ [2025-07-09 17:19:25,852][transformers.modeling_utils][INFO] - Model weights saved in ./results/best_model/model.safetensors
320
+ [2025-07-09 17:19:25,854][transformers.tokenization_utils_base][INFO] - tokenizer config file saved in ./results/best_model/tokenizer_config.json
321
+ [2025-07-09 17:19:25,855][transformers.tokenization_utils_base][INFO] - Special tokens file saved in ./results/best_model/special_tokens_map.json
322
+ [2025-07-09 17:19:25,943][__main__][INFO] - Model and tokenizer saved to ./results/best_model
323
+ [2025-07-09 17:19:25,968][__main__][INFO] - Fine Tuning Finished.
324
+ [2025-07-09 17:19:26,478][__main__][INFO] - Total emissions: 0.0012 kg CO2eq
special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,56 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "100": {
12
+ "content": "[UNK]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "101": {
20
+ "content": "[CLS]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "102": {
28
+ "content": "[SEP]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "103": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "clean_up_tokenization_spaces": false,
45
+ "cls_token": "[CLS]",
46
+ "do_lower_case": false,
47
+ "extra_special_tokens": {},
48
+ "mask_token": "[MASK]",
49
+ "model_max_length": 512,
50
+ "pad_token": "[PAD]",
51
+ "sep_token": "[SEP]",
52
+ "strip_accents": null,
53
+ "tokenize_chinese_chars": true,
54
+ "tokenizer_class": "BertTokenizer",
55
+ "unk_token": "[UNK]"
56
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b7bba114cd668e472e584ecb86303849e427c062e7b9b9ecbe722e353cdaf6ab
3
+ size 5777
vocab.txt ADDED
The diff for this file is too large to render. See raw diff