abarbosa commited on
Commit
b8c6399
·
verified ·
1 Parent(s): d635e1f

Pushing fine-tuned model to Hugging Face Hub

Browse files
README.md ADDED
@@ -0,0 +1,48 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ ---
3
+ language:
4
+ - pt
5
+ - en
6
+ tags:
7
+ - aes
8
+ datasets:
9
+ - kamel-usp/aes_enem_dataset
10
+ base_model: google-bert/bert-base-multilingual-cased
11
+ metrics:
12
+ - accuracy
13
+ - qwk
14
+ library_name: transformers
15
+ model-index:
16
+ - name: bert-base-multilingual-cased-encoder_classification-C4-essay_only
17
+ results:
18
+ - task:
19
+ type: text-classification
20
+ name: Automated Essay Score
21
+ dataset:
22
+ name: Automated Essay Score ENEM Dataset
23
+ type: kamel-usp/aes_enem_dataset
24
+ config: JBCS2025
25
+ split: test
26
+ metrics:
27
+ - name: Macro F1
28
+ type: f1
29
+ value: 0.2949925820893562
30
+ - name: QWK
31
+ type: qwk
32
+ value: 0.4965296316070473
33
+ - name: Weighted Macro F1
34
+ type: f1
35
+ value: 0.5487262828076291
36
+ ---
37
+ # Model ID: bert-base-multilingual-cased-encoder_classification-C4-essay_only
38
+ ## Results
39
+ | | test_data |
40
+ |:-----------------|------------:|
41
+ | eval_accuracy | 0.536232 |
42
+ | eval_RMSE | 30.8338 |
43
+ | eval_QWK | 0.49653 |
44
+ | eval_Macro_F1 | 0.294993 |
45
+ | eval_Weighted_F1 | 0.548726 |
46
+ | eval_Micro_F1 | 0.536232 |
47
+ | eval_HDIV | 0.00724638 |
48
+
config.json ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "BertForSequenceClassification"
4
+ ],
5
+ "attention_probs_dropout_prob": 0.1,
6
+ "classifier_dropout": null,
7
+ "directionality": "bidi",
8
+ "hidden_act": "gelu",
9
+ "hidden_dropout_prob": 0.1,
10
+ "hidden_size": 768,
11
+ "id2label": {
12
+ "0": 0,
13
+ "1": 40,
14
+ "2": 80,
15
+ "3": 120,
16
+ "4": 160,
17
+ "5": 200
18
+ },
19
+ "initializer_range": 0.02,
20
+ "intermediate_size": 3072,
21
+ "label2id": {
22
+ "0": 0,
23
+ "40": 1,
24
+ "80": 2,
25
+ "120": 3,
26
+ "160": 4,
27
+ "200": 5
28
+ },
29
+ "layer_norm_eps": 1e-12,
30
+ "max_position_embeddings": 512,
31
+ "model_type": "bert",
32
+ "num_attention_heads": 12,
33
+ "num_hidden_layers": 12,
34
+ "pad_token_id": 0,
35
+ "pooler_fc_size": 768,
36
+ "pooler_num_attention_heads": 12,
37
+ "pooler_num_fc_layers": 3,
38
+ "pooler_size_per_head": 128,
39
+ "pooler_type": "first_token_transform",
40
+ "position_embedding_type": "absolute",
41
+ "torch_dtype": "float32",
42
+ "transformers_version": "4.53.1",
43
+ "type_vocab_size": 2,
44
+ "use_cache": true,
45
+ "vocab_size": 119547
46
+ }
emissions.csv ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ timestamp,project_name,run_id,experiment_id,duration,emissions,emissions_rate,cpu_power,gpu_power,ram_power,cpu_energy,gpu_energy,ram_energy,energy_consumed,country_name,country_iso_code,region,cloud_provider,cloud_region,os,python_version,codecarbon_version,cpu_count,cpu_model,gpu_count,gpu_model,longitude,latitude,ram_total_size,tracking_mode,on_cloud,pue
2
+ 2025-07-09T17:18:11,jbcs2025,f7043fb1-83d9-4f39-bc50-37d83ab1dd5a,bert-base-multilingual-cased-encoder_classification-C4-essay_only,92.52555080497405,0.0017463849133527085,1.8874623259836084e-05,41.0,109.55187753175237,58.0,0.0010732017221398305,0.004728838783067069,0.0014569300511196423,0.007258970556326542,Romania,ROU,gorj county,,,Linux-5.15.0-143-generic-x86_64-with-glibc2.35,3.12.11,3.0.2,36,Intel(R) Xeon(R) Gold 6248R CPU @ 3.00GHz,1,1 x NVIDIA RTX A6000,23.2904,45.0489,393.6063117980957,machine,N,1.0
evaluation_results.csv ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ eval_loss,eval_model_preparation_time,eval_accuracy,eval_RMSE,eval_QWK,eval_HDIV,eval_Macro_F1,eval_Micro_F1,eval_Weighted_F1,eval_TP_0,eval_TN_0,eval_FP_0,eval_FN_0,eval_TP_1,eval_TN_1,eval_FP_1,eval_FN_1,eval_TP_2,eval_TN_2,eval_FP_2,eval_FN_2,eval_TP_3,eval_TN_3,eval_FP_3,eval_FN_3,eval_TP_4,eval_TN_4,eval_FP_4,eval_FN_4,eval_TP_5,eval_TN_5,eval_FP_5,eval_FN_5,eval_runtime,eval_samples_per_second,eval_steps_per_second,epoch,reference,timestamp,id
2
+ 1.7804791927337646,0.0021,0.30303030303030304,42.78204097869841,0.1050646551724137,0.015151515151515138,0.144,0.30303030303030304,0.19515151515151513,0,131,0,1,0,132,0,0,0,128,0,4,0,68,0,64,34,16,68,14,6,93,24,9,0.6355,207.699,14.161,-1,validation_before_training,2025-07-09 17:16:48,bert-base-multilingual-cased-encoder_classification-C4-essay_only
3
+ 1.272457480430603,0.0021,0.5378787878787878,30.15113445777636,0.5589021564783461,0.007575757575757569,0.33328870666331967,0.5378787878787878,0.5431462985642552,0,131,0,1,0,132,0,0,0,122,6,4,34,52,16,30,27,57,27,21,10,105,12,5,0.3963,333.044,22.708,9.0,validation_after_training,2025-07-09 17:16:48,bert-base-multilingual-cased-encoder_classification-C4-essay_only
4
+ 1.3231467008590698,0.0021,0.5362318840579711,30.833822949500753,0.49652963160704733,0.007246376811594235,0.29499258208935625,0.5362318840579711,0.5487262828076291,0,137,0,1,0,137,0,1,5,112,17,4,44,42,20,32,23,71,21,23,2,127,6,3,0.3847,358.675,23.392,9.0,test_results,2025-07-09 17:16:48,bert-base-multilingual-cased-encoder_classification-C4-essay_only
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:41792349f09db4e47ded8c1f9f5e6a93eff3a7c4c194eef8179085a8117109de
3
+ size 711455760
run_experiment.log ADDED
@@ -0,0 +1,351 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2025-07-09 17:16:35,620][__main__][INFO] - cache_dir: /tmp/
2
+ dataset:
3
+ name: kamel-usp/aes_enem_dataset
4
+ split: JBCS2025
5
+ training_params:
6
+ seed: 42
7
+ num_train_epochs: 20
8
+ logging_steps: 100
9
+ metric_for_best_model: QWK
10
+ bf16: true
11
+ bootstrap:
12
+ enabled: true
13
+ n_bootstrap: 10000
14
+ bootstrap_seed: 42
15
+ metrics:
16
+ - QWK
17
+ - Macro_F1
18
+ - Weighted_F1
19
+ post_training_results:
20
+ model_path: /workspace/jbcs2025/outputs/2025-03-24/20-42-59
21
+ experiments:
22
+ model:
23
+ name: google-bert/bert-base-multilingual-cased
24
+ type: encoder_classification
25
+ num_labels: 6
26
+ output_dir: ./results/
27
+ logging_dir: ./logs/
28
+ best_model_dir: ./results/best_model
29
+ tokenizer:
30
+ name: google-bert/bert-base-multilingual-cased
31
+ dataset:
32
+ grade_index: 3
33
+ use_full_context: false
34
+ training_params:
35
+ weight_decay: 0.01
36
+ warmup_ratio: 0.1
37
+ learning_rate: 5.0e-05
38
+ train_batch_size: 16
39
+ eval_batch_size: 16
40
+ gradient_accumulation_steps: 1
41
+ gradient_checkpointing: false
42
+
43
+ [2025-07-09 17:16:39,482][__main__][INFO] - GPU 0: NVIDIA RTX A6000 | TDP ≈ 300 W
44
+ [2025-07-09 17:16:39,482][__main__][INFO] - Starting the Fine Tuning training process.
45
+ [2025-07-09 17:16:45,186][transformers.configuration_utils][INFO] - loading configuration file config.json from cache at /tmp/models--google-bert--bert-base-multilingual-cased/snapshots/3f076fdb1ab68d5b2880cb87a0886f315b8146f8/config.json
46
+ [2025-07-09 17:16:45,187][transformers.configuration_utils][INFO] - Model config BertConfig {
47
+ "architectures": [
48
+ "BertForMaskedLM"
49
+ ],
50
+ "attention_probs_dropout_prob": 0.1,
51
+ "classifier_dropout": null,
52
+ "directionality": "bidi",
53
+ "hidden_act": "gelu",
54
+ "hidden_dropout_prob": 0.1,
55
+ "hidden_size": 768,
56
+ "initializer_range": 0.02,
57
+ "intermediate_size": 3072,
58
+ "layer_norm_eps": 1e-12,
59
+ "max_position_embeddings": 512,
60
+ "model_type": "bert",
61
+ "num_attention_heads": 12,
62
+ "num_hidden_layers": 12,
63
+ "pad_token_id": 0,
64
+ "pooler_fc_size": 768,
65
+ "pooler_num_attention_heads": 12,
66
+ "pooler_num_fc_layers": 3,
67
+ "pooler_size_per_head": 128,
68
+ "pooler_type": "first_token_transform",
69
+ "position_embedding_type": "absolute",
70
+ "transformers_version": "4.53.1",
71
+ "type_vocab_size": 2,
72
+ "use_cache": true,
73
+ "vocab_size": 119547
74
+ }
75
+
76
+ [2025-07-09 17:16:45,391][transformers.configuration_utils][INFO] - loading configuration file config.json from cache at /tmp/models--google-bert--bert-base-multilingual-cased/snapshots/3f076fdb1ab68d5b2880cb87a0886f315b8146f8/config.json
77
+ [2025-07-09 17:16:45,392][transformers.configuration_utils][INFO] - Model config BertConfig {
78
+ "architectures": [
79
+ "BertForMaskedLM"
80
+ ],
81
+ "attention_probs_dropout_prob": 0.1,
82
+ "classifier_dropout": null,
83
+ "directionality": "bidi",
84
+ "hidden_act": "gelu",
85
+ "hidden_dropout_prob": 0.1,
86
+ "hidden_size": 768,
87
+ "initializer_range": 0.02,
88
+ "intermediate_size": 3072,
89
+ "layer_norm_eps": 1e-12,
90
+ "max_position_embeddings": 512,
91
+ "model_type": "bert",
92
+ "num_attention_heads": 12,
93
+ "num_hidden_layers": 12,
94
+ "pad_token_id": 0,
95
+ "pooler_fc_size": 768,
96
+ "pooler_num_attention_heads": 12,
97
+ "pooler_num_fc_layers": 3,
98
+ "pooler_size_per_head": 128,
99
+ "pooler_type": "first_token_transform",
100
+ "position_embedding_type": "absolute",
101
+ "transformers_version": "4.53.1",
102
+ "type_vocab_size": 2,
103
+ "use_cache": true,
104
+ "vocab_size": 119547
105
+ }
106
+
107
+ [2025-07-09 17:16:46,066][transformers.tokenization_utils_base][INFO] - loading file vocab.txt from cache at /tmp/models--google-bert--bert-base-multilingual-cased/snapshots/3f076fdb1ab68d5b2880cb87a0886f315b8146f8/vocab.txt
108
+ [2025-07-09 17:16:46,067][transformers.tokenization_utils_base][INFO] - loading file tokenizer.json from cache at /tmp/models--google-bert--bert-base-multilingual-cased/snapshots/3f076fdb1ab68d5b2880cb87a0886f315b8146f8/tokenizer.json
109
+ [2025-07-09 17:16:46,067][transformers.tokenization_utils_base][INFO] - loading file added_tokens.json from cache at None
110
+ [2025-07-09 17:16:46,067][transformers.tokenization_utils_base][INFO] - loading file special_tokens_map.json from cache at None
111
+ [2025-07-09 17:16:46,067][transformers.tokenization_utils_base][INFO] - loading file tokenizer_config.json from cache at /tmp/models--google-bert--bert-base-multilingual-cased/snapshots/3f076fdb1ab68d5b2880cb87a0886f315b8146f8/tokenizer_config.json
112
+ [2025-07-09 17:16:46,067][transformers.tokenization_utils_base][INFO] - loading file chat_template.jinja from cache at None
113
+ [2025-07-09 17:16:46,067][transformers.configuration_utils][INFO] - loading configuration file config.json from cache at /tmp/models--google-bert--bert-base-multilingual-cased/snapshots/3f076fdb1ab68d5b2880cb87a0886f315b8146f8/config.json
114
+ [2025-07-09 17:16:46,068][transformers.configuration_utils][INFO] - Model config BertConfig {
115
+ "architectures": [
116
+ "BertForMaskedLM"
117
+ ],
118
+ "attention_probs_dropout_prob": 0.1,
119
+ "classifier_dropout": null,
120
+ "directionality": "bidi",
121
+ "hidden_act": "gelu",
122
+ "hidden_dropout_prob": 0.1,
123
+ "hidden_size": 768,
124
+ "initializer_range": 0.02,
125
+ "intermediate_size": 3072,
126
+ "layer_norm_eps": 1e-12,
127
+ "max_position_embeddings": 512,
128
+ "model_type": "bert",
129
+ "num_attention_heads": 12,
130
+ "num_hidden_layers": 12,
131
+ "pad_token_id": 0,
132
+ "pooler_fc_size": 768,
133
+ "pooler_num_attention_heads": 12,
134
+ "pooler_num_fc_layers": 3,
135
+ "pooler_size_per_head": 128,
136
+ "pooler_type": "first_token_transform",
137
+ "position_embedding_type": "absolute",
138
+ "transformers_version": "4.53.1",
139
+ "type_vocab_size": 2,
140
+ "use_cache": true,
141
+ "vocab_size": 119547
142
+ }
143
+
144
+ [2025-07-09 17:16:46,311][__main__][INFO] - Tokenizer function parameters- Padding:longest; Truncation: True; Use Full Context: False
145
+ [2025-07-09 17:16:47,007][__main__][INFO] -
146
+ Token statistics for 'train' split:
147
+ [2025-07-09 17:16:47,008][__main__][INFO] - Total examples: 500
148
+ [2025-07-09 17:16:47,008][__main__][INFO] - Min tokens: 512
149
+ [2025-07-09 17:16:47,008][__main__][INFO] - Max tokens: 512
150
+ [2025-07-09 17:16:47,008][__main__][INFO] - Avg tokens: 512.00
151
+ [2025-07-09 17:16:47,008][__main__][INFO] - Std tokens: 0.00
152
+ [2025-07-09 17:16:47,103][__main__][INFO] -
153
+ Token statistics for 'validation' split:
154
+ [2025-07-09 17:16:47,103][__main__][INFO] - Total examples: 132
155
+ [2025-07-09 17:16:47,103][__main__][INFO] - Min tokens: 512
156
+ [2025-07-09 17:16:47,103][__main__][INFO] - Max tokens: 512
157
+ [2025-07-09 17:16:47,103][__main__][INFO] - Avg tokens: 512.00
158
+ [2025-07-09 17:16:47,103][__main__][INFO] - Std tokens: 0.00
159
+ [2025-07-09 17:16:47,203][__main__][INFO] -
160
+ Token statistics for 'test' split:
161
+ [2025-07-09 17:16:47,203][__main__][INFO] - Total examples: 138
162
+ [2025-07-09 17:16:47,203][__main__][INFO] - Min tokens: 512
163
+ [2025-07-09 17:16:47,203][__main__][INFO] - Max tokens: 512
164
+ [2025-07-09 17:16:47,203][__main__][INFO] - Avg tokens: 512.00
165
+ [2025-07-09 17:16:47,203][__main__][INFO] - Std tokens: 0.00
166
+ [2025-07-09 17:16:47,204][__main__][INFO] - If token statistics are the same (max, avg, min) keep in mind that this is due to batched tokenization and padding.
167
+ [2025-07-09 17:16:47,204][__main__][INFO] - Model max length: 512. If it is the same as stats, then there is a high chance that sequences are being truncated.
168
+ [2025-07-09 17:16:47,419][transformers.configuration_utils][INFO] - loading configuration file config.json from cache at /tmp/models--google-bert--bert-base-multilingual-cased/snapshots/3f076fdb1ab68d5b2880cb87a0886f315b8146f8/config.json
169
+ [2025-07-09 17:16:47,420][transformers.configuration_utils][INFO] - Model config BertConfig {
170
+ "architectures": [
171
+ "BertForMaskedLM"
172
+ ],
173
+ "attention_probs_dropout_prob": 0.1,
174
+ "classifier_dropout": null,
175
+ "directionality": "bidi",
176
+ "hidden_act": "gelu",
177
+ "hidden_dropout_prob": 0.1,
178
+ "hidden_size": 768,
179
+ "id2label": {
180
+ "0": 0,
181
+ "1": 40,
182
+ "2": 80,
183
+ "3": 120,
184
+ "4": 160,
185
+ "5": 200
186
+ },
187
+ "initializer_range": 0.02,
188
+ "intermediate_size": 3072,
189
+ "label2id": {
190
+ "0": 0,
191
+ "40": 1,
192
+ "80": 2,
193
+ "120": 3,
194
+ "160": 4,
195
+ "200": 5
196
+ },
197
+ "layer_norm_eps": 1e-12,
198
+ "max_position_embeddings": 512,
199
+ "model_type": "bert",
200
+ "num_attention_heads": 12,
201
+ "num_hidden_layers": 12,
202
+ "pad_token_id": 0,
203
+ "pooler_fc_size": 768,
204
+ "pooler_num_attention_heads": 12,
205
+ "pooler_num_fc_layers": 3,
206
+ "pooler_size_per_head": 128,
207
+ "pooler_type": "first_token_transform",
208
+ "position_embedding_type": "absolute",
209
+ "transformers_version": "4.53.1",
210
+ "type_vocab_size": 2,
211
+ "use_cache": true,
212
+ "vocab_size": 119547
213
+ }
214
+
215
+ [2025-07-09 17:16:47,598][transformers.modeling_utils][INFO] - loading weights file model.safetensors from cache at /tmp/models--google-bert--bert-base-multilingual-cased/snapshots/3f076fdb1ab68d5b2880cb87a0886f315b8146f8/model.safetensors
216
+ [2025-07-09 17:16:47,605][transformers.modeling_utils][INFO] - Since the `torch_dtype` attribute can't be found in model's config object, will use torch_dtype={torch_dtype} as derived from model's weights
217
+ [2025-07-09 17:16:47,605][transformers.modeling_utils][INFO] - Instantiating BertForSequenceClassification model under default dtype torch.float32.
218
+ [2025-07-09 17:16:47,773][transformers.modeling_utils][INFO] - A pretrained model of type `BertForSequenceClassification` contains parameters that have been renamed internally (a few are listed below but more are present in the model):
219
+ * `cls.predictions.transform.LayerNorm.beta` -> `cls.predictions.transform.LayerNorm.bias`
220
+ * `cls.predictions.transform.LayerNorm.gamma` -> `cls.predictions.transform.LayerNorm.weight`
221
+ If you are using a model from the Hub, consider submitting a PR to adjust these weights and help future users.
222
+ [2025-07-09 17:16:48,508][transformers.modeling_utils][INFO] - Some weights of the model checkpoint at google-bert/bert-base-multilingual-cased were not used when initializing BertForSequenceClassification: ['cls.predictions.bias', 'cls.predictions.transform.LayerNorm.bias', 'cls.predictions.transform.LayerNorm.weight', 'cls.predictions.transform.dense.bias', 'cls.predictions.transform.dense.weight', 'cls.seq_relationship.bias', 'cls.seq_relationship.weight']
223
+ - This IS expected if you are initializing BertForSequenceClassification from the checkpoint of a model trained on another task or with another architecture (e.g. initializing a BertForSequenceClassification model from a BertForPreTraining model).
224
+ - This IS NOT expected if you are initializing BertForSequenceClassification from the checkpoint of a model that you expect to be exactly identical (initializing a BertForSequenceClassification model from a BertForSequenceClassification model).
225
+ [2025-07-09 17:16:48,508][transformers.modeling_utils][WARNING] - Some weights of BertForSequenceClassification were not initialized from the model checkpoint at google-bert/bert-base-multilingual-cased and are newly initialized: ['classifier.bias', 'classifier.weight']
226
+ You should probably TRAIN this model on a down-stream task to be able to use it for predictions and inference.
227
+ [2025-07-09 17:16:48,514][transformers.training_args][INFO] - PyTorch: setting up devices
228
+ [2025-07-09 17:16:48,577][__main__][INFO] - Total steps: 620. Number of warmup steps: 62
229
+ [2025-07-09 17:16:48,585][transformers.trainer][INFO] - You have loaded a model on multiple GPUs. `is_model_parallel` attribute will be force-set to `True` to avoid any unexpected behavior such as device placement mismatching.
230
+ [2025-07-09 17:16:48,610][transformers.trainer][INFO] - Using auto half precision backend
231
+ [2025-07-09 17:16:48,612][transformers.trainer][INFO] - The following columns in the Evaluation set don't have a corresponding argument in `BertForSequenceClassification.forward` and have been ignored: essay_text, supporting_text, grades, id, id_prompt, essay_year, prompt, reference. If essay_text, supporting_text, grades, id, id_prompt, essay_year, prompt, reference are not expected by `BertForSequenceClassification.forward`, you can safely ignore this message.
232
+ [2025-07-09 17:16:48,618][transformers.trainer][INFO] -
233
+ ***** Running Evaluation *****
234
+ [2025-07-09 17:16:48,618][transformers.trainer][INFO] - Num examples = 132
235
+ [2025-07-09 17:16:48,618][transformers.trainer][INFO] - Batch size = 16
236
+ [2025-07-09 17:16:49,574][transformers.trainer][INFO] - The following columns in the Training set don't have a corresponding argument in `BertForSequenceClassification.forward` and have been ignored: essay_text, supporting_text, grades, id, id_prompt, essay_year, prompt, reference. If essay_text, supporting_text, grades, id, id_prompt, essay_year, prompt, reference are not expected by `BertForSequenceClassification.forward`, you can safely ignore this message.
237
+ [2025-07-09 17:16:49,583][transformers.trainer][INFO] - ***** Running training *****
238
+ [2025-07-09 17:16:49,584][transformers.trainer][INFO] - Num examples = 500
239
+ [2025-07-09 17:16:49,584][transformers.trainer][INFO] - Num Epochs = 20
240
+ [2025-07-09 17:16:49,584][transformers.trainer][INFO] - Instantaneous batch size per device = 16
241
+ [2025-07-09 17:16:49,584][transformers.trainer][INFO] - Total train batch size (w. parallel, distributed & accumulation) = 16
242
+ [2025-07-09 17:16:49,584][transformers.trainer][INFO] - Gradient Accumulation steps = 1
243
+ [2025-07-09 17:16:49,584][transformers.trainer][INFO] - Total optimization steps = 640
244
+ [2025-07-09 17:16:49,584][transformers.trainer][INFO] - Number of trainable parameters = 177,858,054
245
+ [2025-07-09 17:16:54,598][transformers.trainer][INFO] - The following columns in the Evaluation set don't have a corresponding argument in `BertForSequenceClassification.forward` and have been ignored: essay_text, supporting_text, grades, id, id_prompt, essay_year, prompt, reference. If essay_text, supporting_text, grades, id, id_prompt, essay_year, prompt, reference are not expected by `BertForSequenceClassification.forward`, you can safely ignore this message.
246
+ [2025-07-09 17:16:54,600][transformers.trainer][INFO] -
247
+ ***** Running Evaluation *****
248
+ [2025-07-09 17:16:54,600][transformers.trainer][INFO] - Num examples = 132
249
+ [2025-07-09 17:16:54,600][transformers.trainer][INFO] - Batch size = 16
250
+ [2025-07-09 17:16:54,990][transformers.trainer][INFO] - Saving model checkpoint to /workspace/jbcs2025/outputs/2025-07-09/17-16-35/results/checkpoint-32
251
+ [2025-07-09 17:16:54,991][transformers.configuration_utils][INFO] - Configuration saved in /workspace/jbcs2025/outputs/2025-07-09/17-16-35/results/checkpoint-32/config.json
252
+ [2025-07-09 17:16:56,420][transformers.modeling_utils][INFO] - Model weights saved in /workspace/jbcs2025/outputs/2025-07-09/17-16-35/results/checkpoint-32/model.safetensors
253
+ [2025-07-09 17:17:02,794][transformers.trainer][INFO] - The following columns in the Evaluation set don't have a corresponding argument in `BertForSequenceClassification.forward` and have been ignored: essay_text, supporting_text, grades, id, id_prompt, essay_year, prompt, reference. If essay_text, supporting_text, grades, id, id_prompt, essay_year, prompt, reference are not expected by `BertForSequenceClassification.forward`, you can safely ignore this message.
254
+ [2025-07-09 17:17:02,796][transformers.trainer][INFO] -
255
+ ***** Running Evaluation *****
256
+ [2025-07-09 17:17:02,796][transformers.trainer][INFO] - Num examples = 132
257
+ [2025-07-09 17:17:02,796][transformers.trainer][INFO] - Batch size = 16
258
+ [2025-07-09 17:17:03,180][transformers.trainer][INFO] - Saving model checkpoint to /workspace/jbcs2025/outputs/2025-07-09/17-16-35/results/checkpoint-64
259
+ [2025-07-09 17:17:03,181][transformers.configuration_utils][INFO] - Configuration saved in /workspace/jbcs2025/outputs/2025-07-09/17-16-35/results/checkpoint-64/config.json
260
+ [2025-07-09 17:17:04,499][transformers.modeling_utils][INFO] - Model weights saved in /workspace/jbcs2025/outputs/2025-07-09/17-16-35/results/checkpoint-64/model.safetensors
261
+ [2025-07-09 17:17:06,619][transformers.trainer][INFO] - Deleting older checkpoint [/workspace/jbcs2025/outputs/2025-07-09/17-16-35/results/checkpoint-32] due to args.save_total_limit
262
+ [2025-07-09 17:17:11,599][transformers.trainer][INFO] - The following columns in the Evaluation set don't have a corresponding argument in `BertForSequenceClassification.forward` and have been ignored: essay_text, supporting_text, grades, id, id_prompt, essay_year, prompt, reference. If essay_text, supporting_text, grades, id, id_prompt, essay_year, prompt, reference are not expected by `BertForSequenceClassification.forward`, you can safely ignore this message.
263
+ [2025-07-09 17:17:11,601][transformers.trainer][INFO] -
264
+ ***** Running Evaluation *****
265
+ [2025-07-09 17:17:11,601][transformers.trainer][INFO] - Num examples = 132
266
+ [2025-07-09 17:17:11,601][transformers.trainer][INFO] - Batch size = 16
267
+ [2025-07-09 17:17:11,990][transformers.trainer][INFO] - Saving model checkpoint to /workspace/jbcs2025/outputs/2025-07-09/17-16-35/results/checkpoint-96
268
+ [2025-07-09 17:17:11,991][transformers.configuration_utils][INFO] - Configuration saved in /workspace/jbcs2025/outputs/2025-07-09/17-16-35/results/checkpoint-96/config.json
269
+ [2025-07-09 17:17:13,340][transformers.modeling_utils][INFO] - Model weights saved in /workspace/jbcs2025/outputs/2025-07-09/17-16-35/results/checkpoint-96/model.safetensors
270
+ [2025-07-09 17:17:15,231][transformers.trainer][INFO] - Deleting older checkpoint [/workspace/jbcs2025/outputs/2025-07-09/17-16-35/results/checkpoint-64] due to args.save_total_limit
271
+ [2025-07-09 17:17:20,161][transformers.trainer][INFO] - The following columns in the Evaluation set don't have a corresponding argument in `BertForSequenceClassification.forward` and have been ignored: essay_text, supporting_text, grades, id, id_prompt, essay_year, prompt, reference. If essay_text, supporting_text, grades, id, id_prompt, essay_year, prompt, reference are not expected by `BertForSequenceClassification.forward`, you can safely ignore this message.
272
+ [2025-07-09 17:17:20,164][transformers.trainer][INFO] -
273
+ ***** Running Evaluation *****
274
+ [2025-07-09 17:17:20,164][transformers.trainer][INFO] - Num examples = 132
275
+ [2025-07-09 17:17:20,164][transformers.trainer][INFO] - Batch size = 16
276
+ [2025-07-09 17:17:20,551][transformers.trainer][INFO] - Saving model checkpoint to /workspace/jbcs2025/outputs/2025-07-09/17-16-35/results/checkpoint-128
277
+ [2025-07-09 17:17:20,552][transformers.configuration_utils][INFO] - Configuration saved in /workspace/jbcs2025/outputs/2025-07-09/17-16-35/results/checkpoint-128/config.json
278
+ [2025-07-09 17:17:22,038][transformers.modeling_utils][INFO] - Model weights saved in /workspace/jbcs2025/outputs/2025-07-09/17-16-35/results/checkpoint-128/model.safetensors
279
+ [2025-07-09 17:17:24,130][transformers.trainer][INFO] - Deleting older checkpoint [/workspace/jbcs2025/outputs/2025-07-09/17-16-35/results/checkpoint-96] due to args.save_total_limit
280
+ [2025-07-09 17:17:29,080][transformers.trainer][INFO] - The following columns in the Evaluation set don't have a corresponding argument in `BertForSequenceClassification.forward` and have been ignored: essay_text, supporting_text, grades, id, id_prompt, essay_year, prompt, reference. If essay_text, supporting_text, grades, id, id_prompt, essay_year, prompt, reference are not expected by `BertForSequenceClassification.forward`, you can safely ignore this message.
281
+ [2025-07-09 17:17:29,082][transformers.trainer][INFO] -
282
+ ***** Running Evaluation *****
283
+ [2025-07-09 17:17:29,083][transformers.trainer][INFO] - Num examples = 132
284
+ [2025-07-09 17:17:29,083][transformers.trainer][INFO] - Batch size = 16
285
+ [2025-07-09 17:17:29,474][transformers.trainer][INFO] - Saving model checkpoint to /workspace/jbcs2025/outputs/2025-07-09/17-16-35/results/checkpoint-160
286
+ [2025-07-09 17:17:29,476][transformers.configuration_utils][INFO] - Configuration saved in /workspace/jbcs2025/outputs/2025-07-09/17-16-35/results/checkpoint-160/config.json
287
+ [2025-07-09 17:17:30,903][transformers.modeling_utils][INFO] - Model weights saved in /workspace/jbcs2025/outputs/2025-07-09/17-16-35/results/checkpoint-160/model.safetensors
288
+ [2025-07-09 17:17:37,833][transformers.trainer][INFO] - The following columns in the Evaluation set don't have a corresponding argument in `BertForSequenceClassification.forward` and have been ignored: essay_text, supporting_text, grades, id, id_prompt, essay_year, prompt, reference. If essay_text, supporting_text, grades, id, id_prompt, essay_year, prompt, reference are not expected by `BertForSequenceClassification.forward`, you can safely ignore this message.
289
+ [2025-07-09 17:17:37,835][transformers.trainer][INFO] -
290
+ ***** Running Evaluation *****
291
+ [2025-07-09 17:17:37,835][transformers.trainer][INFO] - Num examples = 132
292
+ [2025-07-09 17:17:37,835][transformers.trainer][INFO] - Batch size = 16
293
+ [2025-07-09 17:17:38,223][transformers.trainer][INFO] - Saving model checkpoint to /workspace/jbcs2025/outputs/2025-07-09/17-16-35/results/checkpoint-192
294
+ [2025-07-09 17:17:38,224][transformers.configuration_utils][INFO] - Configuration saved in /workspace/jbcs2025/outputs/2025-07-09/17-16-35/results/checkpoint-192/config.json
295
+ [2025-07-09 17:17:39,538][transformers.modeling_utils][INFO] - Model weights saved in /workspace/jbcs2025/outputs/2025-07-09/17-16-35/results/checkpoint-192/model.safetensors
296
+ [2025-07-09 17:17:41,709][transformers.trainer][INFO] - Deleting older checkpoint [/workspace/jbcs2025/outputs/2025-07-09/17-16-35/results/checkpoint-160] due to args.save_total_limit
297
+ [2025-07-09 17:17:46,670][transformers.trainer][INFO] - The following columns in the Evaluation set don't have a corresponding argument in `BertForSequenceClassification.forward` and have been ignored: essay_text, supporting_text, grades, id, id_prompt, essay_year, prompt, reference. If essay_text, supporting_text, grades, id, id_prompt, essay_year, prompt, reference are not expected by `BertForSequenceClassification.forward`, you can safely ignore this message.
298
+ [2025-07-09 17:17:46,672][transformers.trainer][INFO] -
299
+ ***** Running Evaluation *****
300
+ [2025-07-09 17:17:46,673][transformers.trainer][INFO] - Num examples = 132
301
+ [2025-07-09 17:17:46,673][transformers.trainer][INFO] - Batch size = 16
302
+ [2025-07-09 17:17:47,064][transformers.trainer][INFO] - Saving model checkpoint to /workspace/jbcs2025/outputs/2025-07-09/17-16-35/results/checkpoint-224
303
+ [2025-07-09 17:17:47,066][transformers.configuration_utils][INFO] - Configuration saved in /workspace/jbcs2025/outputs/2025-07-09/17-16-35/results/checkpoint-224/config.json
304
+ [2025-07-09 17:17:48,459][transformers.modeling_utils][INFO] - Model weights saved in /workspace/jbcs2025/outputs/2025-07-09/17-16-35/results/checkpoint-224/model.safetensors
305
+ [2025-07-09 17:17:50,566][transformers.trainer][INFO] - Deleting older checkpoint [/workspace/jbcs2025/outputs/2025-07-09/17-16-35/results/checkpoint-192] due to args.save_total_limit
306
+ [2025-07-09 17:17:55,524][transformers.trainer][INFO] - The following columns in the Evaluation set don't have a corresponding argument in `BertForSequenceClassification.forward` and have been ignored: essay_text, supporting_text, grades, id, id_prompt, essay_year, prompt, reference. If essay_text, supporting_text, grades, id, id_prompt, essay_year, prompt, reference are not expected by `BertForSequenceClassification.forward`, you can safely ignore this message.
307
+ [2025-07-09 17:17:55,527][transformers.trainer][INFO] -
308
+ ***** Running Evaluation *****
309
+ [2025-07-09 17:17:55,527][transformers.trainer][INFO] - Num examples = 132
310
+ [2025-07-09 17:17:55,527][transformers.trainer][INFO] - Batch size = 16
311
+ [2025-07-09 17:17:55,918][transformers.trainer][INFO] - Saving model checkpoint to /workspace/jbcs2025/outputs/2025-07-09/17-16-35/results/checkpoint-256
312
+ [2025-07-09 17:17:55,920][transformers.configuration_utils][INFO] - Configuration saved in /workspace/jbcs2025/outputs/2025-07-09/17-16-35/results/checkpoint-256/config.json
313
+ [2025-07-09 17:17:57,309][transformers.modeling_utils][INFO] - Model weights saved in /workspace/jbcs2025/outputs/2025-07-09/17-16-35/results/checkpoint-256/model.safetensors
314
+ [2025-07-09 17:17:59,425][transformers.trainer][INFO] - Deleting older checkpoint [/workspace/jbcs2025/outputs/2025-07-09/17-16-35/results/checkpoint-224] due to args.save_total_limit
315
+ [2025-07-09 17:18:04,384][transformers.trainer][INFO] - The following columns in the Evaluation set don't have a corresponding argument in `BertForSequenceClassification.forward` and have been ignored: essay_text, supporting_text, grades, id, id_prompt, essay_year, prompt, reference. If essay_text, supporting_text, grades, id, id_prompt, essay_year, prompt, reference are not expected by `BertForSequenceClassification.forward`, you can safely ignore this message.
316
+ [2025-07-09 17:18:04,387][transformers.trainer][INFO] -
317
+ ***** Running Evaluation *****
318
+ [2025-07-09 17:18:04,387][transformers.trainer][INFO] - Num examples = 132
319
+ [2025-07-09 17:18:04,387][transformers.trainer][INFO] - Batch size = 16
320
+ [2025-07-09 17:18:04,777][transformers.trainer][INFO] - Saving model checkpoint to /workspace/jbcs2025/outputs/2025-07-09/17-16-35/results/checkpoint-288
321
+ [2025-07-09 17:18:04,778][transformers.configuration_utils][INFO] - Configuration saved in /workspace/jbcs2025/outputs/2025-07-09/17-16-35/results/checkpoint-288/config.json
322
+ [2025-07-09 17:18:06,261][transformers.modeling_utils][INFO] - Model weights saved in /workspace/jbcs2025/outputs/2025-07-09/17-16-35/results/checkpoint-288/model.safetensors
323
+ [2025-07-09 17:18:08,369][transformers.trainer][INFO] - Deleting older checkpoint [/workspace/jbcs2025/outputs/2025-07-09/17-16-35/results/checkpoint-256] due to args.save_total_limit
324
+ [2025-07-09 17:18:08,507][transformers.trainer][INFO] -
325
+
326
+ Training completed. Do not forget to share your model on huggingface.co/models =)
327
+
328
+
329
+ [2025-07-09 17:18:08,507][transformers.trainer][INFO] - Loading best model from /workspace/jbcs2025/outputs/2025-07-09/17-16-35/results/checkpoint-128 (score: 0.5589021564783461).
330
+ [2025-07-09 17:18:08,871][transformers.trainer][INFO] - Deleting older checkpoint [/workspace/jbcs2025/outputs/2025-07-09/17-16-35/results/checkpoint-288] due to args.save_total_limit
331
+ [2025-07-09 17:18:09,010][transformers.trainer][INFO] - The following columns in the Evaluation set don't have a corresponding argument in `BertForSequenceClassification.forward` and have been ignored: essay_text, supporting_text, grades, id, id_prompt, essay_year, prompt, reference. If essay_text, supporting_text, grades, id, id_prompt, essay_year, prompt, reference are not expected by `BertForSequenceClassification.forward`, you can safely ignore this message.
332
+ [2025-07-09 17:18:09,013][transformers.trainer][INFO] -
333
+ ***** Running Evaluation *****
334
+ [2025-07-09 17:18:09,013][transformers.trainer][INFO] - Num examples = 132
335
+ [2025-07-09 17:18:09,013][transformers.trainer][INFO] - Batch size = 16
336
+ [2025-07-09 17:18:09,416][__main__][INFO] - Training completed successfully.
337
+ [2025-07-09 17:18:09,416][__main__][INFO] - Running on Test
338
+ [2025-07-09 17:18:09,417][transformers.trainer][INFO] - The following columns in the Evaluation set don't have a corresponding argument in `BertForSequenceClassification.forward` and have been ignored: essay_text, supporting_text, grades, id, id_prompt, essay_year, prompt, reference. If essay_text, supporting_text, grades, id, id_prompt, essay_year, prompt, reference are not expected by `BertForSequenceClassification.forward`, you can safely ignore this message.
339
+ [2025-07-09 17:18:09,419][transformers.trainer][INFO] -
340
+ ***** Running Evaluation *****
341
+ [2025-07-09 17:18:09,419][transformers.trainer][INFO] - Num examples = 138
342
+ [2025-07-09 17:18:09,419][transformers.trainer][INFO] - Batch size = 16
343
+ [2025-07-09 17:18:09,810][__main__][INFO] - Test metrics: {'eval_loss': 1.3231467008590698, 'eval_model_preparation_time': 0.0021, 'eval_accuracy': 0.5362318840579711, 'eval_RMSE': 30.833822949500753, 'eval_QWK': 0.49652963160704733, 'eval_HDIV': 0.007246376811594235, 'eval_Macro_F1': 0.29499258208935625, 'eval_Micro_F1': 0.5362318840579711, 'eval_Weighted_F1': 0.5487262828076291, 'eval_TP_0': 0, 'eval_TN_0': 137, 'eval_FP_0': 0, 'eval_FN_0': 1, 'eval_TP_1': 0, 'eval_TN_1': 137, 'eval_FP_1': 0, 'eval_FN_1': 1, 'eval_TP_2': 5, 'eval_TN_2': 112, 'eval_FP_2': 17, 'eval_FN_2': 4, 'eval_TP_3': 44, 'eval_TN_3': 42, 'eval_FP_3': 20, 'eval_FN_3': 32, 'eval_TP_4': 23, 'eval_TN_4': 71, 'eval_FP_4': 21, 'eval_FN_4': 23, 'eval_TP_5': 2, 'eval_TN_5': 127, 'eval_FP_5': 6, 'eval_FN_5': 3, 'eval_runtime': 0.3847, 'eval_samples_per_second': 358.675, 'eval_steps_per_second': 23.392, 'epoch': 9.0}
344
+ [2025-07-09 17:18:09,810][transformers.trainer][INFO] - Saving model checkpoint to ./results/best_model
345
+ [2025-07-09 17:18:09,812][transformers.configuration_utils][INFO] - Configuration saved in ./results/best_model/config.json
346
+ [2025-07-09 17:18:10,908][transformers.modeling_utils][INFO] - Model weights saved in ./results/best_model/model.safetensors
347
+ [2025-07-09 17:18:10,910][transformers.tokenization_utils_base][INFO] - tokenizer config file saved in ./results/best_model/tokenizer_config.json
348
+ [2025-07-09 17:18:10,910][transformers.tokenization_utils_base][INFO] - Special tokens file saved in ./results/best_model/special_tokens_map.json
349
+ [2025-07-09 17:18:10,975][__main__][INFO] - Model and tokenizer saved to ./results/best_model
350
+ [2025-07-09 17:18:10,997][__main__][INFO] - Fine Tuning Finished.
351
+ [2025-07-09 17:18:11,507][__main__][INFO] - Total emissions: 0.0017 kg CO2eq
special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,56 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "100": {
12
+ "content": "[UNK]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "101": {
20
+ "content": "[CLS]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "102": {
28
+ "content": "[SEP]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "103": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "clean_up_tokenization_spaces": false,
45
+ "cls_token": "[CLS]",
46
+ "do_lower_case": false,
47
+ "extra_special_tokens": {},
48
+ "mask_token": "[MASK]",
49
+ "model_max_length": 512,
50
+ "pad_token": "[PAD]",
51
+ "sep_token": "[SEP]",
52
+ "strip_accents": null,
53
+ "tokenize_chinese_chars": true,
54
+ "tokenizer_class": "BertTokenizer",
55
+ "unk_token": "[UNK]"
56
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4e8ca4052f2bbb8925daf9b957acae4217086431184dd4d46c4b83ad1d2370df
3
+ size 5777
vocab.txt ADDED
The diff for this file is too large to render. See raw diff