wlg1 commited on
Commit
3f923ae
·
verified ·
1 Parent(s): 56475e3

Upload folder using huggingface_hub

Browse files
checkpoint-1800/config.json ADDED
@@ -0,0 +1,41 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "RobertaForSequenceClassification"
4
+ ],
5
+ "attention_probs_dropout_prob": 0.1,
6
+ "bos_token_id": 0,
7
+ "classifier_dropout": null,
8
+ "eos_token_id": 2,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "id2label": {
13
+ "0": "LABEL_0",
14
+ "1": "LABEL_1",
15
+ "2": "LABEL_2",
16
+ "3": "LABEL_3",
17
+ "4": "LABEL_4"
18
+ },
19
+ "initializer_range": 0.02,
20
+ "intermediate_size": 3072,
21
+ "label2id": {
22
+ "LABEL_0": 0,
23
+ "LABEL_1": 1,
24
+ "LABEL_2": 2,
25
+ "LABEL_3": 3,
26
+ "LABEL_4": 4
27
+ },
28
+ "layer_norm_eps": 1e-05,
29
+ "max_position_embeddings": 514,
30
+ "model_type": "roberta",
31
+ "num_attention_heads": 12,
32
+ "num_hidden_layers": 12,
33
+ "pad_token_id": 1,
34
+ "position_embedding_type": "absolute",
35
+ "problem_type": "multi_label_classification",
36
+ "torch_dtype": "bfloat16",
37
+ "transformers_version": "4.55.0",
38
+ "type_vocab_size": 1,
39
+ "use_cache": false,
40
+ "vocab_size": 50265
41
+ }
checkpoint-1800/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:76afa0ec590f71d749b95f9bf8b95ac09ab2697d6d7478ec74aa4512713c6193
3
+ size 249323242
checkpoint-1800/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:456ee979e1a2998ca15af1bcd9746414c0cd602292d5fd57f6bbadb564c74973
3
+ size 14645
checkpoint-1800/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3b3aa12c9ba0848bd3e7ec5dc259cdc0335978c8dc0bcac6cebdd7bf38be3cdd
3
+ size 5713
checkpoint-200/config.json ADDED
@@ -0,0 +1,41 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "RobertaForSequenceClassification"
4
+ ],
5
+ "attention_probs_dropout_prob": 0.1,
6
+ "bos_token_id": 0,
7
+ "classifier_dropout": null,
8
+ "eos_token_id": 2,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "id2label": {
13
+ "0": "LABEL_0",
14
+ "1": "LABEL_1",
15
+ "2": "LABEL_2",
16
+ "3": "LABEL_3",
17
+ "4": "LABEL_4"
18
+ },
19
+ "initializer_range": 0.02,
20
+ "intermediate_size": 3072,
21
+ "label2id": {
22
+ "LABEL_0": 0,
23
+ "LABEL_1": 1,
24
+ "LABEL_2": 2,
25
+ "LABEL_3": 3,
26
+ "LABEL_4": 4
27
+ },
28
+ "layer_norm_eps": 1e-05,
29
+ "max_position_embeddings": 514,
30
+ "model_type": "roberta",
31
+ "num_attention_heads": 12,
32
+ "num_hidden_layers": 12,
33
+ "pad_token_id": 1,
34
+ "position_embedding_type": "absolute",
35
+ "problem_type": "multi_label_classification",
36
+ "torch_dtype": "bfloat16",
37
+ "transformers_version": "4.55.0",
38
+ "type_vocab_size": 1,
39
+ "use_cache": false,
40
+ "vocab_size": 50265
41
+ }
checkpoint-200/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3b8ec2df11b29defb9e3fe792dd1082ac14c8ba611d4538056cc25ec54387355
3
+ size 249323242
checkpoint-200/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b39b6009a03e78040640051be9dbef3157dcad15f2bb48f9fb2ccc5804cd6cfb
3
+ size 14645
checkpoint-200/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3b3aa12c9ba0848bd3e7ec5dc259cdc0335978c8dc0bcac6cebdd7bf38be3cdd
3
+ size 5713
choice_distribution.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "gpt-4o": 0.3157894736842105,
3
+ "DeepSeek-V3-0324": 0.09649122807017543,
4
+ "Llama-4-maverick-17b-128e-instruct-fp8": 0.03508771929824561,
5
+ "qwen25-coder-32b-instruct": 0.16666666666666666,
6
+ "gpt-4.1-mini": 0.38596491228070173
7
+ }
config.json ADDED
@@ -0,0 +1,41 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "RobertaForSequenceClassification"
4
+ ],
5
+ "attention_probs_dropout_prob": 0.1,
6
+ "bos_token_id": 0,
7
+ "classifier_dropout": null,
8
+ "eos_token_id": 2,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "id2label": {
13
+ "0": "LABEL_0",
14
+ "1": "LABEL_1",
15
+ "2": "LABEL_2",
16
+ "3": "LABEL_3",
17
+ "4": "LABEL_4"
18
+ },
19
+ "initializer_range": 0.02,
20
+ "intermediate_size": 3072,
21
+ "label2id": {
22
+ "LABEL_0": 0,
23
+ "LABEL_1": 1,
24
+ "LABEL_2": 2,
25
+ "LABEL_3": 3,
26
+ "LABEL_4": 4
27
+ },
28
+ "layer_norm_eps": 1e-05,
29
+ "max_position_embeddings": 514,
30
+ "model_type": "roberta",
31
+ "num_attention_heads": 12,
32
+ "num_hidden_layers": 12,
33
+ "pad_token_id": 1,
34
+ "position_embedding_type": "absolute",
35
+ "problem_type": "multi_label_classification",
36
+ "torch_dtype": "bfloat16",
37
+ "transformers_version": "4.55.0",
38
+ "type_vocab_size": 1,
39
+ "use_cache": false,
40
+ "vocab_size": 50265
41
+ }
logs/events.out.tfevents.1754589352.209-20-159-47.49367.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bb6f7de182cda177ab62f8feecf86f3f9dd94aa6559697e79547ac452b159f9f
3
+ size 14121
logs/events.out.tfevents.1754589352.209-20-159-47.49367.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:813510f8bf8c74eda1323bd387bc48f1acd33c1b06b92013dcc825fda204aece
3
+ size 14121
logs/events.out.tfevents.1754590693.209-20-159-47.50135.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1975b348a714c2c7d43fc3521893f9bed29fa51e6372f9464c58f4feb69f6854
3
+ size 16692
logs/events.out.tfevents.1754590693.209-20-159-47.50135.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:04f9b38305e6d50360c29b10d71dba0d6ad65fa481db91c6fb16fa026268d4f3
3
+ size 16692
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3b8ec2df11b29defb9e3fe792dd1082ac14c8ba611d4538056cc25ec54387355
3
+ size 249323242
per_sample_predictions.csv ADDED
@@ -0,0 +1,119 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ logit_gpt-4o,logit_DeepSeek-V3-0324,logit_Llama-4-maverick-17b-128e-instruct-fp8,logit_qwen25-coder-32b-instruct,logit_gpt-4.1-mini,chosen_executor,correct_prediction,true_gpt-4o,true_DeepSeek-V3-0324,true_Llama-4-maverick-17b-128e-instruct-fp8,true_qwen25-coder-32b-instruct,true_gpt-4.1-mini
2
+ 0.032958984375,0.06298828125,-0.05224609375,0.083984375,0.1201171875,gpt-4.1-mini,1,1.0,1.0,1.0,1.0,1.0
3
+ 0.0712890625,0.046875,-0.0264892578125,0.1416015625,-0.01190185546875,qwen25-coder-32b-instruct,1,1.0,1.0,0.0,1.0,1.0
4
+ 0.1611328125,0.1923828125,0.18359375,0.25390625,0.3671875,gpt-4.1-mini,1,1.0,1.0,1.0,1.0,1.0
5
+ -0.0732421875,-0.12109375,-0.193359375,-0.10498046875,-0.1259765625,gpt-4o,0,0.0,0.0,0.0,0.0,0.0
6
+ -0.2578125,-0.17578125,-0.2265625,-0.10986328125,-0.2158203125,qwen25-coder-32b-instruct,0,0.0,0.0,0.0,0.0,0.0
7
+ -0.267578125,-0.2177734375,-0.310546875,-0.337890625,-0.310546875,DeepSeek-V3-0324,1,0.0,1.0,0.0,0.0,0.0
8
+ 0.11376953125,0.0272216796875,-0.053466796875,0.058837890625,0.08203125,gpt-4o,1,1.0,1.0,1.0,1.0,1.0
9
+ -0.267578125,-0.380859375,-0.4140625,-0.3515625,-0.59765625,gpt-4o,0,0.0,0.0,0.0,0.0,0.0
10
+ -0.2041015625,-0.33203125,-0.39453125,-0.328125,-0.349609375,gpt-4o,1,1.0,1.0,0.0,1.0,1.0
11
+ -0.045166015625,-0.0791015625,-0.07275390625,-0.0091552734375,0.0634765625,gpt-4.1-mini,1,1.0,1.0,1.0,1.0,1.0
12
+ -0.11328125,-0.1103515625,-0.162109375,-0.1455078125,-0.1787109375,DeepSeek-V3-0324,0,0.0,0.0,0.0,0.0,0.0
13
+ 0.043212890625,0.08544921875,0.09033203125,0.1025390625,0.1572265625,gpt-4.1-mini,0,0.0,0.0,0.0,0.0,0.0
14
+ -0.0380859375,-0.0888671875,-0.24609375,-0.12158203125,-0.047607421875,gpt-4o,0,0.0,0.0,0.0,0.0,0.0
15
+ -0.17578125,-0.251953125,-0.259765625,-0.2021484375,-0.2119140625,gpt-4o,1,1.0,0.0,0.0,1.0,0.0
16
+ 0.040771484375,0.0654296875,-0.051513671875,-0.0205078125,0.16796875,gpt-4.1-mini,0,0.0,0.0,0.0,0.0,0.0
17
+ -0.00095367431640625,-0.203125,-0.1884765625,-0.154296875,-0.158203125,gpt-4o,0,0.0,0.0,0.0,0.0,0.0
18
+ -0.318359375,-0.423828125,-0.408203125,-0.38671875,-0.61328125,gpt-4o,0,0.0,0.0,0.0,0.0,0.0
19
+ -0.051025390625,-0.05615234375,-0.2109375,0.0169677734375,0.005889892578125,qwen25-coder-32b-instruct,0,0.0,1.0,0.0,0.0,0.0
20
+ -0.0191650390625,-0.0693359375,-0.140625,-0.103515625,-0.004852294921875,gpt-4.1-mini,0,1.0,1.0,1.0,1.0,0.0
21
+ -0.11962890625,-0.18359375,-0.2236328125,-0.12060546875,-0.12060546875,gpt-4o,1,1.0,1.0,1.0,1.0,1.0
22
+ 0.1943359375,0.1572265625,0.1640625,0.205078125,0.31640625,gpt-4.1-mini,0,0.0,0.0,0.0,0.0,0.0
23
+ -0.228515625,-0.1474609375,-0.2890625,-0.30078125,-0.302734375,DeepSeek-V3-0324,0,0.0,0.0,0.0,0.0,0.0
24
+ -0.1943359375,-0.38671875,-0.388671875,-0.1953125,-0.267578125,gpt-4o,0,0.0,0.0,0.0,0.0,0.0
25
+ -0.423828125,-0.251953125,-0.31640625,-0.2890625,-0.458984375,DeepSeek-V3-0324,0,0.0,0.0,0.0,0.0,0.0
26
+ -0.2421875,-0.2578125,-0.326171875,-0.228515625,-0.4140625,qwen25-coder-32b-instruct,0,0.0,0.0,1.0,0.0,0.0
27
+ 0.07568359375,0.1005859375,0.0264892578125,0.234375,0.2734375,gpt-4.1-mini,1,1.0,1.0,0.0,1.0,1.0
28
+ -0.2294921875,-0.474609375,-0.296875,-0.36328125,-0.6015625,gpt-4o,0,0.0,0.0,0.0,0.0,0.0
29
+ -0.057861328125,-0.083984375,-0.1435546875,-0.0189208984375,-0.0033111572265625,gpt-4.1-mini,1,1.0,1.0,1.0,1.0,1.0
30
+ 0.296875,0.1376953125,0.09716796875,0.275390625,0.427734375,gpt-4.1-mini,1,1.0,1.0,1.0,1.0,1.0
31
+ 0.00335693359375,-0.0673828125,-0.1103515625,-0.0595703125,0.0247802734375,gpt-4.1-mini,1,0.0,1.0,1.0,1.0,1.0
32
+ -0.169921875,-0.28515625,-0.400390625,-0.1533203125,-0.3828125,qwen25-coder-32b-instruct,0,0.0,0.0,0.0,0.0,0.0
33
+ 0.00665283203125,-0.07470703125,-0.1162109375,-0.150390625,0.0306396484375,gpt-4.1-mini,1,1.0,1.0,1.0,1.0,1.0
34
+ -0.283203125,-0.3046875,-0.431640625,-0.21484375,-0.44140625,qwen25-coder-32b-instruct,0,0.0,0.0,0.0,0.0,0.0
35
+ 0.154296875,0.109375,0.00323486328125,0.177734375,0.279296875,gpt-4.1-mini,0,0.0,0.0,0.0,0.0,0.0
36
+ -0.095703125,-0.236328125,-0.28125,-0.0908203125,-0.17578125,qwen25-coder-32b-instruct,1,1.0,1.0,1.0,1.0,1.0
37
+ -0.04150390625,-0.0908203125,-0.154296875,-0.07373046875,-0.030029296875,gpt-4.1-mini,1,1.0,1.0,1.0,1.0,1.0
38
+ -0.35546875,-0.326171875,-0.37890625,-0.328125,-0.46875,DeepSeek-V3-0324,0,0.0,0.0,0.0,0.0,0.0
39
+ 0.004913330078125,-0.00799560546875,-0.07666015625,0.000759124755859375,0.134765625,gpt-4.1-mini,0,0.0,0.0,0.0,0.0,0.0
40
+ -0.318359375,-0.296875,-0.244140625,-0.34765625,-0.439453125,Llama-4-maverick-17b-128e-instruct-fp8,1,0.0,1.0,1.0,0.0,1.0
41
+ -0.044189453125,-0.08203125,-0.09130859375,-0.03271484375,-0.0771484375,qwen25-coder-32b-instruct,1,1.0,0.0,0.0,1.0,0.0
42
+ -0.16015625,-0.37890625,-0.421875,-0.1630859375,-0.22265625,gpt-4o,0,0.0,1.0,1.0,1.0,1.0
43
+ -0.197265625,-0.224609375,-0.1767578125,-0.22265625,-0.3203125,Llama-4-maverick-17b-128e-instruct-fp8,0,0.0,0.0,0.0,0.0,1.0
44
+ 0.1474609375,0.0439453125,-0.05908203125,0.158203125,0.12451171875,qwen25-coder-32b-instruct,0,0.0,0.0,0.0,0.0,0.0
45
+ 0.28515625,0.31640625,0.15234375,0.39453125,0.578125,gpt-4.1-mini,1,1.0,1.0,1.0,1.0,1.0
46
+ -0.032958984375,-0.054931640625,-0.053955078125,-0.07861328125,0.04541015625,gpt-4.1-mini,1,1.0,1.0,0.0,1.0,1.0
47
+ 0.1396484375,0.1474609375,0.0230712890625,0.259765625,0.220703125,qwen25-coder-32b-instruct,1,0.0,0.0,0.0,1.0,0.0
48
+ 0.007659912109375,-0.07177734375,-0.08056640625,-0.0162353515625,-0.0859375,gpt-4o,0,0.0,0.0,0.0,0.0,0.0
49
+ -0.1708984375,-0.12890625,-0.12158203125,-0.2001953125,-0.2216796875,Llama-4-maverick-17b-128e-instruct-fp8,1,1.0,1.0,1.0,1.0,1.0
50
+ -0.1435546875,-0.41015625,-0.330078125,-0.228515625,-0.48828125,gpt-4o,1,1.0,0.0,0.0,1.0,0.0
51
+ 0.06689453125,0.0101318359375,0.037841796875,0.0303955078125,0.1279296875,gpt-4.1-mini,1,1.0,1.0,1.0,1.0,1.0
52
+ -0.06640625,-0.09033203125,-0.2001953125,-0.10546875,-0.07470703125,gpt-4o,0,0.0,1.0,1.0,0.0,1.0
53
+ -0.2412109375,-0.2734375,-0.18359375,-0.28125,-0.388671875,Llama-4-maverick-17b-128e-instruct-fp8,1,0.0,0.0,1.0,0.0,1.0
54
+ 0.296875,0.337890625,0.2734375,0.39453125,0.5234375,gpt-4.1-mini,0,0.0,0.0,0.0,0.0,0.0
55
+ 0.1845703125,0.1240234375,0.0189208984375,0.27734375,0.2451171875,qwen25-coder-32b-instruct,1,1.0,1.0,1.0,1.0,1.0
56
+ -0.0908203125,-0.064453125,-0.2041015625,-0.203125,-0.1689453125,DeepSeek-V3-0324,1,1.0,1.0,1.0,1.0,1.0
57
+ -0.12353515625,-0.314453125,-0.40625,-0.25390625,-0.416015625,gpt-4o,1,1.0,0.0,1.0,0.0,1.0
58
+ -0.1884765625,-0.39453125,-0.39453125,-0.1904296875,-0.2421875,gpt-4o,1,1.0,1.0,0.0,1.0,0.0
59
+ 0.1669921875,0.2158203125,0.15625,0.251953125,0.333984375,gpt-4.1-mini,0,0.0,0.0,0.0,0.0,0.0
60
+ -0.00946044921875,-0.052001953125,-0.0302734375,0.01531982421875,0.036865234375,gpt-4.1-mini,1,1.0,1.0,1.0,0.0,1.0
61
+ -0.1396484375,-0.296875,-0.3515625,-0.361328125,-0.466796875,gpt-4o,0,0.0,0.0,0.0,0.0,0.0
62
+ 0.06396484375,0.1279296875,0.0654296875,0.04931640625,0.2431640625,gpt-4.1-mini,0,0.0,0.0,0.0,0.0,0.0
63
+ 0.2255859375,0.1767578125,0.1884765625,0.259765625,0.392578125,gpt-4.1-mini,1,1.0,1.0,1.0,1.0,1.0
64
+ -0.2578125,-0.314453125,-0.37109375,-0.341796875,-0.478515625,gpt-4o,0,0.0,0.0,0.0,0.0,0.0
65
+ -0.134765625,-0.1484375,-0.3046875,-0.193359375,-0.2197265625,gpt-4o,0,0.0,0.0,0.0,0.0,0.0
66
+ 0.00927734375,-0.0654296875,-0.15625,-0.01373291015625,-0.015869140625,gpt-4o,1,1.0,1.0,1.0,1.0,1.0
67
+ 0.002655029296875,-0.0849609375,-0.0361328125,-0.0233154296875,-0.022216796875,gpt-4o,1,1.0,1.0,1.0,1.0,1.0
68
+ -0.16796875,-0.2080078125,-0.232421875,-0.0279541015625,-0.1708984375,qwen25-coder-32b-instruct,1,0.0,0.0,0.0,1.0,0.0
69
+ 0.0537109375,0.04931640625,0.025390625,0.10498046875,0.1552734375,gpt-4.1-mini,1,1.0,0.0,1.0,1.0,1.0
70
+ -0.1474609375,-0.232421875,-0.255859375,-0.15234375,-0.228515625,gpt-4o,0,0.0,0.0,0.0,0.0,0.0
71
+ 0.0361328125,0.028076171875,-0.09130859375,0.08935546875,0.1279296875,gpt-4.1-mini,1,1.0,1.0,1.0,1.0,1.0
72
+ -0.006103515625,-0.06591796875,-0.2177734375,0.06005859375,0.11328125,gpt-4.1-mini,0,0.0,0.0,0.0,0.0,0.0
73
+ -0.19921875,-0.322265625,-0.294921875,-0.2333984375,-0.44921875,gpt-4o,0,0.0,0.0,0.0,0.0,0.0
74
+ -0.034423828125,-0.0238037109375,-0.12890625,-0.06787109375,-0.024169921875,DeepSeek-V3-0324,0,1.0,0.0,0.0,0.0,0.0
75
+ 0.11962890625,-0.0380859375,0.045166015625,0.1171875,0.0498046875,gpt-4o,0,0.0,1.0,0.0,0.0,0.0
76
+ 0.1552734375,0.0162353515625,-0.0234375,0.177734375,0.296875,gpt-4.1-mini,1,1.0,1.0,1.0,0.0,1.0
77
+ -0.259765625,-0.318359375,-0.376953125,-0.291015625,-0.380859375,gpt-4o,1,1.0,1.0,1.0,1.0,1.0
78
+ -0.0184326171875,-0.03466796875,-0.0245361328125,0.07861328125,0.09521484375,gpt-4.1-mini,0,0.0,0.0,0.0,0.0,0.0
79
+ 0.0771484375,0.06201171875,0.01470947265625,0.10791015625,0.2138671875,gpt-4.1-mini,0,0.0,1.0,1.0,0.0,0.0
80
+ -0.0294189453125,-0.034423828125,-0.1904296875,0.0478515625,0.10888671875,gpt-4.1-mini,1,1.0,1.0,1.0,1.0,1.0
81
+ -0.11083984375,-0.024169921875,-0.14453125,-0.08251953125,-0.0703125,DeepSeek-V3-0324,1,1.0,1.0,1.0,1.0,0.0
82
+ 0.055908203125,0.08984375,0.043701171875,0.130859375,0.138671875,gpt-4.1-mini,1,1.0,1.0,1.0,1.0,1.0
83
+ -0.15234375,-0.236328125,-0.275390625,-0.1767578125,-0.234375,gpt-4o,1,1.0,1.0,1.0,1.0,0.0
84
+ 0.169921875,0.051025390625,0.09716796875,0.158203125,0.21875,gpt-4.1-mini,1,0.0,0.0,0.0,0.0,1.0
85
+ -0.17578125,-0.3359375,-0.328125,-0.302734375,-0.376953125,gpt-4o,0,0.0,0.0,0.0,0.0,0.0
86
+ -0.03564453125,-0.1201171875,-0.05126953125,-0.00372314453125,0.010009765625,gpt-4.1-mini,1,1.0,1.0,1.0,1.0,1.0
87
+ -0.125,-0.263671875,-0.263671875,-0.162109375,-0.251953125,gpt-4o,0,0.0,0.0,0.0,0.0,0.0
88
+ 0.03564453125,-0.0284423828125,-0.083984375,-0.07763671875,0.083984375,gpt-4.1-mini,1,1.0,1.0,1.0,1.0,1.0
89
+ -0.06494140625,-0.1953125,-0.1982421875,-0.166015625,-0.169921875,gpt-4o,1,1.0,1.0,1.0,1.0,1.0
90
+ -0.3125,-0.1845703125,-0.236328125,-0.2197265625,-0.392578125,DeepSeek-V3-0324,0,0.0,0.0,0.0,0.0,0.0
91
+ -0.058349609375,-0.09375,-0.068359375,0.01043701171875,-0.0966796875,qwen25-coder-32b-instruct,0,0.0,0.0,0.0,0.0,0.0
92
+ -0.0673828125,-0.12060546875,-0.1484375,-0.1982421875,-0.05419921875,gpt-4.1-mini,1,1.0,0.0,1.0,0.0,1.0
93
+ -0.306640625,-0.251953125,-0.375,-0.23828125,-0.455078125,qwen25-coder-32b-instruct,1,1.0,1.0,1.0,1.0,1.0
94
+ -0.10302734375,-0.1416015625,-0.236328125,-0.0184326171875,-0.037841796875,qwen25-coder-32b-instruct,0,0.0,0.0,0.0,0.0,0.0
95
+ -0.0244140625,-0.1748046875,-0.248046875,-0.031982421875,-0.00726318359375,gpt-4.1-mini,1,1.0,1.0,1.0,1.0,1.0
96
+ -0.142578125,-0.0908203125,-0.189453125,-0.041259765625,-0.2470703125,qwen25-coder-32b-instruct,0,0.0,0.0,0.0,0.0,0.0
97
+ -0.2412109375,-0.435546875,-0.4453125,-0.296875,-0.53515625,gpt-4o,1,1.0,1.0,1.0,1.0,1.0
98
+ -0.0223388671875,-0.0869140625,-0.1611328125,-0.0093994140625,-0.076171875,qwen25-coder-32b-instruct,0,1.0,0.0,0.0,0.0,1.0
99
+ -0.12451171875,-0.11962890625,-0.1435546875,-0.04248046875,-0.1533203125,qwen25-coder-32b-instruct,1,1.0,1.0,1.0,1.0,1.0
100
+ -0.1103515625,-0.330078125,-0.3359375,-0.1337890625,-0.173828125,gpt-4o,1,1.0,1.0,1.0,1.0,1.0
101
+ -0.1943359375,-0.2060546875,-0.259765625,-0.2294921875,-0.1796875,gpt-4.1-mini,0,0.0,0.0,0.0,0.0,0.0
102
+ 0.0306396484375,0.031982421875,0.00921630859375,-0.045166015625,0.05419921875,gpt-4.1-mini,1,1.0,1.0,1.0,1.0,1.0
103
+ -0.09765625,-0.12890625,-0.265625,-0.11181640625,-0.044677734375,gpt-4.1-mini,1,1.0,1.0,1.0,1.0,1.0
104
+ -0.171875,-0.14453125,-0.26953125,-0.2431640625,-0.212890625,DeepSeek-V3-0324,1,1.0,1.0,1.0,1.0,1.0
105
+ -0.04833984375,-0.125,-0.142578125,-0.04150390625,-0.041748046875,qwen25-coder-32b-instruct,0,0.0,0.0,0.0,0.0,0.0
106
+ -0.02294921875,-0.0712890625,-0.1474609375,-0.0286865234375,-0.06494140625,gpt-4o,0,0.0,0.0,0.0,0.0,0.0
107
+ -0.296875,-0.328125,-0.416015625,-0.365234375,-0.34765625,gpt-4o,1,1.0,1.0,0.0,0.0,1.0
108
+ -0.203125,-0.2041015625,-0.228515625,-0.21875,-0.267578125,gpt-4o,1,1.0,1.0,1.0,0.0,1.0
109
+ 0.189453125,0.146484375,0.1875,0.09375,0.255859375,gpt-4.1-mini,1,1.0,1.0,1.0,1.0,1.0
110
+ 0.2041015625,0.173828125,0.1142578125,0.259765625,0.310546875,gpt-4.1-mini,0,0.0,0.0,0.0,0.0,0.0
111
+ -0.2080078125,-0.2021484375,-0.31640625,-0.255859375,-0.23828125,DeepSeek-V3-0324,1,0.0,1.0,0.0,0.0,0.0
112
+ -0.11572265625,-0.2490234375,-0.318359375,-0.23046875,-0.318359375,gpt-4o,1,1.0,1.0,1.0,1.0,1.0
113
+ 0.162109375,0.224609375,0.19140625,0.2412109375,0.423828125,gpt-4.1-mini,1,1.0,1.0,1.0,1.0,1.0
114
+ -0.02685546875,-0.09228515625,-0.1748046875,-0.08447265625,-0.022216796875,gpt-4.1-mini,1,1.0,1.0,1.0,1.0,1.0
115
+ 0.2177734375,0.14453125,0.1552734375,0.201171875,0.357421875,gpt-4.1-mini,1,1.0,1.0,1.0,1.0,1.0
116
+
117
+ predicted_proportions,0.3158,0.0965,0.0351,0.1667,0.3860
118
+ true_proportions,0.1288,0.1442,0.1156,0.1222,0.1295
119
+ correct_fraction,0.5526
special_tokens_map.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "<s>",
3
+ "cls_token": "<s>",
4
+ "eos_token": "</s>",
5
+ "mask_token": {
6
+ "content": "<mask>",
7
+ "lstrip": true,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false
11
+ },
12
+ "pad_token": "<pad>",
13
+ "sep_token": "</s>",
14
+ "unk_token": "<unk>"
15
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "added_tokens_decoder": {
4
+ "0": {
5
+ "content": "<s>",
6
+ "lstrip": false,
7
+ "normalized": true,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "1": {
13
+ "content": "<pad>",
14
+ "lstrip": false,
15
+ "normalized": true,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
+ },
20
+ "2": {
21
+ "content": "</s>",
22
+ "lstrip": false,
23
+ "normalized": true,
24
+ "rstrip": false,
25
+ "single_word": false,
26
+ "special": true
27
+ },
28
+ "3": {
29
+ "content": "<unk>",
30
+ "lstrip": false,
31
+ "normalized": true,
32
+ "rstrip": false,
33
+ "single_word": false,
34
+ "special": true
35
+ },
36
+ "50264": {
37
+ "content": "<mask>",
38
+ "lstrip": true,
39
+ "normalized": false,
40
+ "rstrip": false,
41
+ "single_word": false,
42
+ "special": true
43
+ }
44
+ },
45
+ "bos_token": "<s>",
46
+ "clean_up_tokenization_spaces": false,
47
+ "cls_token": "<s>",
48
+ "eos_token": "</s>",
49
+ "errors": "replace",
50
+ "extra_special_tokens": {},
51
+ "mask_token": "<mask>",
52
+ "model_max_length": 512,
53
+ "pad_token": "<pad>",
54
+ "sep_token": "</s>",
55
+ "tokenizer_class": "RobertaTokenizer",
56
+ "trim_offsets": true,
57
+ "unk_token": "<unk>"
58
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3b3aa12c9ba0848bd3e7ec5dc259cdc0335978c8dc0bcac6cebdd7bf38be3cdd
3
+ size 5713
vocab.json ADDED
The diff for this file is too large to render. See raw diff