henryscheible commited on
Commit
95cc2d1
·
1 Parent(s): 6ed6edc

End of training

Browse files
.gitignore ADDED
@@ -0,0 +1 @@
 
 
1
+ checkpoint-*/
all_results.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "eval_runtime": 1.3546,
3
+ "eval_samples": 146,
4
+ "eval_samples_per_second": 107.785,
5
+ "eval_steps_per_second": 14.027
6
+ }
config.json ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "bert-base-uncased",
3
+ "architectures": [
4
+ "BertForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "finetuning_task": "wnli",
9
+ "gradient_checkpointing": false,
10
+ "hidden_act": "gelu",
11
+ "hidden_dropout_prob": 0.1,
12
+ "hidden_size": 768,
13
+ "id2label": {
14
+ "0": "not_entailment",
15
+ "1": "entailment"
16
+ },
17
+ "initializer_range": 0.02,
18
+ "intermediate_size": 3072,
19
+ "label2id": {
20
+ "entailment": 1,
21
+ "not_entailment": 0
22
+ },
23
+ "layer_norm_eps": 1e-12,
24
+ "max_position_embeddings": 512,
25
+ "model_type": "bert",
26
+ "num_attention_heads": 12,
27
+ "num_hidden_layers": 12,
28
+ "pad_token_id": 0,
29
+ "position_embedding_type": "absolute",
30
+ "torch_dtype": "float32",
31
+ "transformers_version": "4.23.1",
32
+ "type_vocab_size": 2,
33
+ "use_cache": true,
34
+ "vocab_size": 30522
35
+ }
predict_results_wnli.txt ADDED
@@ -0,0 +1,147 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ index prediction
2
+ 0 entailment
3
+ 1 entailment
4
+ 2 entailment
5
+ 3 entailment
6
+ 4 entailment
7
+ 5 entailment
8
+ 6 entailment
9
+ 7 entailment
10
+ 8 entailment
11
+ 9 entailment
12
+ 10 entailment
13
+ 11 entailment
14
+ 12 entailment
15
+ 13 entailment
16
+ 14 entailment
17
+ 15 entailment
18
+ 16 entailment
19
+ 17 entailment
20
+ 18 entailment
21
+ 19 entailment
22
+ 20 entailment
23
+ 21 entailment
24
+ 22 entailment
25
+ 23 entailment
26
+ 24 entailment
27
+ 25 entailment
28
+ 26 entailment
29
+ 27 entailment
30
+ 28 entailment
31
+ 29 entailment
32
+ 30 entailment
33
+ 31 entailment
34
+ 32 entailment
35
+ 33 entailment
36
+ 34 entailment
37
+ 35 entailment
38
+ 36 entailment
39
+ 37 entailment
40
+ 38 entailment
41
+ 39 entailment
42
+ 40 entailment
43
+ 41 entailment
44
+ 42 entailment
45
+ 43 entailment
46
+ 44 entailment
47
+ 45 entailment
48
+ 46 entailment
49
+ 47 entailment
50
+ 48 entailment
51
+ 49 entailment
52
+ 50 entailment
53
+ 51 entailment
54
+ 52 entailment
55
+ 53 entailment
56
+ 54 entailment
57
+ 55 entailment
58
+ 56 entailment
59
+ 57 entailment
60
+ 58 entailment
61
+ 59 entailment
62
+ 60 entailment
63
+ 61 entailment
64
+ 62 entailment
65
+ 63 entailment
66
+ 64 entailment
67
+ 65 entailment
68
+ 66 entailment
69
+ 67 entailment
70
+ 68 entailment
71
+ 69 entailment
72
+ 70 entailment
73
+ 71 entailment
74
+ 72 entailment
75
+ 73 entailment
76
+ 74 entailment
77
+ 75 entailment
78
+ 76 entailment
79
+ 77 entailment
80
+ 78 entailment
81
+ 79 entailment
82
+ 80 entailment
83
+ 81 entailment
84
+ 82 entailment
85
+ 83 entailment
86
+ 84 entailment
87
+ 85 entailment
88
+ 86 entailment
89
+ 87 entailment
90
+ 88 entailment
91
+ 89 entailment
92
+ 90 entailment
93
+ 91 entailment
94
+ 92 entailment
95
+ 93 entailment
96
+ 94 entailment
97
+ 95 entailment
98
+ 96 entailment
99
+ 97 entailment
100
+ 98 entailment
101
+ 99 entailment
102
+ 100 entailment
103
+ 101 entailment
104
+ 102 entailment
105
+ 103 entailment
106
+ 104 entailment
107
+ 105 entailment
108
+ 106 entailment
109
+ 107 entailment
110
+ 108 entailment
111
+ 109 entailment
112
+ 110 entailment
113
+ 111 entailment
114
+ 112 entailment
115
+ 113 entailment
116
+ 114 entailment
117
+ 115 entailment
118
+ 116 entailment
119
+ 117 entailment
120
+ 118 entailment
121
+ 119 entailment
122
+ 120 entailment
123
+ 121 entailment
124
+ 122 entailment
125
+ 123 entailment
126
+ 124 entailment
127
+ 125 entailment
128
+ 126 entailment
129
+ 127 entailment
130
+ 128 entailment
131
+ 129 entailment
132
+ 130 entailment
133
+ 131 entailment
134
+ 132 entailment
135
+ 133 entailment
136
+ 134 entailment
137
+ 135 entailment
138
+ 136 entailment
139
+ 137 entailment
140
+ 138 entailment
141
+ 139 entailment
142
+ 140 entailment
143
+ 141 entailment
144
+ 142 entailment
145
+ 143 entailment
146
+ 144 entailment
147
+ 145 entailment
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b274880346162c391112b33376b3ecf8583125d151d400ab0e8769d0acaa0460
3
+ size 438006125
special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
test_results.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "eval_runtime": 1.3546,
3
+ "eval_samples": 146,
4
+ "eval_samples_per_second": 107.785,
5
+ "eval_steps_per_second": 14.027
6
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "do_lower_case": true,
4
+ "mask_token": "[MASK]",
5
+ "model_max_length": 512,
6
+ "name_or_path": "bert-base-uncased",
7
+ "pad_token": "[PAD]",
8
+ "sep_token": "[SEP]",
9
+ "special_tokens_map_file": null,
10
+ "strip_accents": null,
11
+ "tokenize_chinese_chars": true,
12
+ "tokenizer_class": "BertTokenizer",
13
+ "unk_token": "[UNK]"
14
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:99681700cce1c539a685672853a41f7a1cb27ef463c17fa642580cfe65965840
3
+ size 3375
vocab.txt ADDED
The diff for this file is too large to render. See raw diff