emanuelaboros commited on
Commit
ddb8104
·
1 Parent(s): 5310465

new model

Browse files
config.json CHANGED
@@ -1,232 +1,23 @@
1
  {
2
- "_name_or_path": "experiments_final/model_dbmdz_bert_medium_historic_multilingual_cased_max_sequence_length_512_epochs_5_run_extended_suffix_baseline/checkpoint-450",
3
  "architectures": [
4
- "ExtendedMultitaskModelForTokenClassification"
5
  ],
6
  "attention_probs_dropout_prob": 0.1,
7
- "auto_map": {
8
- "AutoConfig": "configuration_stacked.ImpressoConfig",
9
- "AutoModelForTokenClassification": "modeling_stacked.ExtendedMultitaskModelForTokenClassification"
10
- },
11
  "classifier_dropout": null,
12
- "custom_pipelines": {
13
- "generic-ner": {
14
- "impl": "generic_ner.MultitaskTokenClassificationPipeline",
15
- "pt": "AutoModelForTokenClassification"
16
- }
17
- },
18
  "hidden_act": "gelu",
19
  "hidden_dropout_prob": 0.1,
20
  "hidden_size": 512,
21
  "initializer_range": 0.02,
22
  "intermediate_size": 2048,
23
- "label_map": {
24
- "NE-COARSE-LIT": {
25
- "B-loc": 8,
26
- "B-org": 0,
27
- "B-pers": 7,
28
- "B-prod": 4,
29
- "B-time": 5,
30
- "I-loc": 1,
31
- "I-org": 2,
32
- "I-pers": 9,
33
- "I-prod": 10,
34
- "I-time": 6,
35
- "O": 3
36
- },
37
- "NE-COARSE-METO": {
38
- "B-loc": 3,
39
- "B-org": 0,
40
- "B-time": 5,
41
- "I-loc": 4,
42
- "I-org": 2,
43
- "O": 1
44
- },
45
- "NE-FINE-COMP": {
46
- "B-comp.demonym": 8,
47
- "B-comp.function": 5,
48
- "B-comp.name": 1,
49
- "B-comp.qualifier": 9,
50
- "B-comp.title": 2,
51
- "I-comp.demonym": 7,
52
- "I-comp.function": 3,
53
- "I-comp.name": 0,
54
- "I-comp.qualifier": 10,
55
- "I-comp.title": 4,
56
- "O": 6
57
- },
58
- "NE-FINE-LIT": {
59
- "B-loc.add.elec": 32,
60
- "B-loc.add.phys": 5,
61
- "B-loc.adm.nat": 34,
62
- "B-loc.adm.reg": 39,
63
- "B-loc.adm.sup": 12,
64
- "B-loc.adm.town": 33,
65
- "B-loc.fac": 36,
66
- "B-loc.oro": 19,
67
- "B-loc.phys.geo": 13,
68
- "B-loc.phys.hydro": 28,
69
- "B-loc.unk": 4,
70
- "B-org.adm": 3,
71
- "B-org.ent": 24,
72
- "B-org.ent.pressagency": 37,
73
- "B-pers.coll": 9,
74
- "B-pers.ind": 0,
75
- "B-pers.ind.articleauthor": 20,
76
- "B-prod.doctr": 2,
77
- "B-prod.media": 10,
78
- "B-time.date.abs": 23,
79
- "I-loc.add.elec": 22,
80
- "I-loc.add.phys": 6,
81
- "I-loc.adm.nat": 11,
82
- "I-loc.adm.reg": 35,
83
- "I-loc.adm.sup": 15,
84
- "I-loc.adm.town": 8,
85
- "I-loc.fac": 27,
86
- "I-loc.oro": 21,
87
- "I-loc.phys.geo": 25,
88
- "I-loc.phys.hydro": 17,
89
- "I-loc.unk": 40,
90
- "I-org.adm": 29,
91
- "I-org.ent": 1,
92
- "I-org.ent.pressagency": 14,
93
- "I-pers.coll": 26,
94
- "I-pers.ind": 16,
95
- "I-pers.ind.articleauthor": 31,
96
- "I-prod.doctr": 30,
97
- "I-prod.media": 38,
98
- "I-time.date.abs": 7,
99
- "O": 18
100
- },
101
- "NE-FINE-METO": {
102
- "B-loc.adm.town": 6,
103
- "B-loc.fac": 3,
104
- "B-loc.oro": 5,
105
- "B-org.adm": 1,
106
- "B-org.ent": 7,
107
- "B-time.date.abs": 9,
108
- "I-loc.fac": 8,
109
- "I-org.adm": 2,
110
- "I-org.ent": 0,
111
- "O": 4
112
- },
113
- "NE-NESTED": {
114
- "B-loc.adm.nat": 13,
115
- "B-loc.adm.reg": 15,
116
- "B-loc.adm.sup": 10,
117
- "B-loc.adm.town": 9,
118
- "B-loc.fac": 18,
119
- "B-loc.oro": 17,
120
- "B-loc.phys.geo": 11,
121
- "B-loc.phys.hydro": 1,
122
- "B-org.adm": 4,
123
- "B-org.ent": 20,
124
- "B-pers.coll": 7,
125
- "B-pers.ind": 2,
126
- "B-prod.media": 23,
127
- "I-loc.adm.nat": 8,
128
- "I-loc.adm.reg": 14,
129
- "I-loc.adm.town": 6,
130
- "I-loc.fac": 0,
131
- "I-loc.oro": 19,
132
- "I-loc.phys.geo": 21,
133
- "I-loc.phys.hydro": 22,
134
- "I-org.adm": 5,
135
- "I-org.ent": 3,
136
- "I-pers.ind": 12,
137
- "I-prod.media": 24,
138
- "O": 16
139
- }
140
- },
141
  "layer_norm_eps": 1e-12,
142
  "max_position_embeddings": 512,
143
- "model_type": "stacked_bert",
144
  "num_attention_heads": 8,
145
  "num_hidden_layers": 8,
146
  "pad_token_id": 0,
147
  "position_embedding_type": "absolute",
148
- "pretrained_config": {
149
- "_name_or_path": "dbmdz/bert-medium-historic-multilingual-cased",
150
- "add_cross_attention": false,
151
- "architectures": [
152
- "BertForMaskedLM"
153
- ],
154
- "attention_probs_dropout_prob": 0.1,
155
- "bad_words_ids": null,
156
- "begin_suppress_tokens": null,
157
- "bos_token_id": null,
158
- "chunk_size_feed_forward": 0,
159
- "classifier_dropout": null,
160
- "cross_attention_hidden_size": null,
161
- "decoder_start_token_id": null,
162
- "diversity_penalty": 0.0,
163
- "do_sample": false,
164
- "early_stopping": false,
165
- "encoder_no_repeat_ngram_size": 0,
166
- "eos_token_id": null,
167
- "exponential_decay_length_penalty": null,
168
- "finetuning_task": null,
169
- "forced_bos_token_id": null,
170
- "forced_eos_token_id": null,
171
- "hidden_act": "gelu",
172
- "hidden_dropout_prob": 0.1,
173
- "hidden_size": 512,
174
- "id2label": {
175
- "0": "LABEL_0",
176
- "1": "LABEL_1"
177
- },
178
- "initializer_range": 0.02,
179
- "intermediate_size": 2048,
180
- "is_decoder": false,
181
- "is_encoder_decoder": false,
182
- "label2id": {
183
- "LABEL_0": 0,
184
- "LABEL_1": 1
185
- },
186
- "layer_norm_eps": 1e-12,
187
- "length_penalty": 1.0,
188
- "max_length": 20,
189
- "max_position_embeddings": 512,
190
- "min_length": 0,
191
- "model_type": "bert",
192
- "no_repeat_ngram_size": 0,
193
- "num_attention_heads": 8,
194
- "num_beam_groups": 1,
195
- "num_beams": 1,
196
- "num_hidden_layers": 8,
197
- "num_return_sequences": 1,
198
- "output_attentions": false,
199
- "output_hidden_states": false,
200
- "output_scores": false,
201
- "pad_token_id": 0,
202
- "position_embedding_type": "absolute",
203
- "prefix": null,
204
- "problem_type": null,
205
- "pruned_heads": {},
206
- "remove_invalid_values": false,
207
- "repetition_penalty": 1.0,
208
- "return_dict": true,
209
- "return_dict_in_generate": false,
210
- "sep_token_id": null,
211
- "suppress_tokens": null,
212
- "task_specific_params": null,
213
- "temperature": 1.0,
214
- "tf_legacy_loss": false,
215
- "tie_encoder_decoder": false,
216
- "tie_word_embeddings": true,
217
- "tokenizer_class": null,
218
- "top_k": 50,
219
- "top_p": 1.0,
220
- "torch_dtype": null,
221
- "torchscript": false,
222
- "type_vocab_size": 2,
223
- "typical_p": 1.0,
224
- "use_bfloat16": false,
225
- "use_cache": true,
226
- "vocab_size": 32000
227
- },
228
- "torch_dtype": "float32",
229
- "transformers_version": "4.40.0.dev0",
230
  "type_vocab_size": 2,
231
  "use_cache": true,
232
  "vocab_size": 32000
 
1
  {
 
2
  "architectures": [
3
+ "MultitaskTimeModelForTokenClassification"
4
  ],
5
  "attention_probs_dropout_prob": 0.1,
 
 
 
 
6
  "classifier_dropout": null,
7
+ "dtype": "float32",
 
 
 
 
 
8
  "hidden_act": "gelu",
9
  "hidden_dropout_prob": 0.1,
10
  "hidden_size": 512,
11
  "initializer_range": 0.02,
12
  "intermediate_size": 2048,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
13
  "layer_norm_eps": 1e-12,
14
  "max_position_embeddings": 512,
15
+ "model_type": "bert",
16
  "num_attention_heads": 8,
17
  "num_hidden_layers": 8,
18
  "pad_token_id": 0,
19
  "position_embedding_type": "absolute",
20
+ "transformers_version": "4.57.0.dev0",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
21
  "type_vocab_size": 2,
22
  "use_cache": true,
23
  "vocab_size": 32000
special_tokens_map.json CHANGED
@@ -1,37 +1,7 @@
1
  {
2
- "cls_token": {
3
- "content": "[CLS]",
4
- "lstrip": false,
5
- "normalized": false,
6
- "rstrip": false,
7
- "single_word": false
8
- },
9
- "mask_token": {
10
- "content": "[MASK]",
11
- "lstrip": false,
12
- "normalized": false,
13
- "rstrip": false,
14
- "single_word": false
15
- },
16
- "pad_token": {
17
- "content": "[PAD]",
18
- "lstrip": false,
19
- "normalized": false,
20
- "rstrip": false,
21
- "single_word": false
22
- },
23
- "sep_token": {
24
- "content": "[SEP]",
25
- "lstrip": false,
26
- "normalized": false,
27
- "rstrip": false,
28
- "single_word": false
29
- },
30
- "unk_token": {
31
- "content": "[UNK]",
32
- "lstrip": false,
33
- "normalized": false,
34
- "rstrip": false,
35
- "single_word": false
36
- }
37
  }
 
1
  {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
7
  }
tokenizer_config.json CHANGED
@@ -45,6 +45,7 @@
45
  "cls_token": "[CLS]",
46
  "do_basic_tokenize": true,
47
  "do_lower_case": false,
 
48
  "mask_token": "[MASK]",
49
  "max_len": 512,
50
  "model_max_length": 512,
 
45
  "cls_token": "[CLS]",
46
  "do_basic_tokenize": true,
47
  "do_lower_case": false,
48
+ "extra_special_tokens": {},
49
  "mask_token": "[MASK]",
50
  "max_len": 512,
51
  "model_max_length": 512,
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cc92dca5d693d80c40bfa708d0ee9551d1f85b832c57710b3edfc72dc86707e1
3
+ size 2104