permutans commited on
Commit
d64c032
·
verified ·
1 Parent(s): 89df35e

Upload folder using huggingface_hub

Browse files
Files changed (5) hide show
  1. README.md +69 -69
  2. config.json +73 -67
  3. head_config.json +1 -1
  4. model.safetensors +2 -2
  5. type_to_idx.json +12 -11
README.md CHANGED
@@ -22,7 +22,7 @@ datasets:
22
 
23
  BERT-based token classifier for detecting **oral and literate markers** in text, based on Walter Ong's "Orality and Literacy" (1982).
24
 
25
- This model performs multi-label span-level detection of 52 rhetorical marker types, where each token independently carries B/I/O labels per type — allowing overlapping spans (e.g. a token that is simultaneously part of a concessive and a nested clause).
26
 
27
  ## Model Details
28
 
@@ -30,8 +30,8 @@ This model performs multi-label span-level detection of 52 rhetorical marker typ
30
  |----------|-------|
31
  | Base model | `bert-base-uncased` |
32
  | Task | Multi-label token classification (independent B/I/O per type) |
33
- | Marker types | 52 (21 oral, 31 literate) |
34
- | Test macro F1 | **0.394** (per-type detection, binary positive = B or I) |
35
  | Training | 20 epochs, batch 24, lr 3e-5, fp16 |
36
  | Regularization | Mixout (p=0.1) — stochastic L2 anchor to pretrained weights |
37
  | Loss | Per-type weighted cross-entropy with inverse-frequency type weights |
@@ -80,16 +80,16 @@ for i, token in enumerate(tokens):
80
  - Types with fewer than 150 annotated spans are excluded from training
81
  - Multi-label BIO annotation: tokens can carry labels for multiple overlapping marker types simultaneously
82
 
83
- ## Marker Types (52)
84
 
85
- ### Oral Markers (21 types)
86
 
87
  Characteristics of oral tradition and spoken discourse:
88
 
89
  | Category | Markers |
90
  |----------|---------|
91
  | **Address & Interaction** | vocative, imperative, second_person, inclusive_we, rhetorical_question, phatic_check, phatic_filler |
92
- | **Repetition & Pattern** | anaphora, tricolon, lexical_repetition, antithesis |
93
  | **Conjunction** | simple_conjunction |
94
  | **Formulas** | discourse_formula, intensifier_doubling |
95
  | **Narrative** | named_individual, specific_place, temporal_anchor, sensory_detail, embodied_action, everyday_example |
@@ -115,75 +115,75 @@ Characteristics of written, analytical discourse:
115
  Per-type detection F1 on test set (binary: B or I = positive, O = negative):
116
 
117
  <details><summary>Click to show per-marker precision/recall/F1/support</summary>
118
-
119
  ```
120
  Type Prec Rec F1 Sup
121
  ========================================================================
122
- literate_abstract_noun 0.283 0.036 0.064 474
123
- literate_additive_formal 0.458 0.388 0.420 85
124
- literate_agent_demoted 0.495 0.569 0.530 288
125
- literate_agentless_passive 0.659 0.592 0.624 1285
126
- literate_aside 0.468 0.524 0.494 481
127
- literate_categorical_statement 0.256 0.141 0.182 389
128
- literate_causal_explicit 0.457 0.196 0.275 382
129
- literate_citation 0.624 0.539 0.578 243
130
- literate_conceptual_metaphor 0.366 0.242 0.291 219
131
- literate_concessive 0.558 0.290 0.382 742
132
- literate_concessive_connector 0.286 0.324 0.304 37
133
- literate_concrete_setting 0.222 0.132 0.166 303
134
- literate_conditional 0.664 0.597 0.629 1642
135
- literate_contrastive 0.481 0.227 0.308 388
136
- literate_cross_reference 0.644 0.326 0.433 89
137
- literate_definitional_move 0.279 0.284 0.281 67
138
- literate_enumeration 0.507 0.580 0.541 855
139
- literate_epistemic_hedge 0.523 0.405 0.456 543
140
- literate_evidential 0.487 0.457 0.471 162
141
- literate_institutional_subject 0.330 0.274 0.300 248
142
- literate_list_structure 0.929 0.464 0.619 653
143
- literate_metadiscourse 0.355 0.251 0.294 355
144
- literate_nested_clauses 0.212 0.140 0.169 1250
145
- literate_nominalization 0.527 0.397 0.453 1147
146
- literate_objectifying_stance 0.593 0.400 0.478 200
147
- literate_probability 0.740 0.544 0.627 136
148
- literate_qualified_assertion 0.153 0.073 0.099 123
149
- literate_relative_chain 0.333 0.179 0.233 1717
150
- literate_technical_abbreviation 0.613 0.725 0.665 153
151
- literate_technical_term 0.490 0.311 0.381 897
152
- literate_temporal_embedding 0.210 0.143 0.170 553
153
- oral_anaphora 0.205 0.128 0.157 141
154
- oral_antithesis 0.389 0.181 0.247 453
155
- oral_discourse_formula 0.557 0.173 0.263 568
156
- oral_embodied_action 0.421 0.213 0.283 489
157
- oral_everyday_example 0.219 0.209 0.214 358
158
- oral_imperative 0.537 0.695 0.606 200
159
- oral_inclusive_we 0.616 0.599 0.608 751
160
- oral_intensifier_doubling 0.632 0.152 0.245 79
161
- oral_lexical_repetition 0.406 0.468 0.435 218
162
- oral_named_individual 0.535 0.566 0.550 813
163
- oral_phatic_check 0.591 0.684 0.634 76
164
- oral_phatic_filler 0.469 0.524 0.495 189
165
- oral_rhetorical_question 0.677 0.646 0.661 1273
166
- oral_second_person 0.618 0.493 0.549 842
167
- oral_self_correction 0.582 0.205 0.303 156
168
- oral_sensory_detail 0.281 0.247 0.263 352
169
- oral_simple_conjunction 0.146 0.085 0.107 71
170
- oral_specific_place 0.534 0.582 0.557 373
171
- oral_temporal_anchor 0.518 0.510 0.514 563
172
- oral_tricolon 0.247 0.185 0.212 562
173
- oral_vocative 0.667 0.684 0.675 158
 
174
  ========================================================================
175
- Macro avg (types w/ support) 0.394
176
  ```
177
 
178
  </details>
179
 
180
- **Missing labels (test set):** 0/52 — all types detected at least once.
181
 
182
  Notable patterns:
183
- - **Strong performers** (F1 > 0.5): vocative (0.675), technical_abbreviation (0.665), rhetorical_question (0.661), phatic_check (0.634), conditional (0.629), probability (0.627), agentless_passive (0.624), list_structure (0.619), inclusive_we (0.608), imperative (0.606), citation (0.578), specific_place (0.557), named_individual (0.550), second_person (0.549), enumeration (0.541), agent_demoted (0.530), temporal_anchor (0.514)
184
- - **Weak performers** (F1 < 0.2): abstract_noun (0.064), qualified_assertion (0.099), simple_conjunction (0.107), anaphora (0.157), concrete_setting (0.166), nested_clauses (0.169), temporal_embedding (0.170), categorical_statement (0.182)
185
- - **Precision-recall tradeoff**: Most types now show higher precision than recall, indicating the model under-predicts rather than over-predicts markers (reversed from the previous release)
186
- - **Dropped type**: `oral_parallelism` was excluded from this training run (fell below the 150-span minimum threshold)
187
 
188
  ## Architecture
189
 
@@ -215,9 +215,9 @@ classifier.bias → randomly initialized
215
 
216
  ## Limitations
217
 
218
- - **Low-precision types**: Several types show precision below 0.25, meaning most predictions for those types are false positives
219
- - **Low-recall types**: `abstract_noun` (0.036 recall), `simple_conjunction` (0.085), and `qualified_assertion` (0.073) are near-invisible to the model despite nonzero precision
220
- - **Excluded type**: `oral_parallelism` fell below the 150-span minimum and was excluded; structural parallelism remains undetected
221
  - **Context window**: 128 tokens max; longer spans may be truncated
222
  - **Domain**: Trained primarily on historical/literary texts; may underperform on modern social media
223
  - **Subjectivity**: Some marker boundaries are inherently ambiguous
@@ -239,4 +239,4 @@ classifier.bias → randomly initialized
239
 
240
  ---
241
 
242
- *Trained: February 2026*
 
22
 
23
  BERT-based token classifier for detecting **oral and literate markers** in text, based on Walter Ong's "Orality and Literacy" (1982).
24
 
25
+ This model performs multi-label span-level detection of 53 rhetorical marker types, where each token independently carries B/I/O labels per type — allowing overlapping spans (e.g. a token that is simultaneously part of a concessive and a nested clause).
26
 
27
  ## Model Details
28
 
 
30
  |----------|-------|
31
  | Base model | `bert-base-uncased` |
32
  | Task | Multi-label token classification (independent B/I/O per type) |
33
+ | Marker types | 53 (22 oral, 31 literate) |
34
+ | Test macro F1 | **0.400** (per-type detection, binary positive = B or I) |
35
  | Training | 20 epochs, batch 24, lr 3e-5, fp16 |
36
  | Regularization | Mixout (p=0.1) — stochastic L2 anchor to pretrained weights |
37
  | Loss | Per-type weighted cross-entropy with inverse-frequency type weights |
 
80
  - Types with fewer than 150 annotated spans are excluded from training
81
  - Multi-label BIO annotation: tokens can carry labels for multiple overlapping marker types simultaneously
82
 
83
+ ## Marker Types (53)
84
 
85
+ ### Oral Markers (22 types)
86
 
87
  Characteristics of oral tradition and spoken discourse:
88
 
89
  | Category | Markers |
90
  |----------|---------|
91
  | **Address & Interaction** | vocative, imperative, second_person, inclusive_we, rhetorical_question, phatic_check, phatic_filler |
92
+ | **Repetition & Pattern** | anaphora, parallelism, tricolon, lexical_repetition, antithesis |
93
  | **Conjunction** | simple_conjunction |
94
  | **Formulas** | discourse_formula, intensifier_doubling |
95
  | **Narrative** | named_individual, specific_place, temporal_anchor, sensory_detail, embodied_action, everyday_example |
 
115
  Per-type detection F1 on test set (binary: B or I = positive, O = negative):
116
 
117
  <details><summary>Click to show per-marker precision/recall/F1/support</summary>
 
118
  ```
119
  Type Prec Rec F1 Sup
120
  ========================================================================
121
+ literate_abstract_noun 0.211 0.319 0.254 464
122
+ literate_additive_formal 0.263 0.506 0.346 83
123
+ literate_agent_demoted 0.364 0.629 0.461 291
124
+ literate_agentless_passive 0.545 0.701 0.613 1274
125
+ literate_aside 0.396 0.565 0.466 467
126
+ literate_categorical_statement 0.246 0.245 0.245 388
127
+ literate_causal_explicit 0.325 0.305 0.315 370
128
+ literate_citation 0.500 0.551 0.524 243
129
+ literate_conceptual_metaphor 0.168 0.297 0.215 219
130
+ literate_concessive 0.542 0.491 0.515 731
131
+ literate_concessive_connector 0.113 0.378 0.174 37
132
+ literate_concrete_setting 0.174 0.279 0.214 301
133
+ literate_conditional 0.586 0.710 0.642 1610
134
+ literate_contrastive 0.374 0.343 0.358 382
135
+ literate_cross_reference 0.351 0.317 0.333 82
136
+ literate_definitional_move 0.217 0.371 0.274 62
137
+ literate_enumeration 0.456 0.570 0.507 899
138
+ literate_epistemic_hedge 0.415 0.511 0.458 534
139
+ literate_evidential 0.364 0.503 0.422 175
140
+ literate_institutional_subject 0.296 0.520 0.378 246
141
+ literate_list_structure 0.709 0.559 0.625 653
142
+ literate_metadiscourse 0.291 0.451 0.354 355
143
+ literate_nested_clauses 0.105 0.266 0.151 1250
144
+ literate_nominalization 0.475 0.554 0.511 1144
145
+ literate_objectifying_stance 0.518 0.448 0.481 194
146
+ literate_probability 0.612 0.548 0.578 115
147
+ literate_qualified_assertion 0.185 0.160 0.172 125
148
+ literate_relative_chain 0.320 0.537 0.401 1713
149
+ literate_technical_abbreviation 0.545 0.783 0.643 161
150
+ literate_technical_term 0.331 0.458 0.384 909
151
+ literate_temporal_embedding 0.222 0.249 0.235 570
152
+ oral_anaphora 0.207 0.248 0.226 137
153
+ oral_antithesis 0.245 0.289 0.265 453
154
+ oral_discourse_formula 0.353 0.384 0.368 563
155
+ oral_embodied_action 0.263 0.374 0.309 470
156
+ oral_everyday_example 0.160 0.164 0.162 366
157
+ oral_imperative 0.519 0.670 0.585 200
158
+ oral_inclusive_we 0.587 0.672 0.626 752
159
+ oral_intensifier_doubling 0.310 0.165 0.215 79
160
+ oral_lexical_repetition 0.293 0.488 0.366 217
161
+ oral_named_individual 0.428 0.676 0.524 791
162
+ oral_parallelism 0.654 0.048 0.089 710
163
+ oral_phatic_check 0.465 0.882 0.609 76
164
+ oral_phatic_filler 0.375 0.582 0.456 182
165
+ oral_rhetorical_question 0.589 0.894 0.710 1264
166
+ oral_second_person 0.614 0.545 0.577 833
167
+ oral_self_correction 0.597 0.295 0.395 156
168
+ oral_sensory_detail 0.275 0.312 0.293 352
169
+ oral_simple_conjunction 0.096 0.211 0.132 71
170
+ oral_specific_place 0.472 0.716 0.569 387
171
+ oral_temporal_anchor 0.397 0.748 0.518 551
172
+ oral_tricolon 0.274 0.285 0.280 557
173
+ oral_vocative 0.634 0.761 0.692 155
174
  ========================================================================
175
+ Macro avg (types w/ support) 0.400
176
  ```
177
 
178
  </details>
179
 
180
+ **Missing labels (test set):** 0/53 — all types detected at least once.
181
 
182
  Notable patterns:
183
+ - **Strong performers** (F1 > 0.5): rhetorical_question (0.710), vocative (0.692), conditional (0.642), technical_abbreviation (0.643), inclusive_we (0.626), list_structure (0.625), agentless_passive (0.613), phatic_check (0.609), imperative (0.585), probability (0.578), second_person (0.577), specific_place (0.569), citation (0.524), named_individual (0.524), temporal_anchor (0.518), concessive (0.515), nominalization (0.511), enumeration (0.507)
184
+ - **Weak performers** (F1 < 0.2): parallelism (0.089), simple_conjunction (0.132), nested_clauses (0.151), everyday_example (0.162), qualified_assertion (0.172), concessive_connector (0.174)
185
+ - **Precision-recall tradeoff**: Most types now show higher recall than precision, indicating the model over-predicts markers reversed from the previous release. Notable exceptions include `parallelism` (0.654 precision / 0.048 recall), `self_correction`, and `intensifier_doubling`, which remain high-precision but low-recall.
186
+ - **Recovered type**: `oral_parallelism` crossed the 150-span threshold and was re-included, though its near-zero recall (0.048) means it is effectively non-functional despite high precision when it does fire.
187
 
188
  ## Architecture
189
 
 
215
 
216
  ## Limitations
217
 
218
+ - **Recall-dominated errors**: Most types over-predict (recall > precision), producing false positives; downstream applications may need confidence thresholding
219
+ - **Near-zero recall types**: `parallelism` (0.048 recall), `intensifier_doubling` (0.165), and `simple_conjunction` (0.211) are rarely detected despite being present in training data
220
+ - **Low-precision types**: `simple_conjunction` (0.096), `nested_clauses` (0.105), and `concessive_connector` (0.113) have precision below 0.15, meaning most predictions for those types are false positives
221
  - **Context window**: 128 tokens max; longer spans may be truncated
222
  - **Domain**: Trained primarily on historical/literary texts; may underperform on modern social media
223
  - **Subjectivity**: Some marker boundaries are inherently ambiguous
 
239
 
240
  ---
241
 
242
+ *Trained: February 2026*
config.json CHANGED
@@ -44,42 +44,45 @@
44
  "120": "O-oral_named_individual",
45
  "121": "B-oral_named_individual",
46
  "122": "I-oral_named_individual",
47
- "123": "O-oral_phatic_check",
48
- "124": "B-oral_phatic_check",
49
- "125": "I-oral_phatic_check",
50
- "126": "O-oral_phatic_filler",
51
- "127": "B-oral_phatic_filler",
52
- "128": "I-oral_phatic_filler",
53
- "129": "O-oral_rhetorical_question",
54
  "13": "B-literate_aside",
55
- "130": "B-oral_rhetorical_question",
56
- "131": "I-oral_rhetorical_question",
57
- "132": "O-oral_second_person",
58
- "133": "B-oral_second_person",
59
- "134": "I-oral_second_person",
60
- "135": "O-oral_self_correction",
61
- "136": "B-oral_self_correction",
62
- "137": "I-oral_self_correction",
63
- "138": "O-oral_sensory_detail",
64
- "139": "B-oral_sensory_detail",
65
  "14": "I-literate_aside",
66
- "140": "I-oral_sensory_detail",
67
- "141": "O-oral_simple_conjunction",
68
- "142": "B-oral_simple_conjunction",
69
- "143": "I-oral_simple_conjunction",
70
- "144": "O-oral_specific_place",
71
- "145": "B-oral_specific_place",
72
- "146": "I-oral_specific_place",
73
- "147": "O-oral_temporal_anchor",
74
- "148": "B-oral_temporal_anchor",
75
- "149": "I-oral_temporal_anchor",
76
  "15": "O-literate_categorical_statement",
77
- "150": "O-oral_tricolon",
78
- "151": "B-oral_tricolon",
79
- "152": "I-oral_tricolon",
80
- "153": "O-oral_vocative",
81
- "154": "B-oral_vocative",
82
- "155": "I-oral_vocative",
 
 
 
83
  "16": "B-literate_categorical_statement",
84
  "17": "I-literate_categorical_statement",
85
  "18": "O-literate_causal_explicit",
@@ -218,17 +221,18 @@
218
  "B-oral_intensifier_doubling": 115,
219
  "B-oral_lexical_repetition": 118,
220
  "B-oral_named_individual": 121,
221
- "B-oral_phatic_check": 124,
222
- "B-oral_phatic_filler": 127,
223
- "B-oral_rhetorical_question": 130,
224
- "B-oral_second_person": 133,
225
- "B-oral_self_correction": 136,
226
- "B-oral_sensory_detail": 139,
227
- "B-oral_simple_conjunction": 142,
228
- "B-oral_specific_place": 145,
229
- "B-oral_temporal_anchor": 148,
230
- "B-oral_tricolon": 151,
231
- "B-oral_vocative": 154,
 
232
  "I-literate_abstract_noun": 2,
233
  "I-literate_additive_formal": 5,
234
  "I-literate_agent_demoted": 8,
@@ -270,17 +274,18 @@
270
  "I-oral_intensifier_doubling": 116,
271
  "I-oral_lexical_repetition": 119,
272
  "I-oral_named_individual": 122,
273
- "I-oral_phatic_check": 125,
274
- "I-oral_phatic_filler": 128,
275
- "I-oral_rhetorical_question": 131,
276
- "I-oral_second_person": 134,
277
- "I-oral_self_correction": 137,
278
- "I-oral_sensory_detail": 140,
279
- "I-oral_simple_conjunction": 143,
280
- "I-oral_specific_place": 146,
281
- "I-oral_temporal_anchor": 149,
282
- "I-oral_tricolon": 152,
283
- "I-oral_vocative": 155,
 
284
  "O-literate_abstract_noun": 0,
285
  "O-literate_additive_formal": 3,
286
  "O-literate_agent_demoted": 6,
@@ -322,24 +327,25 @@
322
  "O-oral_intensifier_doubling": 114,
323
  "O-oral_lexical_repetition": 117,
324
  "O-oral_named_individual": 120,
325
- "O-oral_phatic_check": 123,
326
- "O-oral_phatic_filler": 126,
327
- "O-oral_rhetorical_question": 129,
328
- "O-oral_second_person": 132,
329
- "O-oral_self_correction": 135,
330
- "O-oral_sensory_detail": 138,
331
- "O-oral_simple_conjunction": 141,
332
- "O-oral_specific_place": 144,
333
- "O-oral_temporal_anchor": 147,
334
- "O-oral_tricolon": 150,
335
- "O-oral_vocative": 153
 
336
  },
337
  "layer_norm_eps": 1e-12,
338
  "max_position_embeddings": 512,
339
  "model_type": "bert",
340
  "num_attention_heads": 12,
341
  "num_hidden_layers": 12,
342
- "num_types": 52,
343
  "pad_token_id": 0,
344
  "position_embedding_type": "absolute",
345
  "tie_word_embeddings": true,
 
44
  "120": "O-oral_named_individual",
45
  "121": "B-oral_named_individual",
46
  "122": "I-oral_named_individual",
47
+ "123": "O-oral_parallelism",
48
+ "124": "B-oral_parallelism",
49
+ "125": "I-oral_parallelism",
50
+ "126": "O-oral_phatic_check",
51
+ "127": "B-oral_phatic_check",
52
+ "128": "I-oral_phatic_check",
53
+ "129": "O-oral_phatic_filler",
54
  "13": "B-literate_aside",
55
+ "130": "B-oral_phatic_filler",
56
+ "131": "I-oral_phatic_filler",
57
+ "132": "O-oral_rhetorical_question",
58
+ "133": "B-oral_rhetorical_question",
59
+ "134": "I-oral_rhetorical_question",
60
+ "135": "O-oral_second_person",
61
+ "136": "B-oral_second_person",
62
+ "137": "I-oral_second_person",
63
+ "138": "O-oral_self_correction",
64
+ "139": "B-oral_self_correction",
65
  "14": "I-literate_aside",
66
+ "140": "I-oral_self_correction",
67
+ "141": "O-oral_sensory_detail",
68
+ "142": "B-oral_sensory_detail",
69
+ "143": "I-oral_sensory_detail",
70
+ "144": "O-oral_simple_conjunction",
71
+ "145": "B-oral_simple_conjunction",
72
+ "146": "I-oral_simple_conjunction",
73
+ "147": "O-oral_specific_place",
74
+ "148": "B-oral_specific_place",
75
+ "149": "I-oral_specific_place",
76
  "15": "O-literate_categorical_statement",
77
+ "150": "O-oral_temporal_anchor",
78
+ "151": "B-oral_temporal_anchor",
79
+ "152": "I-oral_temporal_anchor",
80
+ "153": "O-oral_tricolon",
81
+ "154": "B-oral_tricolon",
82
+ "155": "I-oral_tricolon",
83
+ "156": "O-oral_vocative",
84
+ "157": "B-oral_vocative",
85
+ "158": "I-oral_vocative",
86
  "16": "B-literate_categorical_statement",
87
  "17": "I-literate_categorical_statement",
88
  "18": "O-literate_causal_explicit",
 
221
  "B-oral_intensifier_doubling": 115,
222
  "B-oral_lexical_repetition": 118,
223
  "B-oral_named_individual": 121,
224
+ "B-oral_parallelism": 124,
225
+ "B-oral_phatic_check": 127,
226
+ "B-oral_phatic_filler": 130,
227
+ "B-oral_rhetorical_question": 133,
228
+ "B-oral_second_person": 136,
229
+ "B-oral_self_correction": 139,
230
+ "B-oral_sensory_detail": 142,
231
+ "B-oral_simple_conjunction": 145,
232
+ "B-oral_specific_place": 148,
233
+ "B-oral_temporal_anchor": 151,
234
+ "B-oral_tricolon": 154,
235
+ "B-oral_vocative": 157,
236
  "I-literate_abstract_noun": 2,
237
  "I-literate_additive_formal": 5,
238
  "I-literate_agent_demoted": 8,
 
274
  "I-oral_intensifier_doubling": 116,
275
  "I-oral_lexical_repetition": 119,
276
  "I-oral_named_individual": 122,
277
+ "I-oral_parallelism": 125,
278
+ "I-oral_phatic_check": 128,
279
+ "I-oral_phatic_filler": 131,
280
+ "I-oral_rhetorical_question": 134,
281
+ "I-oral_second_person": 137,
282
+ "I-oral_self_correction": 140,
283
+ "I-oral_sensory_detail": 143,
284
+ "I-oral_simple_conjunction": 146,
285
+ "I-oral_specific_place": 149,
286
+ "I-oral_temporal_anchor": 152,
287
+ "I-oral_tricolon": 155,
288
+ "I-oral_vocative": 158,
289
  "O-literate_abstract_noun": 0,
290
  "O-literate_additive_formal": 3,
291
  "O-literate_agent_demoted": 6,
 
327
  "O-oral_intensifier_doubling": 114,
328
  "O-oral_lexical_repetition": 117,
329
  "O-oral_named_individual": 120,
330
+ "O-oral_parallelism": 123,
331
+ "O-oral_phatic_check": 126,
332
+ "O-oral_phatic_filler": 129,
333
+ "O-oral_rhetorical_question": 132,
334
+ "O-oral_second_person": 135,
335
+ "O-oral_self_correction": 138,
336
+ "O-oral_sensory_detail": 141,
337
+ "O-oral_simple_conjunction": 144,
338
+ "O-oral_specific_place": 147,
339
+ "O-oral_temporal_anchor": 150,
340
+ "O-oral_tricolon": 153,
341
+ "O-oral_vocative": 156
342
  },
343
  "layer_norm_eps": 1e-12,
344
  "max_position_embeddings": 512,
345
  "model_type": "bert",
346
  "num_attention_heads": 12,
347
  "num_hidden_layers": 12,
348
+ "num_types": 53,
349
  "pad_token_id": 0,
350
  "position_embedding_type": "absolute",
351
  "tie_word_embeddings": true,
head_config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
  "model_name": "bert-base-uncased",
3
- "num_types": 52,
4
  "hidden_size": 768
5
  }
 
1
  {
2
  "model_name": "bert-base-uncased",
3
+ "num_types": 53,
4
  "hidden_size": 768
5
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e7f65c36ddbc7fa2756a9e31ff2735c9708f2d891e4d637519a90675c6aa7088
3
- size 436073152
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:41995685c78ead06fdda874b90a8bdf7b283997fa076207a33c0bd7136179ef3
3
+ size 436082548
type_to_idx.json CHANGED
@@ -40,15 +40,16 @@
40
  "oral_intensifier_doubling": 38,
41
  "oral_lexical_repetition": 39,
42
  "oral_named_individual": 40,
43
- "oral_phatic_check": 41,
44
- "oral_phatic_filler": 42,
45
- "oral_rhetorical_question": 43,
46
- "oral_second_person": 44,
47
- "oral_self_correction": 45,
48
- "oral_sensory_detail": 46,
49
- "oral_simple_conjunction": 47,
50
- "oral_specific_place": 48,
51
- "oral_temporal_anchor": 49,
52
- "oral_tricolon": 50,
53
- "oral_vocative": 51
 
54
  }
 
40
  "oral_intensifier_doubling": 38,
41
  "oral_lexical_repetition": 39,
42
  "oral_named_individual": 40,
43
+ "oral_parallelism": 41,
44
+ "oral_phatic_check": 42,
45
+ "oral_phatic_filler": 43,
46
+ "oral_rhetorical_question": 44,
47
+ "oral_second_person": 45,
48
+ "oral_self_correction": 46,
49
+ "oral_sensory_detail": 47,
50
+ "oral_simple_conjunction": 48,
51
+ "oral_specific_place": 49,
52
+ "oral_temporal_anchor": 50,
53
+ "oral_tricolon": 51,
54
+ "oral_vocative": 52
55
  }