Upload folder using huggingface_hub
Browse files- README.md +69 -69
- config.json +73 -67
- head_config.json +1 -1
- model.safetensors +2 -2
- type_to_idx.json +12 -11
README.md
CHANGED
|
@@ -22,7 +22,7 @@ datasets:
|
|
| 22 |
|
| 23 |
BERT-based token classifier for detecting **oral and literate markers** in text, based on Walter Ong's "Orality and Literacy" (1982).
|
| 24 |
|
| 25 |
-
This model performs multi-label span-level detection of
|
| 26 |
|
| 27 |
## Model Details
|
| 28 |
|
|
@@ -30,8 +30,8 @@ This model performs multi-label span-level detection of 52 rhetorical marker typ
|
|
| 30 |
|----------|-------|
|
| 31 |
| Base model | `bert-base-uncased` |
|
| 32 |
| Task | Multi-label token classification (independent B/I/O per type) |
|
| 33 |
-
| Marker types |
|
| 34 |
-
| Test macro F1 | **0.
|
| 35 |
| Training | 20 epochs, batch 24, lr 3e-5, fp16 |
|
| 36 |
| Regularization | Mixout (p=0.1) — stochastic L2 anchor to pretrained weights |
|
| 37 |
| Loss | Per-type weighted cross-entropy with inverse-frequency type weights |
|
|
@@ -80,16 +80,16 @@ for i, token in enumerate(tokens):
|
|
| 80 |
- Types with fewer than 150 annotated spans are excluded from training
|
| 81 |
- Multi-label BIO annotation: tokens can carry labels for multiple overlapping marker types simultaneously
|
| 82 |
|
| 83 |
-
## Marker Types (
|
| 84 |
|
| 85 |
-
### Oral Markers (
|
| 86 |
|
| 87 |
Characteristics of oral tradition and spoken discourse:
|
| 88 |
|
| 89 |
| Category | Markers |
|
| 90 |
|----------|---------|
|
| 91 |
| **Address & Interaction** | vocative, imperative, second_person, inclusive_we, rhetorical_question, phatic_check, phatic_filler |
|
| 92 |
-
| **Repetition & Pattern** | anaphora, tricolon, lexical_repetition, antithesis |
|
| 93 |
| **Conjunction** | simple_conjunction |
|
| 94 |
| **Formulas** | discourse_formula, intensifier_doubling |
|
| 95 |
| **Narrative** | named_individual, specific_place, temporal_anchor, sensory_detail, embodied_action, everyday_example |
|
|
@@ -115,75 +115,75 @@ Characteristics of written, analytical discourse:
|
|
| 115 |
Per-type detection F1 on test set (binary: B or I = positive, O = negative):
|
| 116 |
|
| 117 |
<details><summary>Click to show per-marker precision/recall/F1/support</summary>
|
| 118 |
-
|
| 119 |
```
|
| 120 |
Type Prec Rec F1 Sup
|
| 121 |
========================================================================
|
| 122 |
-
literate_abstract_noun 0.
|
| 123 |
-
literate_additive_formal 0.
|
| 124 |
-
literate_agent_demoted 0.
|
| 125 |
-
literate_agentless_passive 0.
|
| 126 |
-
literate_aside 0.
|
| 127 |
-
literate_categorical_statement 0.
|
| 128 |
-
literate_causal_explicit 0.
|
| 129 |
-
literate_citation 0.
|
| 130 |
-
literate_conceptual_metaphor 0.
|
| 131 |
-
literate_concessive 0.
|
| 132 |
-
literate_concessive_connector 0.
|
| 133 |
-
literate_concrete_setting 0.
|
| 134 |
-
literate_conditional 0.
|
| 135 |
-
literate_contrastive 0.
|
| 136 |
-
literate_cross_reference 0.
|
| 137 |
-
literate_definitional_move 0.
|
| 138 |
-
literate_enumeration 0.
|
| 139 |
-
literate_epistemic_hedge 0.
|
| 140 |
-
literate_evidential 0.
|
| 141 |
-
literate_institutional_subject 0.
|
| 142 |
-
literate_list_structure 0.
|
| 143 |
-
literate_metadiscourse 0.
|
| 144 |
-
literate_nested_clauses 0.
|
| 145 |
-
literate_nominalization 0.
|
| 146 |
-
literate_objectifying_stance 0.
|
| 147 |
-
literate_probability 0.
|
| 148 |
-
literate_qualified_assertion 0.
|
| 149 |
-
literate_relative_chain 0.
|
| 150 |
-
literate_technical_abbreviation 0.
|
| 151 |
-
literate_technical_term 0.
|
| 152 |
-
literate_temporal_embedding 0.
|
| 153 |
-
oral_anaphora 0.
|
| 154 |
-
oral_antithesis 0.
|
| 155 |
-
oral_discourse_formula 0.
|
| 156 |
-
oral_embodied_action 0.
|
| 157 |
-
oral_everyday_example 0.
|
| 158 |
-
oral_imperative 0.
|
| 159 |
-
oral_inclusive_we 0.
|
| 160 |
-
oral_intensifier_doubling 0.
|
| 161 |
-
oral_lexical_repetition 0.
|
| 162 |
-
oral_named_individual 0.
|
| 163 |
-
|
| 164 |
-
|
| 165 |
-
|
| 166 |
-
|
| 167 |
-
|
| 168 |
-
|
| 169 |
-
|
| 170 |
-
|
| 171 |
-
|
| 172 |
-
|
| 173 |
-
|
|
|
|
| 174 |
========================================================================
|
| 175 |
-
Macro avg (types w/ support) 0.
|
| 176 |
```
|
| 177 |
|
| 178 |
</details>
|
| 179 |
|
| 180 |
-
**Missing labels (test set):** 0/
|
| 181 |
|
| 182 |
Notable patterns:
|
| 183 |
-
- **Strong performers** (F1 > 0.5):
|
| 184 |
-
- **Weak performers** (F1 < 0.2):
|
| 185 |
-
- **Precision-recall tradeoff**: Most types now show higher
|
| 186 |
-
- **
|
| 187 |
|
| 188 |
## Architecture
|
| 189 |
|
|
@@ -215,9 +215,9 @@ classifier.bias → randomly initialized
|
|
| 215 |
|
| 216 |
## Limitations
|
| 217 |
|
| 218 |
-
- **
|
| 219 |
-
- **
|
| 220 |
-
- **
|
| 221 |
- **Context window**: 128 tokens max; longer spans may be truncated
|
| 222 |
- **Domain**: Trained primarily on historical/literary texts; may underperform on modern social media
|
| 223 |
- **Subjectivity**: Some marker boundaries are inherently ambiguous
|
|
@@ -239,4 +239,4 @@ classifier.bias → randomly initialized
|
|
| 239 |
|
| 240 |
---
|
| 241 |
|
| 242 |
-
*Trained: February 2026*
|
|
|
|
| 22 |
|
| 23 |
BERT-based token classifier for detecting **oral and literate markers** in text, based on Walter Ong's "Orality and Literacy" (1982).
|
| 24 |
|
| 25 |
+
This model performs multi-label span-level detection of 53 rhetorical marker types, where each token independently carries B/I/O labels per type — allowing overlapping spans (e.g. a token that is simultaneously part of a concessive and a nested clause).
|
| 26 |
|
| 27 |
## Model Details
|
| 28 |
|
|
|
|
| 30 |
|----------|-------|
|
| 31 |
| Base model | `bert-base-uncased` |
|
| 32 |
| Task | Multi-label token classification (independent B/I/O per type) |
|
| 33 |
+
| Marker types | 53 (22 oral, 31 literate) |
|
| 34 |
+
| Test macro F1 | **0.400** (per-type detection, binary positive = B or I) |
|
| 35 |
| Training | 20 epochs, batch 24, lr 3e-5, fp16 |
|
| 36 |
| Regularization | Mixout (p=0.1) — stochastic L2 anchor to pretrained weights |
|
| 37 |
| Loss | Per-type weighted cross-entropy with inverse-frequency type weights |
|
|
|
|
| 80 |
- Types with fewer than 150 annotated spans are excluded from training
|
| 81 |
- Multi-label BIO annotation: tokens can carry labels for multiple overlapping marker types simultaneously
|
| 82 |
|
| 83 |
+
## Marker Types (53)
|
| 84 |
|
| 85 |
+
### Oral Markers (22 types)
|
| 86 |
|
| 87 |
Characteristics of oral tradition and spoken discourse:
|
| 88 |
|
| 89 |
| Category | Markers |
|
| 90 |
|----------|---------|
|
| 91 |
| **Address & Interaction** | vocative, imperative, second_person, inclusive_we, rhetorical_question, phatic_check, phatic_filler |
|
| 92 |
+
| **Repetition & Pattern** | anaphora, parallelism, tricolon, lexical_repetition, antithesis |
|
| 93 |
| **Conjunction** | simple_conjunction |
|
| 94 |
| **Formulas** | discourse_formula, intensifier_doubling |
|
| 95 |
| **Narrative** | named_individual, specific_place, temporal_anchor, sensory_detail, embodied_action, everyday_example |
|
|
|
|
| 115 |
Per-type detection F1 on test set (binary: B or I = positive, O = negative):
|
| 116 |
|
| 117 |
<details><summary>Click to show per-marker precision/recall/F1/support</summary>
|
|
|
|
| 118 |
```
|
| 119 |
Type Prec Rec F1 Sup
|
| 120 |
========================================================================
|
| 121 |
+
literate_abstract_noun 0.211 0.319 0.254 464
|
| 122 |
+
literate_additive_formal 0.263 0.506 0.346 83
|
| 123 |
+
literate_agent_demoted 0.364 0.629 0.461 291
|
| 124 |
+
literate_agentless_passive 0.545 0.701 0.613 1274
|
| 125 |
+
literate_aside 0.396 0.565 0.466 467
|
| 126 |
+
literate_categorical_statement 0.246 0.245 0.245 388
|
| 127 |
+
literate_causal_explicit 0.325 0.305 0.315 370
|
| 128 |
+
literate_citation 0.500 0.551 0.524 243
|
| 129 |
+
literate_conceptual_metaphor 0.168 0.297 0.215 219
|
| 130 |
+
literate_concessive 0.542 0.491 0.515 731
|
| 131 |
+
literate_concessive_connector 0.113 0.378 0.174 37
|
| 132 |
+
literate_concrete_setting 0.174 0.279 0.214 301
|
| 133 |
+
literate_conditional 0.586 0.710 0.642 1610
|
| 134 |
+
literate_contrastive 0.374 0.343 0.358 382
|
| 135 |
+
literate_cross_reference 0.351 0.317 0.333 82
|
| 136 |
+
literate_definitional_move 0.217 0.371 0.274 62
|
| 137 |
+
literate_enumeration 0.456 0.570 0.507 899
|
| 138 |
+
literate_epistemic_hedge 0.415 0.511 0.458 534
|
| 139 |
+
literate_evidential 0.364 0.503 0.422 175
|
| 140 |
+
literate_institutional_subject 0.296 0.520 0.378 246
|
| 141 |
+
literate_list_structure 0.709 0.559 0.625 653
|
| 142 |
+
literate_metadiscourse 0.291 0.451 0.354 355
|
| 143 |
+
literate_nested_clauses 0.105 0.266 0.151 1250
|
| 144 |
+
literate_nominalization 0.475 0.554 0.511 1144
|
| 145 |
+
literate_objectifying_stance 0.518 0.448 0.481 194
|
| 146 |
+
literate_probability 0.612 0.548 0.578 115
|
| 147 |
+
literate_qualified_assertion 0.185 0.160 0.172 125
|
| 148 |
+
literate_relative_chain 0.320 0.537 0.401 1713
|
| 149 |
+
literate_technical_abbreviation 0.545 0.783 0.643 161
|
| 150 |
+
literate_technical_term 0.331 0.458 0.384 909
|
| 151 |
+
literate_temporal_embedding 0.222 0.249 0.235 570
|
| 152 |
+
oral_anaphora 0.207 0.248 0.226 137
|
| 153 |
+
oral_antithesis 0.245 0.289 0.265 453
|
| 154 |
+
oral_discourse_formula 0.353 0.384 0.368 563
|
| 155 |
+
oral_embodied_action 0.263 0.374 0.309 470
|
| 156 |
+
oral_everyday_example 0.160 0.164 0.162 366
|
| 157 |
+
oral_imperative 0.519 0.670 0.585 200
|
| 158 |
+
oral_inclusive_we 0.587 0.672 0.626 752
|
| 159 |
+
oral_intensifier_doubling 0.310 0.165 0.215 79
|
| 160 |
+
oral_lexical_repetition 0.293 0.488 0.366 217
|
| 161 |
+
oral_named_individual 0.428 0.676 0.524 791
|
| 162 |
+
oral_parallelism 0.654 0.048 0.089 710
|
| 163 |
+
oral_phatic_check 0.465 0.882 0.609 76
|
| 164 |
+
oral_phatic_filler 0.375 0.582 0.456 182
|
| 165 |
+
oral_rhetorical_question 0.589 0.894 0.710 1264
|
| 166 |
+
oral_second_person 0.614 0.545 0.577 833
|
| 167 |
+
oral_self_correction 0.597 0.295 0.395 156
|
| 168 |
+
oral_sensory_detail 0.275 0.312 0.293 352
|
| 169 |
+
oral_simple_conjunction 0.096 0.211 0.132 71
|
| 170 |
+
oral_specific_place 0.472 0.716 0.569 387
|
| 171 |
+
oral_temporal_anchor 0.397 0.748 0.518 551
|
| 172 |
+
oral_tricolon 0.274 0.285 0.280 557
|
| 173 |
+
oral_vocative 0.634 0.761 0.692 155
|
| 174 |
========================================================================
|
| 175 |
+
Macro avg (types w/ support) 0.400
|
| 176 |
```
|
| 177 |
|
| 178 |
</details>
|
| 179 |
|
| 180 |
+
**Missing labels (test set):** 0/53 — all types detected at least once.
|
| 181 |
|
| 182 |
Notable patterns:
|
| 183 |
+
- **Strong performers** (F1 > 0.5): rhetorical_question (0.710), vocative (0.692), conditional (0.642), technical_abbreviation (0.643), inclusive_we (0.626), list_structure (0.625), agentless_passive (0.613), phatic_check (0.609), imperative (0.585), probability (0.578), second_person (0.577), specific_place (0.569), citation (0.524), named_individual (0.524), temporal_anchor (0.518), concessive (0.515), nominalization (0.511), enumeration (0.507)
|
| 184 |
+
- **Weak performers** (F1 < 0.2): parallelism (0.089), simple_conjunction (0.132), nested_clauses (0.151), everyday_example (0.162), qualified_assertion (0.172), concessive_connector (0.174)
|
| 185 |
+
- **Precision-recall tradeoff**: Most types now show higher recall than precision, indicating the model over-predicts markers — reversed from the previous release. Notable exceptions include `parallelism` (0.654 precision / 0.048 recall), `self_correction`, and `intensifier_doubling`, which remain high-precision but low-recall.
|
| 186 |
+
- **Recovered type**: `oral_parallelism` crossed the 150-span threshold and was re-included, though its near-zero recall (0.048) means it is effectively non-functional despite high precision when it does fire.
|
| 187 |
|
| 188 |
## Architecture
|
| 189 |
|
|
|
|
| 215 |
|
| 216 |
## Limitations
|
| 217 |
|
| 218 |
+
- **Recall-dominated errors**: Most types over-predict (recall > precision), producing false positives; downstream applications may need confidence thresholding
|
| 219 |
+
- **Near-zero recall types**: `parallelism` (0.048 recall), `intensifier_doubling` (0.165), and `simple_conjunction` (0.211) are rarely detected despite being present in training data
|
| 220 |
+
- **Low-precision types**: `simple_conjunction` (0.096), `nested_clauses` (0.105), and `concessive_connector` (0.113) have precision below 0.15, meaning most predictions for those types are false positives
|
| 221 |
- **Context window**: 128 tokens max; longer spans may be truncated
|
| 222 |
- **Domain**: Trained primarily on historical/literary texts; may underperform on modern social media
|
| 223 |
- **Subjectivity**: Some marker boundaries are inherently ambiguous
|
|
|
|
| 239 |
|
| 240 |
---
|
| 241 |
|
| 242 |
+
*Trained: February 2026*
|
config.json
CHANGED
|
@@ -44,42 +44,45 @@
|
|
| 44 |
"120": "O-oral_named_individual",
|
| 45 |
"121": "B-oral_named_individual",
|
| 46 |
"122": "I-oral_named_individual",
|
| 47 |
-
"123": "O-
|
| 48 |
-
"124": "B-
|
| 49 |
-
"125": "I-
|
| 50 |
-
"126": "O-
|
| 51 |
-
"127": "B-
|
| 52 |
-
"128": "I-
|
| 53 |
-
"129": "O-
|
| 54 |
"13": "B-literate_aside",
|
| 55 |
-
"130": "B-
|
| 56 |
-
"131": "I-
|
| 57 |
-
"132": "O-
|
| 58 |
-
"133": "B-
|
| 59 |
-
"134": "I-
|
| 60 |
-
"135": "O-
|
| 61 |
-
"136": "B-
|
| 62 |
-
"137": "I-
|
| 63 |
-
"138": "O-
|
| 64 |
-
"139": "B-
|
| 65 |
"14": "I-literate_aside",
|
| 66 |
-
"140": "I-
|
| 67 |
-
"141": "O-
|
| 68 |
-
"142": "B-
|
| 69 |
-
"143": "I-
|
| 70 |
-
"144": "O-
|
| 71 |
-
"145": "B-
|
| 72 |
-
"146": "I-
|
| 73 |
-
"147": "O-
|
| 74 |
-
"148": "B-
|
| 75 |
-
"149": "I-
|
| 76 |
"15": "O-literate_categorical_statement",
|
| 77 |
-
"150": "O-
|
| 78 |
-
"151": "B-
|
| 79 |
-
"152": "I-
|
| 80 |
-
"153": "O-
|
| 81 |
-
"154": "B-
|
| 82 |
-
"155": "I-
|
|
|
|
|
|
|
|
|
|
| 83 |
"16": "B-literate_categorical_statement",
|
| 84 |
"17": "I-literate_categorical_statement",
|
| 85 |
"18": "O-literate_causal_explicit",
|
|
@@ -218,17 +221,18 @@
|
|
| 218 |
"B-oral_intensifier_doubling": 115,
|
| 219 |
"B-oral_lexical_repetition": 118,
|
| 220 |
"B-oral_named_individual": 121,
|
| 221 |
-
"B-
|
| 222 |
-
"B-
|
| 223 |
-
"B-
|
| 224 |
-
"B-
|
| 225 |
-
"B-
|
| 226 |
-
"B-
|
| 227 |
-
"B-
|
| 228 |
-
"B-
|
| 229 |
-
"B-
|
| 230 |
-
"B-
|
| 231 |
-
"B-
|
|
|
|
| 232 |
"I-literate_abstract_noun": 2,
|
| 233 |
"I-literate_additive_formal": 5,
|
| 234 |
"I-literate_agent_demoted": 8,
|
|
@@ -270,17 +274,18 @@
|
|
| 270 |
"I-oral_intensifier_doubling": 116,
|
| 271 |
"I-oral_lexical_repetition": 119,
|
| 272 |
"I-oral_named_individual": 122,
|
| 273 |
-
"I-
|
| 274 |
-
"I-
|
| 275 |
-
"I-
|
| 276 |
-
"I-
|
| 277 |
-
"I-
|
| 278 |
-
"I-
|
| 279 |
-
"I-
|
| 280 |
-
"I-
|
| 281 |
-
"I-
|
| 282 |
-
"I-
|
| 283 |
-
"I-
|
|
|
|
| 284 |
"O-literate_abstract_noun": 0,
|
| 285 |
"O-literate_additive_formal": 3,
|
| 286 |
"O-literate_agent_demoted": 6,
|
|
@@ -322,24 +327,25 @@
|
|
| 322 |
"O-oral_intensifier_doubling": 114,
|
| 323 |
"O-oral_lexical_repetition": 117,
|
| 324 |
"O-oral_named_individual": 120,
|
| 325 |
-
"O-
|
| 326 |
-
"O-
|
| 327 |
-
"O-
|
| 328 |
-
"O-
|
| 329 |
-
"O-
|
| 330 |
-
"O-
|
| 331 |
-
"O-
|
| 332 |
-
"O-
|
| 333 |
-
"O-
|
| 334 |
-
"O-
|
| 335 |
-
"O-
|
|
|
|
| 336 |
},
|
| 337 |
"layer_norm_eps": 1e-12,
|
| 338 |
"max_position_embeddings": 512,
|
| 339 |
"model_type": "bert",
|
| 340 |
"num_attention_heads": 12,
|
| 341 |
"num_hidden_layers": 12,
|
| 342 |
-
"num_types":
|
| 343 |
"pad_token_id": 0,
|
| 344 |
"position_embedding_type": "absolute",
|
| 345 |
"tie_word_embeddings": true,
|
|
|
|
| 44 |
"120": "O-oral_named_individual",
|
| 45 |
"121": "B-oral_named_individual",
|
| 46 |
"122": "I-oral_named_individual",
|
| 47 |
+
"123": "O-oral_parallelism",
|
| 48 |
+
"124": "B-oral_parallelism",
|
| 49 |
+
"125": "I-oral_parallelism",
|
| 50 |
+
"126": "O-oral_phatic_check",
|
| 51 |
+
"127": "B-oral_phatic_check",
|
| 52 |
+
"128": "I-oral_phatic_check",
|
| 53 |
+
"129": "O-oral_phatic_filler",
|
| 54 |
"13": "B-literate_aside",
|
| 55 |
+
"130": "B-oral_phatic_filler",
|
| 56 |
+
"131": "I-oral_phatic_filler",
|
| 57 |
+
"132": "O-oral_rhetorical_question",
|
| 58 |
+
"133": "B-oral_rhetorical_question",
|
| 59 |
+
"134": "I-oral_rhetorical_question",
|
| 60 |
+
"135": "O-oral_second_person",
|
| 61 |
+
"136": "B-oral_second_person",
|
| 62 |
+
"137": "I-oral_second_person",
|
| 63 |
+
"138": "O-oral_self_correction",
|
| 64 |
+
"139": "B-oral_self_correction",
|
| 65 |
"14": "I-literate_aside",
|
| 66 |
+
"140": "I-oral_self_correction",
|
| 67 |
+
"141": "O-oral_sensory_detail",
|
| 68 |
+
"142": "B-oral_sensory_detail",
|
| 69 |
+
"143": "I-oral_sensory_detail",
|
| 70 |
+
"144": "O-oral_simple_conjunction",
|
| 71 |
+
"145": "B-oral_simple_conjunction",
|
| 72 |
+
"146": "I-oral_simple_conjunction",
|
| 73 |
+
"147": "O-oral_specific_place",
|
| 74 |
+
"148": "B-oral_specific_place",
|
| 75 |
+
"149": "I-oral_specific_place",
|
| 76 |
"15": "O-literate_categorical_statement",
|
| 77 |
+
"150": "O-oral_temporal_anchor",
|
| 78 |
+
"151": "B-oral_temporal_anchor",
|
| 79 |
+
"152": "I-oral_temporal_anchor",
|
| 80 |
+
"153": "O-oral_tricolon",
|
| 81 |
+
"154": "B-oral_tricolon",
|
| 82 |
+
"155": "I-oral_tricolon",
|
| 83 |
+
"156": "O-oral_vocative",
|
| 84 |
+
"157": "B-oral_vocative",
|
| 85 |
+
"158": "I-oral_vocative",
|
| 86 |
"16": "B-literate_categorical_statement",
|
| 87 |
"17": "I-literate_categorical_statement",
|
| 88 |
"18": "O-literate_causal_explicit",
|
|
|
|
| 221 |
"B-oral_intensifier_doubling": 115,
|
| 222 |
"B-oral_lexical_repetition": 118,
|
| 223 |
"B-oral_named_individual": 121,
|
| 224 |
+
"B-oral_parallelism": 124,
|
| 225 |
+
"B-oral_phatic_check": 127,
|
| 226 |
+
"B-oral_phatic_filler": 130,
|
| 227 |
+
"B-oral_rhetorical_question": 133,
|
| 228 |
+
"B-oral_second_person": 136,
|
| 229 |
+
"B-oral_self_correction": 139,
|
| 230 |
+
"B-oral_sensory_detail": 142,
|
| 231 |
+
"B-oral_simple_conjunction": 145,
|
| 232 |
+
"B-oral_specific_place": 148,
|
| 233 |
+
"B-oral_temporal_anchor": 151,
|
| 234 |
+
"B-oral_tricolon": 154,
|
| 235 |
+
"B-oral_vocative": 157,
|
| 236 |
"I-literate_abstract_noun": 2,
|
| 237 |
"I-literate_additive_formal": 5,
|
| 238 |
"I-literate_agent_demoted": 8,
|
|
|
|
| 274 |
"I-oral_intensifier_doubling": 116,
|
| 275 |
"I-oral_lexical_repetition": 119,
|
| 276 |
"I-oral_named_individual": 122,
|
| 277 |
+
"I-oral_parallelism": 125,
|
| 278 |
+
"I-oral_phatic_check": 128,
|
| 279 |
+
"I-oral_phatic_filler": 131,
|
| 280 |
+
"I-oral_rhetorical_question": 134,
|
| 281 |
+
"I-oral_second_person": 137,
|
| 282 |
+
"I-oral_self_correction": 140,
|
| 283 |
+
"I-oral_sensory_detail": 143,
|
| 284 |
+
"I-oral_simple_conjunction": 146,
|
| 285 |
+
"I-oral_specific_place": 149,
|
| 286 |
+
"I-oral_temporal_anchor": 152,
|
| 287 |
+
"I-oral_tricolon": 155,
|
| 288 |
+
"I-oral_vocative": 158,
|
| 289 |
"O-literate_abstract_noun": 0,
|
| 290 |
"O-literate_additive_formal": 3,
|
| 291 |
"O-literate_agent_demoted": 6,
|
|
|
|
| 327 |
"O-oral_intensifier_doubling": 114,
|
| 328 |
"O-oral_lexical_repetition": 117,
|
| 329 |
"O-oral_named_individual": 120,
|
| 330 |
+
"O-oral_parallelism": 123,
|
| 331 |
+
"O-oral_phatic_check": 126,
|
| 332 |
+
"O-oral_phatic_filler": 129,
|
| 333 |
+
"O-oral_rhetorical_question": 132,
|
| 334 |
+
"O-oral_second_person": 135,
|
| 335 |
+
"O-oral_self_correction": 138,
|
| 336 |
+
"O-oral_sensory_detail": 141,
|
| 337 |
+
"O-oral_simple_conjunction": 144,
|
| 338 |
+
"O-oral_specific_place": 147,
|
| 339 |
+
"O-oral_temporal_anchor": 150,
|
| 340 |
+
"O-oral_tricolon": 153,
|
| 341 |
+
"O-oral_vocative": 156
|
| 342 |
},
|
| 343 |
"layer_norm_eps": 1e-12,
|
| 344 |
"max_position_embeddings": 512,
|
| 345 |
"model_type": "bert",
|
| 346 |
"num_attention_heads": 12,
|
| 347 |
"num_hidden_layers": 12,
|
| 348 |
+
"num_types": 53,
|
| 349 |
"pad_token_id": 0,
|
| 350 |
"position_embedding_type": "absolute",
|
| 351 |
"tie_word_embeddings": true,
|
head_config.json
CHANGED
|
@@ -1,5 +1,5 @@
|
|
| 1 |
{
|
| 2 |
"model_name": "bert-base-uncased",
|
| 3 |
-
"num_types":
|
| 4 |
"hidden_size": 768
|
| 5 |
}
|
|
|
|
| 1 |
{
|
| 2 |
"model_name": "bert-base-uncased",
|
| 3 |
+
"num_types": 53,
|
| 4 |
"hidden_size": 768
|
| 5 |
}
|
model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:41995685c78ead06fdda874b90a8bdf7b283997fa076207a33c0bd7136179ef3
|
| 3 |
+
size 436082548
|
type_to_idx.json
CHANGED
|
@@ -40,15 +40,16 @@
|
|
| 40 |
"oral_intensifier_doubling": 38,
|
| 41 |
"oral_lexical_repetition": 39,
|
| 42 |
"oral_named_individual": 40,
|
| 43 |
-
"
|
| 44 |
-
"
|
| 45 |
-
"
|
| 46 |
-
"
|
| 47 |
-
"
|
| 48 |
-
"
|
| 49 |
-
"
|
| 50 |
-
"
|
| 51 |
-
"
|
| 52 |
-
"
|
| 53 |
-
"
|
|
|
|
| 54 |
}
|
|
|
|
| 40 |
"oral_intensifier_doubling": 38,
|
| 41 |
"oral_lexical_repetition": 39,
|
| 42 |
"oral_named_individual": 40,
|
| 43 |
+
"oral_parallelism": 41,
|
| 44 |
+
"oral_phatic_check": 42,
|
| 45 |
+
"oral_phatic_filler": 43,
|
| 46 |
+
"oral_rhetorical_question": 44,
|
| 47 |
+
"oral_second_person": 45,
|
| 48 |
+
"oral_self_correction": 46,
|
| 49 |
+
"oral_sensory_detail": 47,
|
| 50 |
+
"oral_simple_conjunction": 48,
|
| 51 |
+
"oral_specific_place": 49,
|
| 52 |
+
"oral_temporal_anchor": 50,
|
| 53 |
+
"oral_tricolon": 51,
|
| 54 |
+
"oral_vocative": 52
|
| 55 |
}
|