Commit
·
406d54a
1
Parent(s):
987a753
feat: updated conll model
Browse files- conll2012_dataset_maker.py +120 -0
- models/o3-mini_20250218/README.md +0 -199
- models/o3-mini_20250218/added_tokens.json +0 -3
- models/o3-mini_20250218/config.json +0 -150
- models/o3-mini_20250218/model.safetensors +0 -3
- models/o3-mini_20250218/special_tokens_map.json +0 -15
- models/o3-mini_20250218/spm.model +0 -3
- models/o3-mini_20250218/tokenizer.json +0 -0
- models/o3-mini_20250218/tokenizer_config.json +0 -60
- models/o3-mini_20250218/training_args.bin +0 -3
- multi_head_trainer.py +1 -1
conll2012_dataset_maker.py
ADDED
|
@@ -0,0 +1,120 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
from datasets import load_dataset, DatasetDict
|
| 2 |
+
import argparse
|
| 3 |
+
import logging
|
| 4 |
+
|
| 5 |
+
from utils import default_logging_config, get_uniq_training_labels, show_examples
|
| 6 |
+
|
| 7 |
+
logger = logging.getLogger(__name__)
|
| 8 |
+
|
| 9 |
+
|
| 10 |
+
allowed_pos = {'``', '$', "''", ',', '-LRB-', '-RRB-', '.', ':', 'ADD', 'CC', 'CD', 'DT', 'EX',
|
| 11 |
+
'FW', 'HYPH', 'IN', 'JJ', 'JJR', 'JJS', 'LS', 'MD', 'NFP', 'NN', 'NNP', 'NNPS', 'NNS', 'PDT', 'POS',
|
| 12 |
+
'PRP', 'PRP$', 'RB', 'RBR', 'RBS', 'RP', 'SYM', 'TO', 'UH', 'VB', 'VBD', 'VBG', 'VBN', 'VBP', 'VBZ',
|
| 13 |
+
'WDT', 'WP', 'WP$', 'WRB'}
|
| 14 |
+
|
| 15 |
+
allowed_ner = {'O', 'B-PERSON', 'I-PERSON', 'B-NORP', 'I-NORP', 'B-FAC', 'I-FAC', 'B-ORG', 'I-ORG', 'B-GPE', 'I-GPE',
|
| 16 |
+
'B-LOC', 'I-LOC', 'B-PRODUCT', 'I-PRODUCT', 'B-DATE', 'I-DATE', 'B-TIME', 'I-TIME', 'B-PERCENT',
|
| 17 |
+
'I-PERCENT', 'B-MONEY', 'I-MONEY', 'B-QUANTITY', 'I-QUANTITY', 'B-ORDINAL', 'I-ORDINAL',
|
| 18 |
+
'B-CARDINAL', 'I-CARDINAL', 'B-EVENT', 'I-EVENT', 'B-WORK_OF_ART', 'I-WORK_OF_ART', 'B-LAW', 'I-LAW',
|
| 19 |
+
'B-LANGUAGE', 'I-LANGUAGE'}
|
| 20 |
+
|
| 21 |
+
|
| 22 |
+
def is_valid_example(exp):
|
| 23 |
+
"""
|
| 24 |
+
Simple filter that checks if all pos_tags are in allowed_pos
|
| 25 |
+
and all ner_tags are in allowed_ner. If you do not want any
|
| 26 |
+
filtering, simply return True.
|
| 27 |
+
"""
|
| 28 |
+
# You can skip filtering by just returning True:
|
| 29 |
+
# return True
|
| 30 |
+
|
| 31 |
+
# If your dataset has multiple tokens with possibly different tags,
|
| 32 |
+
# check them all:
|
| 33 |
+
for pos_tag in exp["pos_tags"]:
|
| 34 |
+
if pos_tag not in allowed_pos:
|
| 35 |
+
return False
|
| 36 |
+
|
| 37 |
+
for ner_tag in exp["ner_tags"]:
|
| 38 |
+
if ner_tag not in allowed_ner:
|
| 39 |
+
return False
|
| 40 |
+
|
| 41 |
+
return True
|
| 42 |
+
|
| 43 |
+
|
| 44 |
+
def transform_and_filter_dataset(onto_ds):
|
| 45 |
+
"""
|
| 46 |
+
onto_ds is a DatasetDict with splits: 'train', 'validation', 'test', etc.
|
| 47 |
+
Return a new DatasetDict with the same splits but:
|
| 48 |
+
- Filter out unwanted examples
|
| 49 |
+
- Possibly rename or remove columns
|
| 50 |
+
- Possibly introduce new columns
|
| 51 |
+
"""
|
| 52 |
+
pos_tag_int2str = onto_ds["train"].features["sentences"][0]["pos_tags"].feature.names
|
| 53 |
+
ner_tag_int2str = onto_ds["train"].features["sentences"][0]["named_entities"].feature.names
|
| 54 |
+
|
| 55 |
+
def flatten_ontonotes(batch):
|
| 56 |
+
out = {
|
| 57 |
+
"tokens": [],
|
| 58 |
+
"ner_tags": [],
|
| 59 |
+
"pos_tags": [],
|
| 60 |
+
"verb_predicate": [],
|
| 61 |
+
}
|
| 62 |
+
for doc_id, sents in zip(batch["document_id"], batch["sentences"]):
|
| 63 |
+
for sent_info in sents:
|
| 64 |
+
out["tokens"].append(sent_info["words"])
|
| 65 |
+
out["ner_tags"].append([ner_tag_int2str[i] for i in sent_info["named_entities"]])
|
| 66 |
+
out["pos_tags"].append([pos_tag_int2str[i] for i in sent_info["pos_tags"]])
|
| 67 |
+
out["verb_predicate"].append([("Yes" if s else "O") for s in sent_info["predicate_lemmas"]])
|
| 68 |
+
return out
|
| 69 |
+
|
| 70 |
+
new_splits = {}
|
| 71 |
+
for split_name, split_ds in onto_ds.items():
|
| 72 |
+
# Flatten
|
| 73 |
+
flattened_ds = split_ds.map(
|
| 74 |
+
flatten_ontonotes,
|
| 75 |
+
batched=True,
|
| 76 |
+
remove_columns=["sentences", "document_id"], # remove old columns
|
| 77 |
+
)
|
| 78 |
+
|
| 79 |
+
# Filter out invalid examples
|
| 80 |
+
filtered_split = flattened_ds.filter(is_valid_example)
|
| 81 |
+
new_splits[split_name] = filtered_split
|
| 82 |
+
|
| 83 |
+
return DatasetDict(new_splits)
|
| 84 |
+
|
| 85 |
+
|
| 86 |
+
# ------------------------------------------------------------------------------
|
| 87 |
+
# 6) Main Script
|
| 88 |
+
# ------------------------------------------------------------------------------
|
| 89 |
+
if __name__ == "__main__":
|
| 90 |
+
import logging.config
|
| 91 |
+
|
| 92 |
+
arg_parser = argparse.ArgumentParser(description="Process OntoNotes CoNLL-2012 (English).")
|
| 93 |
+
arg_parser.add_argument("--log-level", help="Log level.", action="store",
|
| 94 |
+
default="INFO", choices=["DEBUG", "INFO", "WARNING", "ERROR", "CRITICAL"])
|
| 95 |
+
arg_parser.add_argument("--save", help="Save final dataset to disk.", action="store_true", default=False)
|
| 96 |
+
arg_parser.add_argument("--save-path", help="Where to save final dataset.", default="./conll2012_en12_training_data")
|
| 97 |
+
arg_parser.add_argument("--show", help="Show examples: <split>/<col>/<label>/<count>", default=None)
|
| 98 |
+
args = arg_parser.parse_args()
|
| 99 |
+
|
| 100 |
+
logging.config.dictConfig(default_logging_config)
|
| 101 |
+
logger.setLevel(args.log_level)
|
| 102 |
+
|
| 103 |
+
# 6a) Load OntoNotes (English) from the 'conll2012_ontonotesv5' script
|
| 104 |
+
# This usually yields "train", "validation", "test" splits.
|
| 105 |
+
ontonotes_ds = load_dataset("conll2012_ontonotesv5", "english_v12")
|
| 106 |
+
logger.info(f"Splits loaded: {ontonotes_ds}")
|
| 107 |
+
|
| 108 |
+
# 6b) Transform & Filter
|
| 109 |
+
final_dataset = transform_and_filter_dataset(ontonotes_ds)
|
| 110 |
+
|
| 111 |
+
# 6d) Show examples if user requested
|
| 112 |
+
show_examples(final_dataset, args.show)
|
| 113 |
+
|
| 114 |
+
# 6e) Log unique training labels (POS/NER) if you like
|
| 115 |
+
get_uniq_training_labels(final_dataset)
|
| 116 |
+
|
| 117 |
+
# 6f) Save to disk if requested
|
| 118 |
+
if args.save:
|
| 119 |
+
final_dataset.save_to_disk(args.save_path)
|
| 120 |
+
logger.info("Saved dataset to %s", args.save_path)
|
models/o3-mini_20250218/README.md
DELETED
|
@@ -1,199 +0,0 @@
|
|
| 1 |
-
---
|
| 2 |
-
license: bsd-2-clause
|
| 3 |
-
---
|
| 4 |
-
|
| 5 |
-
### Dataset: o3-mini_20250218
|
| 6 |
-
```text
|
| 7 |
-
DatasetDict({
|
| 8 |
-
test: Dataset({
|
| 9 |
-
features: ['text', 'tokens', 'adj', 'adv', 'det', 'enc', 'func', 'misc', 'ner1', 'ner2', 'noun', 'pronoun', 'punct', 'verb', 'wh'],
|
| 10 |
-
num_rows: 2571
|
| 11 |
-
})
|
| 12 |
-
train: Dataset({
|
| 13 |
-
features: ['text', 'tokens', 'adj', 'adv', 'det', 'enc', 'func', 'misc', 'ner1', 'ner2', 'noun', 'pronoun', 'punct', 'verb', 'wh'],
|
| 14 |
-
num_rows: 23389
|
| 15 |
-
})
|
| 16 |
-
validation: Dataset({
|
| 17 |
-
features: ['text', 'tokens', 'adj', 'adv', 'det', 'enc', 'func', 'misc', 'ner1', 'ner2', 'noun', 'pronoun', 'punct', 'verb', 'wh'],
|
| 18 |
-
num_rows: 2599
|
| 19 |
-
})
|
| 20 |
-
})
|
| 21 |
-
```
|
| 22 |
-
|
| 23 |
-
|
| 24 |
-
### Classification Reports
|
| 25 |
-
```text
|
| 26 |
-
----- adj classification report -----
|
| 27 |
-
precision recall f1-score support
|
| 28 |
-
|
| 29 |
-
JJ 0.90 0.87 0.88 3187
|
| 30 |
-
JJR 0.95 0.88 0.91 162
|
| 31 |
-
JJS 0.88 0.84 0.86 102
|
| 32 |
-
O 0.99 0.99 0.99 29414
|
| 33 |
-
|
| 34 |
-
accuracy 0.98 32865
|
| 35 |
-
macro avg 0.93 0.89 0.91 32865
|
| 36 |
-
weighted avg 0.98 0.98 0.98 32865
|
| 37 |
-
|
| 38 |
-
----- adv classification report -----
|
| 39 |
-
precision recall f1-score support
|
| 40 |
-
|
| 41 |
-
O 0.99 0.99 0.99 30468
|
| 42 |
-
RB 0.91 0.91 0.91 2157
|
| 43 |
-
RBR 0.89 0.90 0.89 146
|
| 44 |
-
RBS 0.80 0.79 0.79 94
|
| 45 |
-
|
| 46 |
-
accuracy 0.99 32865
|
| 47 |
-
macro avg 0.90 0.90 0.90 32865
|
| 48 |
-
weighted avg 0.99 0.99 0.99 32865
|
| 49 |
-
|
| 50 |
-
----- det classification report -----
|
| 51 |
-
precision recall f1-score support
|
| 52 |
-
|
| 53 |
-
DT 0.96 0.95 0.96 4447
|
| 54 |
-
EX 0.96 0.90 0.93 82
|
| 55 |
-
O 0.99 0.99 0.99 28163
|
| 56 |
-
PDT 0.63 0.55 0.59 173
|
| 57 |
-
|
| 58 |
-
accuracy 0.99 32865
|
| 59 |
-
macro avg 0.89 0.85 0.87 32865
|
| 60 |
-
weighted avg 0.99 0.99 0.99 32865
|
| 61 |
-
|
| 62 |
-
----- enc classification report -----
|
| 63 |
-
precision recall f1-score support
|
| 64 |
-
|
| 65 |
-
BRACKET 0.79 0.89 0.84 385
|
| 66 |
-
O 0.99 0.99 0.99 31944
|
| 67 |
-
QUOTE 0.75 0.76 0.76 536
|
| 68 |
-
|
| 69 |
-
accuracy 0.99 32865
|
| 70 |
-
macro avg 0.85 0.88 0.86 32865
|
| 71 |
-
weighted avg 0.99 0.99 0.99 32865
|
| 72 |
-
|
| 73 |
-
----- func classification report -----
|
| 74 |
-
precision recall f1-score support
|
| 75 |
-
|
| 76 |
-
CC 0.98 0.99 0.98 1153
|
| 77 |
-
IN 0.97 0.98 0.97 3805
|
| 78 |
-
O 0.99 0.99 0.99 26444
|
| 79 |
-
RP 0.87 0.77 0.82 373
|
| 80 |
-
TO 1.00 0.99 0.99 871
|
| 81 |
-
UH 0.77 0.68 0.72 219
|
| 82 |
-
|
| 83 |
-
accuracy 0.99 32865
|
| 84 |
-
macro avg 0.93 0.90 0.91 32865
|
| 85 |
-
weighted avg 0.99 0.99 0.99 32865
|
| 86 |
-
|
| 87 |
-
----- misc classification report -----
|
| 88 |
-
precision recall f1-score support
|
| 89 |
-
|
| 90 |
-
$ 0.92 0.86 0.89 64
|
| 91 |
-
ADD 0.77 0.71 0.74 719
|
| 92 |
-
CD 0.89 0.89 0.89 558
|
| 93 |
-
EMOJI 1.00 0.73 0.85 15
|
| 94 |
-
O 0.99 0.99 0.99 30608
|
| 95 |
-
TIME 0.88 0.90 0.89 901
|
| 96 |
-
|
| 97 |
-
accuracy 0.98 32865
|
| 98 |
-
macro avg 0.91 0.85 0.87 32865
|
| 99 |
-
weighted avg 0.98 0.98 0.98 32865
|
| 100 |
-
|
| 101 |
-
----- ner1 classification report -----
|
| 102 |
-
precision recall f1-score support
|
| 103 |
-
|
| 104 |
-
B-GPE 0.87 0.90 0.89 473
|
| 105 |
-
B-ORG 0.86 0.82 0.84 424
|
| 106 |
-
B-PER 0.95 0.93 0.94 649
|
| 107 |
-
I-GPE 0.85 0.90 0.87 147
|
| 108 |
-
I-ORG 0.85 0.82 0.83 310
|
| 109 |
-
I-PER 0.96 0.96 0.96 261
|
| 110 |
-
O 0.99 0.99 0.99 30601
|
| 111 |
-
|
| 112 |
-
accuracy 0.99 32865
|
| 113 |
-
macro avg 0.90 0.90 0.90 32865
|
| 114 |
-
weighted avg 0.99 0.99 0.99 32865
|
| 115 |
-
|
| 116 |
-
----- ner2 classification report -----
|
| 117 |
-
precision recall f1-score support
|
| 118 |
-
|
| 119 |
-
B-EVENT 0.62 0.52 0.56 621
|
| 120 |
-
B-LOC 0.78 0.78 0.78 909
|
| 121 |
-
I-EVENT 0.54 0.32 0.40 1033
|
| 122 |
-
I-LOC 0.73 0.66 0.70 597
|
| 123 |
-
O 0.96 0.98 0.97 29705
|
| 124 |
-
|
| 125 |
-
accuracy 0.94 32865
|
| 126 |
-
macro avg 0.73 0.65 0.68 32865
|
| 127 |
-
weighted avg 0.93 0.94 0.93 32865
|
| 128 |
-
|
| 129 |
-
----- noun classification report -----
|
| 130 |
-
precision recall f1-score support
|
| 131 |
-
|
| 132 |
-
NN 0.96 0.96 0.96 4400
|
| 133 |
-
NNP 0.94 0.96 0.95 2410
|
| 134 |
-
NNPS 0.67 0.72 0.69 61
|
| 135 |
-
NNS 0.97 0.97 0.97 1698
|
| 136 |
-
O 0.99 0.99 0.99 24296
|
| 137 |
-
|
| 138 |
-
accuracy 0.98 32865
|
| 139 |
-
macro avg 0.91 0.92 0.91 32865
|
| 140 |
-
weighted avg 0.98 0.98 0.98 32865
|
| 141 |
-
|
| 142 |
-
----- pronoun classification report -----
|
| 143 |
-
precision recall f1-score support
|
| 144 |
-
|
| 145 |
-
O 1.00 1.00 1.00 29952
|
| 146 |
-
POS 0.97 0.97 0.97 154
|
| 147 |
-
PRP 0.97 0.97 0.97 2139
|
| 148 |
-
PRP$ 0.99 0.98 0.99 620
|
| 149 |
-
|
| 150 |
-
accuracy 1.00 32865
|
| 151 |
-
macro avg 0.98 0.98 0.98 32865
|
| 152 |
-
weighted avg 1.00 1.00 1.00 32865
|
| 153 |
-
|
| 154 |
-
----- punct classification report -----
|
| 155 |
-
precision recall f1-score support
|
| 156 |
-
|
| 157 |
-
COLON 0.99 0.95 0.97 201
|
| 158 |
-
COMMA 0.99 1.00 0.99 1454
|
| 159 |
-
EXCLAIM 0.99 0.97 0.98 107
|
| 160 |
-
HYPH 0.96 0.95 0.95 321
|
| 161 |
-
LS 0.57 0.53 0.55 15
|
| 162 |
-
O 1.00 1.00 1.00 28545
|
| 163 |
-
PERIOD 0.98 0.99 0.99 2022
|
| 164 |
-
QUESTION 0.99 0.99 0.99 156
|
| 165 |
-
SEP 0.75 0.41 0.53 44
|
| 166 |
-
|
| 167 |
-
accuracy 1.00 32865
|
| 168 |
-
macro avg 0.91 0.87 0.88 32865
|
| 169 |
-
weighted avg 1.00 1.00 1.00 32865
|
| 170 |
-
|
| 171 |
-
----- verb classification report -----
|
| 172 |
-
precision recall f1-score support
|
| 173 |
-
|
| 174 |
-
MD 1.00 0.98 0.99 527
|
| 175 |
-
O 1.00 0.99 0.99 26452
|
| 176 |
-
VB 0.95 0.94 0.94 1540
|
| 177 |
-
VBD 0.96 0.96 0.96 1330
|
| 178 |
-
VBG 0.94 0.96 0.95 625
|
| 179 |
-
VBN 0.88 0.93 0.90 766
|
| 180 |
-
VBP 0.88 0.92 0.90 766
|
| 181 |
-
VBZ 0.99 0.98 0.98 859
|
| 182 |
-
|
| 183 |
-
accuracy 0.98 32865
|
| 184 |
-
macro avg 0.95 0.96 0.95 32865
|
| 185 |
-
weighted avg 0.99 0.98 0.98 32865
|
| 186 |
-
|
| 187 |
-
----- wh classification report -----
|
| 188 |
-
precision recall f1-score support
|
| 189 |
-
|
| 190 |
-
O 0.99 1.00 0.99 32019
|
| 191 |
-
WDT 0.75 0.57 0.65 186
|
| 192 |
-
WP 0.84 0.71 0.77 164
|
| 193 |
-
WP$ 0.62 0.58 0.60 238
|
| 194 |
-
WRB 0.94 0.72 0.81 258
|
| 195 |
-
|
| 196 |
-
accuracy 0.99 32865
|
| 197 |
-
macro avg 0.83 0.72 0.77 32865
|
| 198 |
-
weighted avg 0.99 0.99 0.99 32865
|
| 199 |
-
```
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
models/o3-mini_20250218/added_tokens.json
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
{
|
| 2 |
-
"[MASK]": 128000
|
| 3 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
models/o3-mini_20250218/config.json
DELETED
|
@@ -1,150 +0,0 @@
|
|
| 1 |
-
{
|
| 2 |
-
"_name_or_path": "microsoft/deberta-v3-base",
|
| 3 |
-
"architectures": [
|
| 4 |
-
"MultiHeadModel"
|
| 5 |
-
],
|
| 6 |
-
"attention_probs_dropout_prob": 0.1,
|
| 7 |
-
"hidden_act": "gelu",
|
| 8 |
-
"hidden_dropout_prob": 0.1,
|
| 9 |
-
"hidden_size": 768,
|
| 10 |
-
"initializer_range": 0.02,
|
| 11 |
-
"intermediate_size": 3072,
|
| 12 |
-
"label_maps": {
|
| 13 |
-
"adj": [
|
| 14 |
-
"JJ",
|
| 15 |
-
"JJS",
|
| 16 |
-
"JJR",
|
| 17 |
-
"O"
|
| 18 |
-
],
|
| 19 |
-
"adv": [
|
| 20 |
-
"RBR",
|
| 21 |
-
"RB",
|
| 22 |
-
"RBS",
|
| 23 |
-
"O"
|
| 24 |
-
],
|
| 25 |
-
"det": [
|
| 26 |
-
"PDT",
|
| 27 |
-
"DT",
|
| 28 |
-
"EX",
|
| 29 |
-
"O"
|
| 30 |
-
],
|
| 31 |
-
"enc": [
|
| 32 |
-
"QUOTE",
|
| 33 |
-
"TICK",
|
| 34 |
-
"BRACKET",
|
| 35 |
-
"O"
|
| 36 |
-
],
|
| 37 |
-
"func": [
|
| 38 |
-
"UH",
|
| 39 |
-
"RP",
|
| 40 |
-
"TO",
|
| 41 |
-
"O",
|
| 42 |
-
"IN",
|
| 43 |
-
"CC"
|
| 44 |
-
],
|
| 45 |
-
"misc": [
|
| 46 |
-
"EMOJI",
|
| 47 |
-
"TIME",
|
| 48 |
-
"ADD",
|
| 49 |
-
"CD",
|
| 50 |
-
"O",
|
| 51 |
-
"$"
|
| 52 |
-
],
|
| 53 |
-
"ner1": [
|
| 54 |
-
"I-ORG",
|
| 55 |
-
"B-ORG",
|
| 56 |
-
"I-GPE",
|
| 57 |
-
"B-PER",
|
| 58 |
-
"O",
|
| 59 |
-
"B-GPE",
|
| 60 |
-
"I-PER"
|
| 61 |
-
],
|
| 62 |
-
"ner2": [
|
| 63 |
-
"I-LOC",
|
| 64 |
-
"B-LOC",
|
| 65 |
-
"I-EVENT",
|
| 66 |
-
"O",
|
| 67 |
-
"B-EVENT"
|
| 68 |
-
],
|
| 69 |
-
"noun": [
|
| 70 |
-
"NNS",
|
| 71 |
-
"O",
|
| 72 |
-
"NNP",
|
| 73 |
-
"NN",
|
| 74 |
-
"NNPS"
|
| 75 |
-
],
|
| 76 |
-
"pronoun": [
|
| 77 |
-
"PRP$",
|
| 78 |
-
"PRP",
|
| 79 |
-
"POS",
|
| 80 |
-
"O"
|
| 81 |
-
],
|
| 82 |
-
"punct": [
|
| 83 |
-
"QUESTION",
|
| 84 |
-
"LS",
|
| 85 |
-
"COMMA",
|
| 86 |
-
"EXCLAIM",
|
| 87 |
-
"COLON",
|
| 88 |
-
"PERIOD",
|
| 89 |
-
"SEP",
|
| 90 |
-
"O",
|
| 91 |
-
"HYPH"
|
| 92 |
-
],
|
| 93 |
-
"verb": [
|
| 94 |
-
"MD",
|
| 95 |
-
"VBG",
|
| 96 |
-
"O",
|
| 97 |
-
"VB",
|
| 98 |
-
"VBP",
|
| 99 |
-
"VBZ",
|
| 100 |
-
"VBN",
|
| 101 |
-
"VBD"
|
| 102 |
-
],
|
| 103 |
-
"wh": [
|
| 104 |
-
"WP$",
|
| 105 |
-
"O",
|
| 106 |
-
"WP",
|
| 107 |
-
"WRB",
|
| 108 |
-
"WDT"
|
| 109 |
-
]
|
| 110 |
-
},
|
| 111 |
-
"layer_norm_eps": 1e-07,
|
| 112 |
-
"legacy": true,
|
| 113 |
-
"max_position_embeddings": 512,
|
| 114 |
-
"max_relative_positions": -1,
|
| 115 |
-
"model_type": "deberta-v2",
|
| 116 |
-
"norm_rel_ebd": "layer_norm",
|
| 117 |
-
"num_attention_heads": 12,
|
| 118 |
-
"num_hidden_layers": 12,
|
| 119 |
-
"num_labels_dict": {
|
| 120 |
-
"adj": 4,
|
| 121 |
-
"adv": 4,
|
| 122 |
-
"det": 4,
|
| 123 |
-
"enc": 4,
|
| 124 |
-
"func": 6,
|
| 125 |
-
"misc": 6,
|
| 126 |
-
"ner1": 7,
|
| 127 |
-
"ner2": 5,
|
| 128 |
-
"noun": 5,
|
| 129 |
-
"pronoun": 4,
|
| 130 |
-
"punct": 9,
|
| 131 |
-
"verb": 8,
|
| 132 |
-
"wh": 5
|
| 133 |
-
},
|
| 134 |
-
"pad_token_id": 0,
|
| 135 |
-
"pooler_dropout": 0,
|
| 136 |
-
"pooler_hidden_act": "gelu",
|
| 137 |
-
"pooler_hidden_size": 768,
|
| 138 |
-
"pos_att_type": [
|
| 139 |
-
"p2c",
|
| 140 |
-
"c2p"
|
| 141 |
-
],
|
| 142 |
-
"position_biased_input": false,
|
| 143 |
-
"position_buckets": 256,
|
| 144 |
-
"relative_attention": true,
|
| 145 |
-
"share_att_key": true,
|
| 146 |
-
"torch_dtype": "float32",
|
| 147 |
-
"transformers_version": "4.48.2",
|
| 148 |
-
"type_vocab_size": 0,
|
| 149 |
-
"vocab_size": 128100
|
| 150 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
models/o3-mini_20250218/model.safetensors
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:d7fc80d3a8526faa41c3c79c97c87d72ca6f01fb6ef3812cd3a7764787b9949f
|
| 3 |
-
size 735571028
|
|
|
|
|
|
|
|
|
|
|
|
models/o3-mini_20250218/special_tokens_map.json
DELETED
|
@@ -1,15 +0,0 @@
|
|
| 1 |
-
{
|
| 2 |
-
"bos_token": "[CLS]",
|
| 3 |
-
"cls_token": "[CLS]",
|
| 4 |
-
"eos_token": "[SEP]",
|
| 5 |
-
"mask_token": "[MASK]",
|
| 6 |
-
"pad_token": "[PAD]",
|
| 7 |
-
"sep_token": "[SEP]",
|
| 8 |
-
"unk_token": {
|
| 9 |
-
"content": "[UNK]",
|
| 10 |
-
"lstrip": false,
|
| 11 |
-
"normalized": true,
|
| 12 |
-
"rstrip": false,
|
| 13 |
-
"single_word": false
|
| 14 |
-
}
|
| 15 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
models/o3-mini_20250218/spm.model
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:c679fbf93643d19aab7ee10c0b99e460bdbc02fedf34b92b05af343b4af586fd
|
| 3 |
-
size 2464616
|
|
|
|
|
|
|
|
|
|
|
|
models/o3-mini_20250218/tokenizer.json
DELETED
|
The diff for this file is too large to render.
See raw diff
|
|
|
models/o3-mini_20250218/tokenizer_config.json
DELETED
|
@@ -1,60 +0,0 @@
|
|
| 1 |
-
{
|
| 2 |
-
"add_prefix_space": true,
|
| 3 |
-
"added_tokens_decoder": {
|
| 4 |
-
"0": {
|
| 5 |
-
"content": "[PAD]",
|
| 6 |
-
"lstrip": false,
|
| 7 |
-
"normalized": false,
|
| 8 |
-
"rstrip": false,
|
| 9 |
-
"single_word": false,
|
| 10 |
-
"special": true
|
| 11 |
-
},
|
| 12 |
-
"1": {
|
| 13 |
-
"content": "[CLS]",
|
| 14 |
-
"lstrip": false,
|
| 15 |
-
"normalized": false,
|
| 16 |
-
"rstrip": false,
|
| 17 |
-
"single_word": false,
|
| 18 |
-
"special": true
|
| 19 |
-
},
|
| 20 |
-
"2": {
|
| 21 |
-
"content": "[SEP]",
|
| 22 |
-
"lstrip": false,
|
| 23 |
-
"normalized": false,
|
| 24 |
-
"rstrip": false,
|
| 25 |
-
"single_word": false,
|
| 26 |
-
"special": true
|
| 27 |
-
},
|
| 28 |
-
"3": {
|
| 29 |
-
"content": "[UNK]",
|
| 30 |
-
"lstrip": false,
|
| 31 |
-
"normalized": true,
|
| 32 |
-
"rstrip": false,
|
| 33 |
-
"single_word": false,
|
| 34 |
-
"special": true
|
| 35 |
-
},
|
| 36 |
-
"128000": {
|
| 37 |
-
"content": "[MASK]",
|
| 38 |
-
"lstrip": false,
|
| 39 |
-
"normalized": false,
|
| 40 |
-
"rstrip": false,
|
| 41 |
-
"single_word": false,
|
| 42 |
-
"special": true
|
| 43 |
-
}
|
| 44 |
-
},
|
| 45 |
-
"bos_token": "[CLS]",
|
| 46 |
-
"clean_up_tokenization_spaces": false,
|
| 47 |
-
"cls_token": "[CLS]",
|
| 48 |
-
"do_lower_case": false,
|
| 49 |
-
"eos_token": "[SEP]",
|
| 50 |
-
"extra_special_tokens": {},
|
| 51 |
-
"mask_token": "[MASK]",
|
| 52 |
-
"model_max_length": 1000000000000000019884624838656,
|
| 53 |
-
"pad_token": "[PAD]",
|
| 54 |
-
"sep_token": "[SEP]",
|
| 55 |
-
"sp_model_kwargs": {},
|
| 56 |
-
"split_by_punct": false,
|
| 57 |
-
"tokenizer_class": "DebertaV2Tokenizer",
|
| 58 |
-
"unk_token": "[UNK]",
|
| 59 |
-
"vocab_type": "spm"
|
| 60 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
models/o3-mini_20250218/training_args.bin
DELETED
|
@@ -1,3 +0,0 @@
|
|
| 1 |
-
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:393cbff7e2678a2b8c4e3190f5be4af291a4d8e9e2ca5376460939e460fa5ce5
|
| 3 |
-
size 5304
|
|
|
|
|
|
|
|
|
|
|
|
multi_head_trainer.py
CHANGED
|
@@ -305,7 +305,7 @@ if __name__ == "__main__":
|
|
| 305 |
arg_parser.add_argument("--mini", help='Train model using small subset of examples for pipeline testing.',
|
| 306 |
action="store_true", default=False)
|
| 307 |
arg_parser.add_argument("--save-path", help="Save final model to specified path.",
|
| 308 |
-
action="store", default="./
|
| 309 |
arg_parser.add_argument("--show", help="Show examples: <split>/<col>/<label>/<count>",
|
| 310 |
action="store", default=None)
|
| 311 |
arg_parser.add_argument("--train", help='Train model using loaded examples.',
|
|
|
|
| 305 |
arg_parser.add_argument("--mini", help='Train model using small subset of examples for pipeline testing.',
|
| 306 |
action="store_true", default=False)
|
| 307 |
arg_parser.add_argument("--save-path", help="Save final model to specified path.",
|
| 308 |
+
action="store", default="./final")
|
| 309 |
arg_parser.add_argument("--show", help="Show examples: <split>/<col>/<label>/<count>",
|
| 310 |
action="store", default=None)
|
| 311 |
arg_parser.add_argument("--train", help='Train model using loaded examples.',
|