veryfansome commited on
Commit
406d54a
·
1 Parent(s): 987a753

feat: updated conll model

Browse files
conll2012_dataset_maker.py ADDED
@@ -0,0 +1,120 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ from datasets import load_dataset, DatasetDict
2
+ import argparse
3
+ import logging
4
+
5
+ from utils import default_logging_config, get_uniq_training_labels, show_examples
6
+
7
+ logger = logging.getLogger(__name__)
8
+
9
+
10
+ allowed_pos = {'``', '$', "''", ',', '-LRB-', '-RRB-', '.', ':', 'ADD', 'CC', 'CD', 'DT', 'EX',
11
+ 'FW', 'HYPH', 'IN', 'JJ', 'JJR', 'JJS', 'LS', 'MD', 'NFP', 'NN', 'NNP', 'NNPS', 'NNS', 'PDT', 'POS',
12
+ 'PRP', 'PRP$', 'RB', 'RBR', 'RBS', 'RP', 'SYM', 'TO', 'UH', 'VB', 'VBD', 'VBG', 'VBN', 'VBP', 'VBZ',
13
+ 'WDT', 'WP', 'WP$', 'WRB'}
14
+
15
+ allowed_ner = {'O', 'B-PERSON', 'I-PERSON', 'B-NORP', 'I-NORP', 'B-FAC', 'I-FAC', 'B-ORG', 'I-ORG', 'B-GPE', 'I-GPE',
16
+ 'B-LOC', 'I-LOC', 'B-PRODUCT', 'I-PRODUCT', 'B-DATE', 'I-DATE', 'B-TIME', 'I-TIME', 'B-PERCENT',
17
+ 'I-PERCENT', 'B-MONEY', 'I-MONEY', 'B-QUANTITY', 'I-QUANTITY', 'B-ORDINAL', 'I-ORDINAL',
18
+ 'B-CARDINAL', 'I-CARDINAL', 'B-EVENT', 'I-EVENT', 'B-WORK_OF_ART', 'I-WORK_OF_ART', 'B-LAW', 'I-LAW',
19
+ 'B-LANGUAGE', 'I-LANGUAGE'}
20
+
21
+
22
+ def is_valid_example(exp):
23
+ """
24
+ Simple filter that checks if all pos_tags are in allowed_pos
25
+ and all ner_tags are in allowed_ner. If you do not want any
26
+ filtering, simply return True.
27
+ """
28
+ # You can skip filtering by just returning True:
29
+ # return True
30
+
31
+ # If your dataset has multiple tokens with possibly different tags,
32
+ # check them all:
33
+ for pos_tag in exp["pos_tags"]:
34
+ if pos_tag not in allowed_pos:
35
+ return False
36
+
37
+ for ner_tag in exp["ner_tags"]:
38
+ if ner_tag not in allowed_ner:
39
+ return False
40
+
41
+ return True
42
+
43
+
44
+ def transform_and_filter_dataset(onto_ds):
45
+ """
46
+ onto_ds is a DatasetDict with splits: 'train', 'validation', 'test', etc.
47
+ Return a new DatasetDict with the same splits but:
48
+ - Filter out unwanted examples
49
+ - Possibly rename or remove columns
50
+ - Possibly introduce new columns
51
+ """
52
+ pos_tag_int2str = onto_ds["train"].features["sentences"][0]["pos_tags"].feature.names
53
+ ner_tag_int2str = onto_ds["train"].features["sentences"][0]["named_entities"].feature.names
54
+
55
+ def flatten_ontonotes(batch):
56
+ out = {
57
+ "tokens": [],
58
+ "ner_tags": [],
59
+ "pos_tags": [],
60
+ "verb_predicate": [],
61
+ }
62
+ for doc_id, sents in zip(batch["document_id"], batch["sentences"]):
63
+ for sent_info in sents:
64
+ out["tokens"].append(sent_info["words"])
65
+ out["ner_tags"].append([ner_tag_int2str[i] for i in sent_info["named_entities"]])
66
+ out["pos_tags"].append([pos_tag_int2str[i] for i in sent_info["pos_tags"]])
67
+ out["verb_predicate"].append([("Yes" if s else "O") for s in sent_info["predicate_lemmas"]])
68
+ return out
69
+
70
+ new_splits = {}
71
+ for split_name, split_ds in onto_ds.items():
72
+ # Flatten
73
+ flattened_ds = split_ds.map(
74
+ flatten_ontonotes,
75
+ batched=True,
76
+ remove_columns=["sentences", "document_id"], # remove old columns
77
+ )
78
+
79
+ # Filter out invalid examples
80
+ filtered_split = flattened_ds.filter(is_valid_example)
81
+ new_splits[split_name] = filtered_split
82
+
83
+ return DatasetDict(new_splits)
84
+
85
+
86
+ # ------------------------------------------------------------------------------
87
+ # 6) Main Script
88
+ # ------------------------------------------------------------------------------
89
+ if __name__ == "__main__":
90
+ import logging.config
91
+
92
+ arg_parser = argparse.ArgumentParser(description="Process OntoNotes CoNLL-2012 (English).")
93
+ arg_parser.add_argument("--log-level", help="Log level.", action="store",
94
+ default="INFO", choices=["DEBUG", "INFO", "WARNING", "ERROR", "CRITICAL"])
95
+ arg_parser.add_argument("--save", help="Save final dataset to disk.", action="store_true", default=False)
96
+ arg_parser.add_argument("--save-path", help="Where to save final dataset.", default="./conll2012_en12_training_data")
97
+ arg_parser.add_argument("--show", help="Show examples: <split>/<col>/<label>/<count>", default=None)
98
+ args = arg_parser.parse_args()
99
+
100
+ logging.config.dictConfig(default_logging_config)
101
+ logger.setLevel(args.log_level)
102
+
103
+ # 6a) Load OntoNotes (English) from the 'conll2012_ontonotesv5' script
104
+ # This usually yields "train", "validation", "test" splits.
105
+ ontonotes_ds = load_dataset("conll2012_ontonotesv5", "english_v12")
106
+ logger.info(f"Splits loaded: {ontonotes_ds}")
107
+
108
+ # 6b) Transform & Filter
109
+ final_dataset = transform_and_filter_dataset(ontonotes_ds)
110
+
111
+ # 6d) Show examples if user requested
112
+ show_examples(final_dataset, args.show)
113
+
114
+ # 6e) Log unique training labels (POS/NER) if you like
115
+ get_uniq_training_labels(final_dataset)
116
+
117
+ # 6f) Save to disk if requested
118
+ if args.save:
119
+ final_dataset.save_to_disk(args.save_path)
120
+ logger.info("Saved dataset to %s", args.save_path)
models/o3-mini_20250218/README.md DELETED
@@ -1,199 +0,0 @@
1
- ---
2
- license: bsd-2-clause
3
- ---
4
-
5
- ### Dataset: o3-mini_20250218
6
- ```text
7
- DatasetDict({
8
- test: Dataset({
9
- features: ['text', 'tokens', 'adj', 'adv', 'det', 'enc', 'func', 'misc', 'ner1', 'ner2', 'noun', 'pronoun', 'punct', 'verb', 'wh'],
10
- num_rows: 2571
11
- })
12
- train: Dataset({
13
- features: ['text', 'tokens', 'adj', 'adv', 'det', 'enc', 'func', 'misc', 'ner1', 'ner2', 'noun', 'pronoun', 'punct', 'verb', 'wh'],
14
- num_rows: 23389
15
- })
16
- validation: Dataset({
17
- features: ['text', 'tokens', 'adj', 'adv', 'det', 'enc', 'func', 'misc', 'ner1', 'ner2', 'noun', 'pronoun', 'punct', 'verb', 'wh'],
18
- num_rows: 2599
19
- })
20
- })
21
- ```
22
-
23
-
24
- ### Classification Reports
25
- ```text
26
- ----- adj classification report -----
27
- precision recall f1-score support
28
-
29
- JJ 0.90 0.87 0.88 3187
30
- JJR 0.95 0.88 0.91 162
31
- JJS 0.88 0.84 0.86 102
32
- O 0.99 0.99 0.99 29414
33
-
34
- accuracy 0.98 32865
35
- macro avg 0.93 0.89 0.91 32865
36
- weighted avg 0.98 0.98 0.98 32865
37
-
38
- ----- adv classification report -----
39
- precision recall f1-score support
40
-
41
- O 0.99 0.99 0.99 30468
42
- RB 0.91 0.91 0.91 2157
43
- RBR 0.89 0.90 0.89 146
44
- RBS 0.80 0.79 0.79 94
45
-
46
- accuracy 0.99 32865
47
- macro avg 0.90 0.90 0.90 32865
48
- weighted avg 0.99 0.99 0.99 32865
49
-
50
- ----- det classification report -----
51
- precision recall f1-score support
52
-
53
- DT 0.96 0.95 0.96 4447
54
- EX 0.96 0.90 0.93 82
55
- O 0.99 0.99 0.99 28163
56
- PDT 0.63 0.55 0.59 173
57
-
58
- accuracy 0.99 32865
59
- macro avg 0.89 0.85 0.87 32865
60
- weighted avg 0.99 0.99 0.99 32865
61
-
62
- ----- enc classification report -----
63
- precision recall f1-score support
64
-
65
- BRACKET 0.79 0.89 0.84 385
66
- O 0.99 0.99 0.99 31944
67
- QUOTE 0.75 0.76 0.76 536
68
-
69
- accuracy 0.99 32865
70
- macro avg 0.85 0.88 0.86 32865
71
- weighted avg 0.99 0.99 0.99 32865
72
-
73
- ----- func classification report -----
74
- precision recall f1-score support
75
-
76
- CC 0.98 0.99 0.98 1153
77
- IN 0.97 0.98 0.97 3805
78
- O 0.99 0.99 0.99 26444
79
- RP 0.87 0.77 0.82 373
80
- TO 1.00 0.99 0.99 871
81
- UH 0.77 0.68 0.72 219
82
-
83
- accuracy 0.99 32865
84
- macro avg 0.93 0.90 0.91 32865
85
- weighted avg 0.99 0.99 0.99 32865
86
-
87
- ----- misc classification report -----
88
- precision recall f1-score support
89
-
90
- $ 0.92 0.86 0.89 64
91
- ADD 0.77 0.71 0.74 719
92
- CD 0.89 0.89 0.89 558
93
- EMOJI 1.00 0.73 0.85 15
94
- O 0.99 0.99 0.99 30608
95
- TIME 0.88 0.90 0.89 901
96
-
97
- accuracy 0.98 32865
98
- macro avg 0.91 0.85 0.87 32865
99
- weighted avg 0.98 0.98 0.98 32865
100
-
101
- ----- ner1 classification report -----
102
- precision recall f1-score support
103
-
104
- B-GPE 0.87 0.90 0.89 473
105
- B-ORG 0.86 0.82 0.84 424
106
- B-PER 0.95 0.93 0.94 649
107
- I-GPE 0.85 0.90 0.87 147
108
- I-ORG 0.85 0.82 0.83 310
109
- I-PER 0.96 0.96 0.96 261
110
- O 0.99 0.99 0.99 30601
111
-
112
- accuracy 0.99 32865
113
- macro avg 0.90 0.90 0.90 32865
114
- weighted avg 0.99 0.99 0.99 32865
115
-
116
- ----- ner2 classification report -----
117
- precision recall f1-score support
118
-
119
- B-EVENT 0.62 0.52 0.56 621
120
- B-LOC 0.78 0.78 0.78 909
121
- I-EVENT 0.54 0.32 0.40 1033
122
- I-LOC 0.73 0.66 0.70 597
123
- O 0.96 0.98 0.97 29705
124
-
125
- accuracy 0.94 32865
126
- macro avg 0.73 0.65 0.68 32865
127
- weighted avg 0.93 0.94 0.93 32865
128
-
129
- ----- noun classification report -----
130
- precision recall f1-score support
131
-
132
- NN 0.96 0.96 0.96 4400
133
- NNP 0.94 0.96 0.95 2410
134
- NNPS 0.67 0.72 0.69 61
135
- NNS 0.97 0.97 0.97 1698
136
- O 0.99 0.99 0.99 24296
137
-
138
- accuracy 0.98 32865
139
- macro avg 0.91 0.92 0.91 32865
140
- weighted avg 0.98 0.98 0.98 32865
141
-
142
- ----- pronoun classification report -----
143
- precision recall f1-score support
144
-
145
- O 1.00 1.00 1.00 29952
146
- POS 0.97 0.97 0.97 154
147
- PRP 0.97 0.97 0.97 2139
148
- PRP$ 0.99 0.98 0.99 620
149
-
150
- accuracy 1.00 32865
151
- macro avg 0.98 0.98 0.98 32865
152
- weighted avg 1.00 1.00 1.00 32865
153
-
154
- ----- punct classification report -----
155
- precision recall f1-score support
156
-
157
- COLON 0.99 0.95 0.97 201
158
- COMMA 0.99 1.00 0.99 1454
159
- EXCLAIM 0.99 0.97 0.98 107
160
- HYPH 0.96 0.95 0.95 321
161
- LS 0.57 0.53 0.55 15
162
- O 1.00 1.00 1.00 28545
163
- PERIOD 0.98 0.99 0.99 2022
164
- QUESTION 0.99 0.99 0.99 156
165
- SEP 0.75 0.41 0.53 44
166
-
167
- accuracy 1.00 32865
168
- macro avg 0.91 0.87 0.88 32865
169
- weighted avg 1.00 1.00 1.00 32865
170
-
171
- ----- verb classification report -----
172
- precision recall f1-score support
173
-
174
- MD 1.00 0.98 0.99 527
175
- O 1.00 0.99 0.99 26452
176
- VB 0.95 0.94 0.94 1540
177
- VBD 0.96 0.96 0.96 1330
178
- VBG 0.94 0.96 0.95 625
179
- VBN 0.88 0.93 0.90 766
180
- VBP 0.88 0.92 0.90 766
181
- VBZ 0.99 0.98 0.98 859
182
-
183
- accuracy 0.98 32865
184
- macro avg 0.95 0.96 0.95 32865
185
- weighted avg 0.99 0.98 0.98 32865
186
-
187
- ----- wh classification report -----
188
- precision recall f1-score support
189
-
190
- O 0.99 1.00 0.99 32019
191
- WDT 0.75 0.57 0.65 186
192
- WP 0.84 0.71 0.77 164
193
- WP$ 0.62 0.58 0.60 238
194
- WRB 0.94 0.72 0.81 258
195
-
196
- accuracy 0.99 32865
197
- macro avg 0.83 0.72 0.77 32865
198
- weighted avg 0.99 0.99 0.99 32865
199
- ```
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
models/o3-mini_20250218/added_tokens.json DELETED
@@ -1,3 +0,0 @@
1
- {
2
- "[MASK]": 128000
3
- }
 
 
 
 
models/o3-mini_20250218/config.json DELETED
@@ -1,150 +0,0 @@
1
- {
2
- "_name_or_path": "microsoft/deberta-v3-base",
3
- "architectures": [
4
- "MultiHeadModel"
5
- ],
6
- "attention_probs_dropout_prob": 0.1,
7
- "hidden_act": "gelu",
8
- "hidden_dropout_prob": 0.1,
9
- "hidden_size": 768,
10
- "initializer_range": 0.02,
11
- "intermediate_size": 3072,
12
- "label_maps": {
13
- "adj": [
14
- "JJ",
15
- "JJS",
16
- "JJR",
17
- "O"
18
- ],
19
- "adv": [
20
- "RBR",
21
- "RB",
22
- "RBS",
23
- "O"
24
- ],
25
- "det": [
26
- "PDT",
27
- "DT",
28
- "EX",
29
- "O"
30
- ],
31
- "enc": [
32
- "QUOTE",
33
- "TICK",
34
- "BRACKET",
35
- "O"
36
- ],
37
- "func": [
38
- "UH",
39
- "RP",
40
- "TO",
41
- "O",
42
- "IN",
43
- "CC"
44
- ],
45
- "misc": [
46
- "EMOJI",
47
- "TIME",
48
- "ADD",
49
- "CD",
50
- "O",
51
- "$"
52
- ],
53
- "ner1": [
54
- "I-ORG",
55
- "B-ORG",
56
- "I-GPE",
57
- "B-PER",
58
- "O",
59
- "B-GPE",
60
- "I-PER"
61
- ],
62
- "ner2": [
63
- "I-LOC",
64
- "B-LOC",
65
- "I-EVENT",
66
- "O",
67
- "B-EVENT"
68
- ],
69
- "noun": [
70
- "NNS",
71
- "O",
72
- "NNP",
73
- "NN",
74
- "NNPS"
75
- ],
76
- "pronoun": [
77
- "PRP$",
78
- "PRP",
79
- "POS",
80
- "O"
81
- ],
82
- "punct": [
83
- "QUESTION",
84
- "LS",
85
- "COMMA",
86
- "EXCLAIM",
87
- "COLON",
88
- "PERIOD",
89
- "SEP",
90
- "O",
91
- "HYPH"
92
- ],
93
- "verb": [
94
- "MD",
95
- "VBG",
96
- "O",
97
- "VB",
98
- "VBP",
99
- "VBZ",
100
- "VBN",
101
- "VBD"
102
- ],
103
- "wh": [
104
- "WP$",
105
- "O",
106
- "WP",
107
- "WRB",
108
- "WDT"
109
- ]
110
- },
111
- "layer_norm_eps": 1e-07,
112
- "legacy": true,
113
- "max_position_embeddings": 512,
114
- "max_relative_positions": -1,
115
- "model_type": "deberta-v2",
116
- "norm_rel_ebd": "layer_norm",
117
- "num_attention_heads": 12,
118
- "num_hidden_layers": 12,
119
- "num_labels_dict": {
120
- "adj": 4,
121
- "adv": 4,
122
- "det": 4,
123
- "enc": 4,
124
- "func": 6,
125
- "misc": 6,
126
- "ner1": 7,
127
- "ner2": 5,
128
- "noun": 5,
129
- "pronoun": 4,
130
- "punct": 9,
131
- "verb": 8,
132
- "wh": 5
133
- },
134
- "pad_token_id": 0,
135
- "pooler_dropout": 0,
136
- "pooler_hidden_act": "gelu",
137
- "pooler_hidden_size": 768,
138
- "pos_att_type": [
139
- "p2c",
140
- "c2p"
141
- ],
142
- "position_biased_input": false,
143
- "position_buckets": 256,
144
- "relative_attention": true,
145
- "share_att_key": true,
146
- "torch_dtype": "float32",
147
- "transformers_version": "4.48.2",
148
- "type_vocab_size": 0,
149
- "vocab_size": 128100
150
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
models/o3-mini_20250218/model.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:d7fc80d3a8526faa41c3c79c97c87d72ca6f01fb6ef3812cd3a7764787b9949f
3
- size 735571028
 
 
 
 
models/o3-mini_20250218/special_tokens_map.json DELETED
@@ -1,15 +0,0 @@
1
- {
2
- "bos_token": "[CLS]",
3
- "cls_token": "[CLS]",
4
- "eos_token": "[SEP]",
5
- "mask_token": "[MASK]",
6
- "pad_token": "[PAD]",
7
- "sep_token": "[SEP]",
8
- "unk_token": {
9
- "content": "[UNK]",
10
- "lstrip": false,
11
- "normalized": true,
12
- "rstrip": false,
13
- "single_word": false
14
- }
15
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
models/o3-mini_20250218/spm.model DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:c679fbf93643d19aab7ee10c0b99e460bdbc02fedf34b92b05af343b4af586fd
3
- size 2464616
 
 
 
 
models/o3-mini_20250218/tokenizer.json DELETED
The diff for this file is too large to render. See raw diff
 
models/o3-mini_20250218/tokenizer_config.json DELETED
@@ -1,60 +0,0 @@
1
- {
2
- "add_prefix_space": true,
3
- "added_tokens_decoder": {
4
- "0": {
5
- "content": "[PAD]",
6
- "lstrip": false,
7
- "normalized": false,
8
- "rstrip": false,
9
- "single_word": false,
10
- "special": true
11
- },
12
- "1": {
13
- "content": "[CLS]",
14
- "lstrip": false,
15
- "normalized": false,
16
- "rstrip": false,
17
- "single_word": false,
18
- "special": true
19
- },
20
- "2": {
21
- "content": "[SEP]",
22
- "lstrip": false,
23
- "normalized": false,
24
- "rstrip": false,
25
- "single_word": false,
26
- "special": true
27
- },
28
- "3": {
29
- "content": "[UNK]",
30
- "lstrip": false,
31
- "normalized": true,
32
- "rstrip": false,
33
- "single_word": false,
34
- "special": true
35
- },
36
- "128000": {
37
- "content": "[MASK]",
38
- "lstrip": false,
39
- "normalized": false,
40
- "rstrip": false,
41
- "single_word": false,
42
- "special": true
43
- }
44
- },
45
- "bos_token": "[CLS]",
46
- "clean_up_tokenization_spaces": false,
47
- "cls_token": "[CLS]",
48
- "do_lower_case": false,
49
- "eos_token": "[SEP]",
50
- "extra_special_tokens": {},
51
- "mask_token": "[MASK]",
52
- "model_max_length": 1000000000000000019884624838656,
53
- "pad_token": "[PAD]",
54
- "sep_token": "[SEP]",
55
- "sp_model_kwargs": {},
56
- "split_by_punct": false,
57
- "tokenizer_class": "DebertaV2Tokenizer",
58
- "unk_token": "[UNK]",
59
- "vocab_type": "spm"
60
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
models/o3-mini_20250218/training_args.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:393cbff7e2678a2b8c4e3190f5be4af291a4d8e9e2ca5376460939e460fa5ce5
3
- size 5304
 
 
 
 
multi_head_trainer.py CHANGED
@@ -305,7 +305,7 @@ if __name__ == "__main__":
305
  arg_parser.add_argument("--mini", help='Train model using small subset of examples for pipeline testing.',
306
  action="store_true", default=False)
307
  arg_parser.add_argument("--save-path", help="Save final model to specified path.",
308
- action="store", default="./ud_final")
309
  arg_parser.add_argument("--show", help="Show examples: <split>/<col>/<label>/<count>",
310
  action="store", default=None)
311
  arg_parser.add_argument("--train", help='Train model using loaded examples.',
 
305
  arg_parser.add_argument("--mini", help='Train model using small subset of examples for pipeline testing.',
306
  action="store_true", default=False)
307
  arg_parser.add_argument("--save-path", help="Save final model to specified path.",
308
+ action="store", default="./final")
309
  arg_parser.add_argument("--show", help="Show examples: <split>/<col>/<label>/<count>",
310
  action="store", default=None)
311
  arg_parser.add_argument("--train", help='Train model using loaded examples.',