teru00801 commited on
Commit
3769c68
·
verified ·
1 Parent(s): 69e5048

Upload folder using huggingface_hub

Browse files
20260305_170447/inference_bundle/config.json ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "DebertaV2ForMaskedLM"
4
+ ],
5
+ "attention_probs_dropout_prob": 0.1,
6
+ "bos_token_id": 96871,
7
+ "dtype": "float32",
8
+ "eos_token_id": 96868,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 3072,
14
+ "layer_norm_eps": 1e-07,
15
+ "legacy": true,
16
+ "max_position_embeddings": 512,
17
+ "max_relative_positions": -1,
18
+ "model_type": "deberta-v2",
19
+ "norm_rel_ebd": "layer_norm",
20
+ "num_attention_heads": 12,
21
+ "num_hidden_layers": 12,
22
+ "pad_token_id": 96869,
23
+ "pooler_dropout": 0,
24
+ "pooler_hidden_act": "gelu",
25
+ "pooler_hidden_size": 768,
26
+ "pos_att_type": [
27
+ "p2c",
28
+ "c2p"
29
+ ],
30
+ "position_biased_input": false,
31
+ "position_buckets": 256,
32
+ "relative_attention": true,
33
+ "share_att_key": true,
34
+ "tie_word_embeddings": true,
35
+ "transformers_version": "5.1.0",
36
+ "type_vocab_size": 0,
37
+ "vocab_size": 96873
38
+ }
20260305_170447/inference_bundle/export_report.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "run_dir": "outputs/task1_positive_binary_runs/20260305_170447",
3
+ "source_model_dir": "models/deberta_v3_mlm",
4
+ "best_epoch": 3,
5
+ "best_score": 0.864963503649635,
6
+ "val_metrics": {
7
+ "acc": 0.8603681833970128,
8
+ "f1": 0.8039024390243903,
9
+ "precision": 0.8538860103626943,
10
+ "recall": 0.759447004608295
11
+ },
12
+ "final_eval_metrics": {
13
+ "acc": 0.8603681833970128,
14
+ "f1": 0.8039024390243903,
15
+ "precision": 0.8538860103626943,
16
+ "recall": 0.759447004608295
17
+ },
18
+ "files": [
19
+ "config.json",
20
+ "export_report.json",
21
+ "inference_config.json",
22
+ "label_map.json",
23
+ "model.safetensors",
24
+ "special_tokens_map.json",
25
+ "tokenizer.json",
26
+ "tokenizer_config.json"
27
+ ]
28
+ }
20260305_170447/inference_bundle/inference_config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "model_type": "task1_deberta_binary",
3
+ "max_length": 384,
4
+ "title_empty_token": "[TITLE_EMPTY]",
5
+ "label_map_file": "label_map.json",
6
+ "weights_file": "model.safetensors"
7
+ }
20260305_170447/inference_bundle/label_map.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "Decline": 0,
3
+ "Pick": 1
4
+ }
20260305_170447/inference_bundle/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0823c204c77eefde383ba1dd017f60171cd5066f0531faae5064a6bf1f7a3810
3
+ size 639427360
20260305_170447/inference_bundle/special_tokens_map.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "[CLS]",
3
+ "eos_token": "[SEP]",
4
+ "unk_token": "[UNK]",
5
+ "sep_token": "[SEP]",
6
+ "pad_token": "[PAD]",
7
+ "cls_token": "[CLS]",
8
+ "mask_token": "[MASK]"
9
+ }
20260305_170447/inference_bundle/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
20260305_170447/inference_bundle/tokenizer_config.json ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": true,
3
+ "backend": "tokenizers",
4
+ "bos_token": "[CLS]",
5
+ "clean_up_tokenization_spaces": false,
6
+ "cls_token": "[CLS]",
7
+ "do_lower_case": false,
8
+ "eos_token": "[SEP]",
9
+ "extra_special_tokens": [
10
+ "[TITLE_EMPTY]"
11
+ ],
12
+ "is_local": true,
13
+ "keep_accents": true,
14
+ "mask_token": "[MASK]",
15
+ "model_max_length": 1000000000000000019884624838656,
16
+ "pad_token": "[PAD]",
17
+ "sep_token": "[SEP]",
18
+ "sp_model_kwargs": {},
19
+ "split_by_punct": false,
20
+ "tokenizer_class": "DebertaV2Tokenizer",
21
+ "unk_id": 1,
22
+ "unk_token": "[UNK]"
23
+ }