jimboHsueh commited on
Commit
5336bcd
1 Parent(s): 442ad3e

End of training

Browse files
all_results.json CHANGED
@@ -1,4 +1,4 @@
1
  {
2
- "eval_exact_match": 77.8996344300432,
3
- "eval_f1": 77.8996344300432
4
  }
 
1
  {
2
+ "eval_exact_match": 78.92987703555998,
3
+ "eval_f1": 78.92987703555998
4
  }
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "cyclone/simcse-chinese-roberta-wwm-ext",
3
  "architectures": [
4
  "BertForQuestionAnswering"
5
  ],
@@ -8,19 +8,18 @@
8
  "classifier_dropout": null,
9
  "directionality": "bidi",
10
  "eos_token_id": 2,
11
- "gradient_checkpointing": false,
12
  "hidden_act": "gelu",
13
  "hidden_dropout_prob": 0.1,
14
- "hidden_size": 768,
15
  "initializer_range": 0.02,
16
- "intermediate_size": 3072,
17
  "layer_norm_eps": 1e-12,
18
  "max_position_embeddings": 512,
19
  "model_type": "bert",
20
- "num_attention_heads": 12,
21
- "num_hidden_layers": 12,
22
  "output_past": true,
23
- "pad_token_id": 1,
24
  "pooler_fc_size": 768,
25
  "pooler_num_attention_heads": 12,
26
  "pooler_num_fc_layers": 3,
 
1
  {
2
+ "_name_or_path": "hfl/chinese-roberta-wwm-ext-large",
3
  "architectures": [
4
  "BertForQuestionAnswering"
5
  ],
 
8
  "classifier_dropout": null,
9
  "directionality": "bidi",
10
  "eos_token_id": 2,
 
11
  "hidden_act": "gelu",
12
  "hidden_dropout_prob": 0.1,
13
+ "hidden_size": 1024,
14
  "initializer_range": 0.02,
15
+ "intermediate_size": 4096,
16
  "layer_norm_eps": 1e-12,
17
  "max_position_embeddings": 512,
18
  "model_type": "bert",
19
+ "num_attention_heads": 16,
20
+ "num_hidden_layers": 24,
21
  "output_past": true,
22
+ "pad_token_id": 0,
23
  "pooler_fc_size": 768,
24
  "pooler_num_attention_heads": 12,
25
  "pooler_num_fc_layers": 3,
eval_nbest_predictions.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6c16a6b1731bbb1b3a9c4b65e27f5895044521eb61db41691bcbb46e07054e26
3
- size 14332554
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:17f5d4c7aedd8688d51fea050aa3d206769f956c4fc79fc586ce63151da64112
3
+ size 14265158
eval_predictions.json CHANGED
The diff for this file is too large to render. See raw diff
 
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c2e1ffe6f83bb62cc52589411bfeab349b5c23fc57b5fa900fdce8d146ea06a2
3
- size 406782121
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:83376d941c0d8824fa13dec8e4f8efb571d462452919396ac2069d1bc0a4c0d6
3
+ size 1298033321
tokenizer_config.json CHANGED
@@ -46,18 +46,11 @@
46
  "cls_token": "[CLS]",
47
  "do_lower_case": true,
48
  "mask_token": "[MASK]",
49
- "max_length": 64,
50
  "model_max_length": 1000000000000000019884624838656,
51
- "pad_to_multiple_of": null,
52
  "pad_token": "[PAD]",
53
- "pad_token_type_id": 0,
54
- "padding_side": "right",
55
  "sep_token": "[SEP]",
56
- "stride": 0,
57
  "strip_accents": null,
58
  "tokenize_chinese_chars": true,
59
  "tokenizer_class": "BertTokenizer",
60
- "truncation_side": "right",
61
- "truncation_strategy": "longest_first",
62
  "unk_token": "[UNK]"
63
  }
 
46
  "cls_token": "[CLS]",
47
  "do_lower_case": true,
48
  "mask_token": "[MASK]",
 
49
  "model_max_length": 1000000000000000019884624838656,
 
50
  "pad_token": "[PAD]",
 
 
51
  "sep_token": "[SEP]",
 
52
  "strip_accents": null,
53
  "tokenize_chinese_chars": true,
54
  "tokenizer_class": "BertTokenizer",
 
 
55
  "unk_token": "[UNK]"
56
  }