joseluhf11 commited on
Commit
6bd6192
·
1 Parent(s): 9108fbc

Upload 12 files

Browse files
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "bert-base-uncased",
3
  "architectures": [
4
  "BertModel"
5
  ],
 
1
  {
2
+ "_name_or_path": "/root/.cache/torch/sentence_transformers/joseluhf11_symptom_encoder_v2/",
3
  "architectures": [
4
  "BertModel"
5
  ],
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:54cf56e12308619d029556b44c9dac06c837c3e39dc3b76b8ea04af5ce6dbb63
3
  size 437995689
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9d8e16a4691418be2c0a4df18c6dc4c4fdf9b1b3ed4440cf7cd7edd0808294d7
3
  size 437995689
special_tokens_map.json CHANGED
@@ -1,4 +1,11 @@
1
  {
 
 
 
 
 
 
 
2
  "cls_token": "[CLS]",
3
  "mask_token": "[MASK]",
4
  "pad_token": "[PAD]",
 
1
  {
2
+ "additional_special_tokens": [
3
+ "[PAD]",
4
+ "[UNK]",
5
+ "[CLS]",
6
+ "[SEP]",
7
+ "[MASK]"
8
+ ],
9
  "cls_token": "[CLS]",
10
  "mask_token": "[MASK]",
11
  "pad_token": "[PAD]",
tokenizer_config.json CHANGED
@@ -41,16 +41,29 @@
41
  "special": true
42
  }
43
  },
44
- "additional_special_tokens": [],
 
 
 
 
 
 
45
  "clean_up_tokenization_spaces": true,
46
  "cls_token": "[CLS]",
47
  "do_lower_case": true,
48
  "mask_token": "[MASK]",
 
49
  "model_max_length": 512,
 
50
  "pad_token": "[PAD]",
 
 
51
  "sep_token": "[SEP]",
 
52
  "strip_accents": null,
53
  "tokenize_chinese_chars": true,
54
  "tokenizer_class": "BertTokenizer",
 
 
55
  "unk_token": "[UNK]"
56
  }
 
41
  "special": true
42
  }
43
  },
44
+ "additional_special_tokens": [
45
+ "[PAD]",
46
+ "[UNK]",
47
+ "[CLS]",
48
+ "[SEP]",
49
+ "[MASK]"
50
+ ],
51
  "clean_up_tokenization_spaces": true,
52
  "cls_token": "[CLS]",
53
  "do_lower_case": true,
54
  "mask_token": "[MASK]",
55
+ "max_length": 512,
56
  "model_max_length": 512,
57
+ "pad_to_multiple_of": null,
58
  "pad_token": "[PAD]",
59
+ "pad_token_type_id": 0,
60
+ "padding_side": "right",
61
  "sep_token": "[SEP]",
62
+ "stride": 0,
63
  "strip_accents": null,
64
  "tokenize_chinese_chars": true,
65
  "tokenizer_class": "BertTokenizer",
66
+ "truncation_side": "right",
67
+ "truncation_strategy": "longest_first",
68
  "unk_token": "[UNK]"
69
  }