dragosnicolae555 commited on
Commit
a4fe2dd
·
1 Parent(s): a324d2c

update model config

Browse files
Files changed (6) hide show
  1. added_tokens.json +1 -0
  2. albert_ +0 -0
  3. config.json +0 -4
  4. special_tokens_map.json +1 -0
  5. spiece.model +3 -0
  6. spiece.vocab +0 -0
added_tokens.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"[CLS]": 50000, "[SEP]": 50001, "<pad>": 50002, "[MASK]": 50003}
albert_ ADDED
File without changes
config.json CHANGED
@@ -1,7 +1,4 @@
1
  {
2
- "architectures": [
3
- "AlbertForMaskedLM"
4
- ],
5
  "attention_probs_dropout_prob": 0,
6
  "bos_token_id": 2,
7
  "classifier_dropout_prob": 0.1,
@@ -21,7 +18,6 @@
21
  "num_hidden_layers": 12,
22
  "pad_token_id": 0,
23
  "position_embedding_type": "absolute",
24
- "torch_dtype": "float32",
25
  "transformers_version": "4.10.2",
26
  "type_vocab_size": 2,
27
  "vocab_size": 50000
 
1
  {
 
 
 
2
  "attention_probs_dropout_prob": 0,
3
  "bos_token_id": 2,
4
  "classifier_dropout_prob": 0.1,
 
18
  "num_hidden_layers": 12,
19
  "pad_token_id": 0,
20
  "position_embedding_type": "absolute",
 
21
  "transformers_version": "4.10.2",
22
  "type_vocab_size": 2,
23
  "vocab_size": 50000
special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"bos_token": "[CLS]", "eos_token": "[SEP]", "unk_token": "<unk>", "sep_token": "[SEP]", "pad_token": "<pad>", "cls_token": "[CLS]", "mask_token": {"content": "[MASK]", "single_word": false, "lstrip": true, "rstrip": false, "normalized": true}}
spiece.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8c96b07a7eced4dbc93cdf74ed1de56322f118f41f4bae680c8762799139c8b5
3
+ size 1131184
spiece.vocab ADDED
The diff for this file is too large to render. See raw diff