MutazYoune commited on
Commit
f12b6c4
·
1 Parent(s): d6b0be7

Upload 7 files

Browse files
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "/content/drive/MyDrive/Data_For_Bert/BERT_racism",
3
  "architectures": [
4
  "BertForMaskedLM"
5
  ],
 
1
  {
2
+ "_name_or_path": "/content/drive/MyDrive/Data_For_Bert/BERT_racism/checkpoint-50000",
3
  "architectures": [
4
  "BertForMaskedLM"
5
  ],
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
tokenizer_config.json ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "do_basic_tokenize": true,
4
+ "do_lower_case": true,
5
+ "mask_token": "[MASK]",
6
+ "model_max_length": 512,
7
+ "name_or_path": "/content/drive/MyDrive/Data_For_Bert/BERT_racism",
8
+ "never_split": null,
9
+ "pad_token": "[PAD]",
10
+ "sep_token": "[SEP]",
11
+ "special_tokens_map_file": null,
12
+ "strip_accents": null,
13
+ "tokenize_chinese_chars": true,
14
+ "tokenizer_class": "BertTokenizer",
15
+ "unk_token": "[UNK]"
16
+ }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:92d789e0c38f0bead545414681fd3192f4675d1ddd6a69ec78f7b802415e3275
3
- size 3643
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f4d53b5744def8c6041beb8c113c38b90d34f39043189e6b9fae97fbd99284d1
3
+ size 3515
vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
vocab.txt ADDED
The diff for this file is too large to render. See raw diff