gsl22 commited on
Commit
724eaae
·
1 Parent(s): 0f2cf00

Training in progress, epoch 1

Browse files
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:68a200f42fd87ac4c988adfe5efae1e3013b36ba014d8b9ab1cdb7948c0ed34f
3
  size 267883181
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a69cc0addaa785dfcc3f39f703dfc6bca91ac7b8f4d61964e57b4be5b267a94e
3
  size 267883181
runs/Oct02_14-39-43_be5a2759b431/events.out.tfevents.1696257664.be5a2759b431.340.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:97d5a37b898ed62725530e9d0d076cf2e98200e6fc1904cf48d0041abdcb4597
3
+ size 4653
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -1,13 +1,13 @@
1
  {
2
  "clean_up_tokenization_spaces": true,
3
  "cls_token": "[CLS]",
4
- "do_lower_case": false,
5
  "mask_token": "[MASK]",
6
  "model_max_length": 512,
7
  "pad_token": "[PAD]",
8
  "sep_token": "[SEP]",
9
  "strip_accents": null,
10
  "tokenize_chinese_chars": true,
11
- "tokenizer_class": "BertTokenizer",
12
  "unk_token": "[UNK]"
13
  }
 
1
  {
2
  "clean_up_tokenization_spaces": true,
3
  "cls_token": "[CLS]",
4
+ "do_lower_case": true,
5
  "mask_token": "[MASK]",
6
  "model_max_length": 512,
7
  "pad_token": "[PAD]",
8
  "sep_token": "[SEP]",
9
  "strip_accents": null,
10
  "tokenize_chinese_chars": true,
11
+ "tokenizer_class": "DistilBertTokenizer",
12
  "unk_token": "[UNK]"
13
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bd4c084730b5e5563c5f247cfa2a231d12f9e2d12e1cd8b970950db410a95c1b
3
  size 3963
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:78687938a28cf921358ad6f5e4e8392e84b3b89e67778e228d41816bdfd3d6d7
3
  size 3963
vocab.txt CHANGED
The diff for this file is too large to render. See raw diff