Sayan01 commited on
Commit
b2c35f3
·
1 Parent(s): 7de64dc

Training in progress, epoch 1

Browse files
logs/1655874243.7970035/events.out.tfevents.1655874243.851641e31c58.72.5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8abb11e34a555327091ba44e0a4f88b680b97a70a9745abcfa6e568f1fbfa19e
3
+ size 5347
logs/events.out.tfevents.1655872621.851641e31c58.72.2 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b4115aa25c824bc91a7aef9c7f5212887a9c1378c63e7bd32ab6c4a5031a5938
3
- size 8476
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cbf649d928e473f1158068f33ce1bab388e360c277192117300a3d7037e52032
3
+ size 8830
logs/events.out.tfevents.1655874243.851641e31c58.72.4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cb9de08ae11fb87651142792ec10c23560a65f124d714aac796e5b9af1dc3c17
3
+ size 4156
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:64ef6f5a9848e770a4fd93ab1bf8b96a718ed39f0e67062c412082b922d13a3f
3
  size 38383463
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:83834275f3c045cf8aaf46797170de426ba910a0f869e3601ec3d324dd499384
3
  size 38383463
tokenizer_config.json CHANGED
@@ -1,14 +1,12 @@
1
  {
2
  "cls_token": "[CLS]",
3
- "do_basic_tokenize": true,
4
  "do_lower_case": true,
5
  "mask_token": "[MASK]",
6
  "model_max_length": 512,
7
- "name_or_path": "textattack/bert-base-uncased-SST-2",
8
- "never_split": null,
9
  "pad_token": "[PAD]",
10
  "sep_token": "[SEP]",
11
- "special_tokens_map_file": "/root/.cache/huggingface/transformers/c7c9b9c5d8bab3ba2ddaa08b138aa385f9790f30e8dce3bfe47e3f10bd97f4ad.dd8bd9bfd3664b530ea4e645105f557769387b3da9f79bdb55ed556bdd80611d",
12
  "strip_accents": null,
13
  "tokenize_chinese_chars": true,
14
  "tokenizer_class": "BertTokenizer",
 
1
  {
2
  "cls_token": "[CLS]",
 
3
  "do_lower_case": true,
4
  "mask_token": "[MASK]",
5
  "model_max_length": 512,
6
+ "name_or_path": "bert-base-uncased",
 
7
  "pad_token": "[PAD]",
8
  "sep_token": "[SEP]",
9
+ "special_tokens_map_file": null,
10
  "strip_accents": null,
11
  "tokenize_chinese_chars": true,
12
  "tokenizer_class": "BertTokenizer",