Sayan01 commited on
Commit
0530f61
·
1 Parent(s): ad3d038

Training in progress, epoch 1

Browse files
logs/1655835855.0405195/events.out.tfevents.1655835855.4dede0cb6010.74.39 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:53fd206ad4b5bc8347b9214f30b8c94d6e353965c34604085702c7095d6e586c
3
+ size 5343
logs/events.out.tfevents.1655834463.4dede0cb6010.74.36 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a650e0d2eee62b71afa77b71d396604f202f1b76ba635cb10eb96c51e816cf2c
3
- size 8610
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6fb0469aea3eced215f7f2895fb79fc9371e63f496ff2b20724b5e6a9b8e65bf
3
+ size 8964
logs/events.out.tfevents.1655835855.4dede0cb6010.74.38 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:390dcde4b1bac3f41f32ac7ad2d8221a4766c71359c42f88a5fd0aa427a790e8
3
+ size 4151
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:408d2927158829b519ed3906ae35bd561f3353bd0a1bb7f53914cec42596fdc3
3
  size 17561831
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:40e1d2ae2d6c02efe01a9b8ae965059f84295f5319554dedbb211e8ac048f84d
3
  size 17561831
tokenizer_config.json CHANGED
@@ -1,14 +1,12 @@
1
  {
2
  "cls_token": "[CLS]",
3
- "do_basic_tokenize": true,
4
  "do_lower_case": true,
5
  "mask_token": "[MASK]",
6
  "model_max_length": 512,
7
- "name_or_path": "textattack/bert-base-uncased-RTE",
8
- "never_split": null,
9
  "pad_token": "[PAD]",
10
  "sep_token": "[SEP]",
11
- "special_tokens_map_file": "/root/.cache/huggingface/transformers/a402e21b2325f5e1b9c95d97c746c5bdd55fd55558528bc4b52faa344ef64d5b.dd8bd9bfd3664b530ea4e645105f557769387b3da9f79bdb55ed556bdd80611d",
12
  "strip_accents": null,
13
  "tokenize_chinese_chars": true,
14
  "tokenizer_class": "BertTokenizer",
 
1
  {
2
  "cls_token": "[CLS]",
 
3
  "do_lower_case": true,
4
  "mask_token": "[MASK]",
5
  "model_max_length": 512,
6
+ "name_or_path": "bert-base-uncased",
 
7
  "pad_token": "[PAD]",
8
  "sep_token": "[SEP]",
9
+ "special_tokens_map_file": null,
10
  "strip_accents": null,
11
  "tokenize_chinese_chars": true,
12
  "tokenizer_class": "BertTokenizer",