Sayan01 commited on
Commit
0e37ce6
·
1 Parent(s): 5122e9e

Training in progress, epoch 1

Browse files
logs/1656840351.2595713/events.out.tfevents.1656840351.6d65cfcd351c.80.31 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e1f5128111ca0b00177042d21d5f4c2c67fcc7f85b97a7e0841ba8ef974b461e
3
+ size 5347
logs/events.out.tfevents.1656839869.6d65cfcd351c.80.28 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:88a87bbeaf58078631be73a877d8d7cdd84ce176dfc330c98a16841157c2fd3f
3
- size 4210
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:43d3ceaf4ea5679a11b2b36d5d41b4a9597dec429af8ae9bace063fdf2958257
3
+ size 6121
logs/events.out.tfevents.1656840351.6d65cfcd351c.80.30 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ff8087cf4164e71d7ce659d2a39f65bf5c5acebdde8cadce28cbb0031165765f
3
+ size 4210
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9848216ea2558a8a522d97007b6e08215e0a772c54a71a926283f5176d460c50
3
  size 57429255
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:95059a299e1902b8112393ec476ac6228dbb95eb5e5a645fa49bffbd556a8f92
3
  size 57429255
tokenizer_config.json CHANGED
@@ -1,12 +1,13 @@
1
  {
2
  "cls_token": "[CLS]",
 
3
  "do_lower_case": true,
4
  "mask_token": "[MASK]",
5
- "model_max_length": 512,
6
- "name_or_path": "bert-base-uncased",
7
  "pad_token": "[PAD]",
8
  "sep_token": "[SEP]",
9
- "special_tokens_map_file": null,
10
  "strip_accents": null,
11
  "tokenize_chinese_chars": true,
12
  "tokenizer_class": "BertTokenizer",
 
1
  {
2
  "cls_token": "[CLS]",
3
+ "do_basic_tokenize": true,
4
  "do_lower_case": true,
5
  "mask_token": "[MASK]",
6
+ "name_or_path": "sentence-transformers/stsb-bert-base",
7
+ "never_split": null,
8
  "pad_token": "[PAD]",
9
  "sep_token": "[SEP]",
10
+ "special_tokens_map_file": "old_models/stsb-bert-base/0_BERT/special_tokens_map.json",
11
  "strip_accents": null,
12
  "tokenize_chinese_chars": true,
13
  "tokenizer_class": "BertTokenizer",