Training in progress, epoch 1
Browse files
logs/1656840351.2595713/events.out.tfevents.1656840351.6d65cfcd351c.80.31
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e1f5128111ca0b00177042d21d5f4c2c67fcc7f85b97a7e0841ba8ef974b461e
|
| 3 |
+
size 5347
|
logs/events.out.tfevents.1656839869.6d65cfcd351c.80.28
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:43d3ceaf4ea5679a11b2b36d5d41b4a9597dec429af8ae9bace063fdf2958257
|
| 3 |
+
size 6121
|
logs/events.out.tfevents.1656840351.6d65cfcd351c.80.30
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ff8087cf4164e71d7ce659d2a39f65bf5c5acebdde8cadce28cbb0031165765f
|
| 3 |
+
size 4210
|
pytorch_model.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 57429255
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:95059a299e1902b8112393ec476ac6228dbb95eb5e5a645fa49bffbd556a8f92
|
| 3 |
size 57429255
|
tokenizer_config.json
CHANGED
|
@@ -1,12 +1,13 @@
|
|
| 1 |
{
|
| 2 |
"cls_token": "[CLS]",
|
|
|
|
| 3 |
"do_lower_case": true,
|
| 4 |
"mask_token": "[MASK]",
|
| 5 |
-
"
|
| 6 |
-
"
|
| 7 |
"pad_token": "[PAD]",
|
| 8 |
"sep_token": "[SEP]",
|
| 9 |
-
"special_tokens_map_file":
|
| 10 |
"strip_accents": null,
|
| 11 |
"tokenize_chinese_chars": true,
|
| 12 |
"tokenizer_class": "BertTokenizer",
|
|
|
|
| 1 |
{
|
| 2 |
"cls_token": "[CLS]",
|
| 3 |
+
"do_basic_tokenize": true,
|
| 4 |
"do_lower_case": true,
|
| 5 |
"mask_token": "[MASK]",
|
| 6 |
+
"name_or_path": "sentence-transformers/stsb-bert-base",
|
| 7 |
+
"never_split": null,
|
| 8 |
"pad_token": "[PAD]",
|
| 9 |
"sep_token": "[SEP]",
|
| 10 |
+
"special_tokens_map_file": "old_models/stsb-bert-base/0_BERT/special_tokens_map.json",
|
| 11 |
"strip_accents": null,
|
| 12 |
"tokenize_chinese_chars": true,
|
| 13 |
"tokenizer_class": "BertTokenizer",
|