Training in progress, epoch 1
Browse files
model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 267832560
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ffdcb176bcf3d1eb40a8e9c53abf70881ada7dcf18ee84ef0bb393abc13b82ea
|
| 3 |
size 267832560
|
runs/Nov04_20-07-35_e462e2c16979/events.out.tfevents.1730750882.e462e2c16979.1531.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7072f35fb908769516797c1ad953ff2d314c78e9e6fb801ff0ce23d3146c57a2
|
| 3 |
+
size 6162
|
tokenizer_config.json
CHANGED
|
@@ -50,6 +50,6 @@
|
|
| 50 |
"sep_token": "[SEP]",
|
| 51 |
"strip_accents": null,
|
| 52 |
"tokenize_chinese_chars": true,
|
| 53 |
-
"tokenizer_class": "
|
| 54 |
"unk_token": "[UNK]"
|
| 55 |
}
|
|
|
|
| 50 |
"sep_token": "[SEP]",
|
| 51 |
"strip_accents": null,
|
| 52 |
"tokenize_chinese_chars": true,
|
| 53 |
+
"tokenizer_class": "BertTokenizer",
|
| 54 |
"unk_token": "[UNK]"
|
| 55 |
}
|
training_args.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 5240
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ae7b3c9e13af7a8ab4b466b1e72c58ab7bdfaa97cc2afa581d8925216b9b0710
|
| 3 |
size 5240
|