Training in progress, epoch 2
Browse files- config.json +0 -1
- logs/events.out.tfevents.1727718130.spartan-gpgpu147.hpc.unimelb.edu.au.226463.0 +2 -2
- logs/events.out.tfevents.1727718386.spartan-gpgpu147.hpc.unimelb.edu.au.226463.1 +3 -0
- logs/events.out.tfevents.1727718438.spartan-gpgpu147.hpc.unimelb.edu.au.234866.0 +3 -0
- model.safetensors +1 -1
- tokenizer.json +2 -16
config.json
CHANGED
|
@@ -4,7 +4,6 @@
|
|
| 4 |
],
|
| 5 |
"attention_probs_dropout_prob": 0.1,
|
| 6 |
"classifier_dropout": null,
|
| 7 |
-
"finetuning_task": "glue:rte",
|
| 8 |
"gradient_checkpointing": false,
|
| 9 |
"hidden_act": "gelu",
|
| 10 |
"hidden_dropout_prob": 0.1,
|
|
|
|
| 4 |
],
|
| 5 |
"attention_probs_dropout_prob": 0.1,
|
| 6 |
"classifier_dropout": null,
|
|
|
|
| 7 |
"gradient_checkpointing": false,
|
| 8 |
"hidden_act": "gelu",
|
| 9 |
"hidden_dropout_prob": 0.1,
|
logs/events.out.tfevents.1727718130.spartan-gpgpu147.hpc.unimelb.edu.au.226463.0
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:14c55a05809c15323b633ee0ced7ffb05f58c8ff8847335988e099822993d491
|
| 3 |
+
size 11637
|
logs/events.out.tfevents.1727718386.spartan-gpgpu147.hpc.unimelb.edu.au.226463.1
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9e920b46fe64cdb206c1d6e0e8ace218b0e438e2eaed87eea898ef6ba2770532
|
| 3 |
+
size 477
|
logs/events.out.tfevents.1727718438.spartan-gpgpu147.hpc.unimelb.edu.au.234866.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1a6ba77d823ab5327467780ed075829547b71f9aedfcf2d526b08341c014e55b
|
| 3 |
+
size 6447
|
model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 90867952
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a2125a095e7bf7716ce78051ea4fe1b687980b59f7fac1ab17878f10453bcc5a
|
| 3 |
size 90867952
|
tokenizer.json
CHANGED
|
@@ -1,21 +1,7 @@
|
|
| 1 |
{
|
| 2 |
"version": "1.0",
|
| 3 |
-
"truncation":
|
| 4 |
-
|
| 5 |
-
"max_length": 512,
|
| 6 |
-
"strategy": "LongestFirst",
|
| 7 |
-
"stride": 0
|
| 8 |
-
},
|
| 9 |
-
"padding": {
|
| 10 |
-
"strategy": {
|
| 11 |
-
"Fixed": 512
|
| 12 |
-
},
|
| 13 |
-
"direction": "Right",
|
| 14 |
-
"pad_to_multiple_of": null,
|
| 15 |
-
"pad_id": 0,
|
| 16 |
-
"pad_type_id": 0,
|
| 17 |
-
"pad_token": "[PAD]"
|
| 18 |
-
},
|
| 19 |
"added_tokens": [
|
| 20 |
{
|
| 21 |
"id": 0,
|
|
|
|
| 1 |
{
|
| 2 |
"version": "1.0",
|
| 3 |
+
"truncation": null,
|
| 4 |
+
"padding": null,
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 5 |
"added_tokens": [
|
| 6 |
{
|
| 7 |
"id": 0,
|