Training in progress, step 2500
Browse files
config.json
CHANGED
|
@@ -11,10 +11,10 @@
|
|
| 11 |
"layer_norm_epsilon": 1e-05,
|
| 12 |
"model_type": "gpt2",
|
| 13 |
"n_ctx": 1024,
|
| 14 |
-
"n_embd":
|
| 15 |
-
"n_head":
|
| 16 |
"n_inner": null,
|
| 17 |
-
"n_layer":
|
| 18 |
"n_positions": 1024,
|
| 19 |
"reorder_and_upcast_attn": false,
|
| 20 |
"resid_pdrop": 0.1,
|
|
|
|
| 11 |
"layer_norm_epsilon": 1e-05,
|
| 12 |
"model_type": "gpt2",
|
| 13 |
"n_ctx": 1024,
|
| 14 |
+
"n_embd": 1280,
|
| 15 |
+
"n_head": 20,
|
| 16 |
"n_inner": null,
|
| 17 |
+
"n_layer": 36,
|
| 18 |
"n_positions": 1024,
|
| 19 |
"reorder_and_upcast_attn": false,
|
| 20 |
"resid_pdrop": 0.1,
|
model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:84ed619d408eab16225bf5544805c57f71c4641475a66cab5fca85892e29b828
|
| 3 |
+
size 3096165928
|
runs/Jul10_16-19-34_c273f810eaf5/events.out.tfevents.1752164378.c273f810eaf5.851.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:96c71b8901c135aed11e438096589d4d40e213d0eb3f2ba5730ff2dbfabe4080
|
| 3 |
+
size 6305
|
training_args.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 5432
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:15587a0db87a68a2780844527c839fa0601790316b4c67087fcb386edebb8637
|
| 3 |
size 5432
|