Upload folder using huggingface_hub
Browse files- checkpoint-1135/model.safetensors +1 -1
- checkpoint-1135/optimizer.pt +1 -1
- checkpoint-1135/trainer_state.json +30 -30
- model.safetensors +1 -1
checkpoint-1135/model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 343236280
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9c039f89d5a020f4aee173cb1bd811e28adaa8a47bf908e5373a426ce5acba03
|
| 3 |
size 343236280
|
checkpoint-1135/optimizer.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 686593029
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ebca17f410bc0de1b33ee988a830a8da5c5b3ec219150c98d372a8a530d4112a
|
| 3 |
size 686593029
|
checkpoint-1135/trainer_state.json
CHANGED
|
@@ -1,5 +1,5 @@
|
|
| 1 |
{
|
| 2 |
-
"best_metric": 0.
|
| 3 |
"best_model_checkpoint": "face_emotions_image_detection/checkpoint-1135",
|
| 4 |
"epoch": 5.0,
|
| 5 |
"eval_steps": 500,
|
|
@@ -10,61 +10,61 @@
|
|
| 10 |
"log_history": [
|
| 11 |
{
|
| 12 |
"epoch": 1.0,
|
| 13 |
-
"eval_accuracy": 0.
|
| 14 |
-
"eval_loss": 0.
|
| 15 |
-
"eval_runtime":
|
| 16 |
-
"eval_samples_per_second": 67.
|
| 17 |
-
"eval_steps_per_second": 2.
|
| 18 |
"step": 227
|
| 19 |
},
|
| 20 |
{
|
| 21 |
"epoch": 2.0,
|
| 22 |
-
"eval_accuracy": 0.
|
| 23 |
-
"eval_loss": 0.
|
| 24 |
-
"eval_runtime":
|
| 25 |
-
"eval_samples_per_second":
|
| 26 |
-
"eval_steps_per_second": 2.
|
| 27 |
"step": 454
|
| 28 |
},
|
| 29 |
{
|
| 30 |
"epoch": 2.2,
|
| 31 |
-
"grad_norm": 2.
|
| 32 |
"learning_rate": 1.7557603686635944e-07,
|
| 33 |
-
"loss": 0.
|
| 34 |
"step": 500
|
| 35 |
},
|
| 36 |
{
|
| 37 |
"epoch": 3.0,
|
| 38 |
-
"eval_accuracy": 0.
|
| 39 |
-
"eval_loss": 0.
|
| 40 |
-
"eval_runtime":
|
| 41 |
-
"eval_samples_per_second": 67.
|
| 42 |
-
"eval_steps_per_second": 2.
|
| 43 |
"step": 681
|
| 44 |
},
|
| 45 |
{
|
| 46 |
"epoch": 4.0,
|
| 47 |
-
"eval_accuracy": 0.
|
| 48 |
-
"eval_loss": 0.
|
| 49 |
-
"eval_runtime": 143.
|
| 50 |
-
"eval_samples_per_second": 67.
|
| 51 |
-
"eval_steps_per_second": 2.
|
| 52 |
"step": 908
|
| 53 |
},
|
| 54 |
{
|
| 55 |
"epoch": 4.41,
|
| 56 |
-
"grad_norm":
|
| 57 |
"learning_rate": 3.7327188940092163e-08,
|
| 58 |
-
"loss": 0.
|
| 59 |
"step": 1000
|
| 60 |
},
|
| 61 |
{
|
| 62 |
"epoch": 5.0,
|
| 63 |
-
"eval_accuracy": 0.
|
| 64 |
-
"eval_loss": 0.
|
| 65 |
-
"eval_runtime":
|
| 66 |
-
"eval_samples_per_second":
|
| 67 |
-
"eval_steps_per_second": 2.
|
| 68 |
"step": 1135
|
| 69 |
}
|
| 70 |
],
|
|
|
|
| 1 |
{
|
| 2 |
+
"best_metric": 0.718735933303833,
|
| 3 |
"best_model_checkpoint": "face_emotions_image_detection/checkpoint-1135",
|
| 4 |
"epoch": 5.0,
|
| 5 |
"eval_steps": 500,
|
|
|
|
| 10 |
"log_history": [
|
| 11 |
{
|
| 12 |
"epoch": 1.0,
|
| 13 |
+
"eval_accuracy": 0.778375581996896,
|
| 14 |
+
"eval_loss": 0.7267366051673889,
|
| 15 |
+
"eval_runtime": 143.7439,
|
| 16 |
+
"eval_samples_per_second": 67.238,
|
| 17 |
+
"eval_steps_per_second": 2.108,
|
| 18 |
"step": 227
|
| 19 |
},
|
| 20 |
{
|
| 21 |
"epoch": 2.0,
|
| 22 |
+
"eval_accuracy": 0.7781686497672012,
|
| 23 |
+
"eval_loss": 0.7245084643363953,
|
| 24 |
+
"eval_runtime": 143.5108,
|
| 25 |
+
"eval_samples_per_second": 67.347,
|
| 26 |
+
"eval_steps_per_second": 2.111,
|
| 27 |
"step": 454
|
| 28 |
},
|
| 29 |
{
|
| 30 |
"epoch": 2.2,
|
| 31 |
+
"grad_norm": 2.988050937652588,
|
| 32 |
"learning_rate": 1.7557603686635944e-07,
|
| 33 |
+
"loss": 0.6424,
|
| 34 |
"step": 500
|
| 35 |
},
|
| 36 |
{
|
| 37 |
"epoch": 3.0,
|
| 38 |
+
"eval_accuracy": 0.779720641489912,
|
| 39 |
+
"eval_loss": 0.7209171056747437,
|
| 40 |
+
"eval_runtime": 142.63,
|
| 41 |
+
"eval_samples_per_second": 67.763,
|
| 42 |
+
"eval_steps_per_second": 2.124,
|
| 43 |
"step": 681
|
| 44 |
},
|
| 45 |
{
|
| 46 |
"epoch": 4.0,
|
| 47 |
+
"eval_accuracy": 0.7789963786859804,
|
| 48 |
+
"eval_loss": 0.7192764282226562,
|
| 49 |
+
"eval_runtime": 143.1166,
|
| 50 |
+
"eval_samples_per_second": 67.532,
|
| 51 |
+
"eval_steps_per_second": 2.117,
|
| 52 |
"step": 908
|
| 53 |
},
|
| 54 |
{
|
| 55 |
"epoch": 4.41,
|
| 56 |
+
"grad_norm": 2.910348653793335,
|
| 57 |
"learning_rate": 3.7327188940092163e-08,
|
| 58 |
+
"loss": 0.6345,
|
| 59 |
"step": 1000
|
| 60 |
},
|
| 61 |
{
|
| 62 |
"epoch": 5.0,
|
| 63 |
+
"eval_accuracy": 0.7794102431453699,
|
| 64 |
+
"eval_loss": 0.718735933303833,
|
| 65 |
+
"eval_runtime": 142.01,
|
| 66 |
+
"eval_samples_per_second": 68.059,
|
| 67 |
+
"eval_steps_per_second": 2.134,
|
| 68 |
"step": 1135
|
| 69 |
}
|
| 70 |
],
|
model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 343236280
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9c039f89d5a020f4aee173cb1bd811e28adaa8a47bf908e5373a426ce5acba03
|
| 3 |
size 343236280
|