model_15M_medium_ds_masking_0.7_explicit_hs_predicted_hparamas
Browse files- README.md +28 -39
- config.json +1 -1
- model.safetensors +1 -1
- training_args.bin +1 -1
README.md
CHANGED
|
@@ -16,8 +16,8 @@ should probably proofread and complete it, then remove this comment. -->
|
|
| 16 |
|
| 17 |
This model is a fine-tuned version of [](https://huggingface.co/) on an unknown dataset.
|
| 18 |
It achieves the following results on the evaluation set:
|
| 19 |
-
- Loss: 0.
|
| 20 |
-
- Accuracy: 0.
|
| 21 |
|
| 22 |
## Model description
|
| 23 |
|
|
@@ -36,7 +36,7 @@ More information needed
|
|
| 36 |
### Training hyperparameters
|
| 37 |
|
| 38 |
The following hyperparameters were used during training:
|
| 39 |
-
- learning_rate: 0.
|
| 40 |
- train_batch_size: 256
|
| 41 |
- eval_batch_size: 256
|
| 42 |
- seed: 42
|
|
@@ -48,44 +48,33 @@ The following hyperparameters were used during training:
|
|
| 48 |
|
| 49 |
### Training results
|
| 50 |
|
| 51 |
-
| Training Loss | Epoch
|
| 52 |
-
|
| 53 |
-
| No log | 0
|
| 54 |
-
| 0.
|
| 55 |
-
| 0.
|
| 56 |
-
| 0.
|
| 57 |
-
| 0.
|
| 58 |
-
| 0.
|
| 59 |
-
| 0.
|
| 60 |
-
| 0.
|
| 61 |
-
| 0.
|
| 62 |
-
| 0.
|
| 63 |
-
| 0.
|
| 64 |
-
| 0.
|
| 65 |
-
| 0.
|
| 66 |
-
| 0.
|
| 67 |
-
| 0.
|
| 68 |
-
| 0.
|
| 69 |
-
| 0.
|
| 70 |
-
| 0.
|
| 71 |
-
| 0.
|
| 72 |
-
| 0.
|
| 73 |
-
| 0.3904 | 8.6035 | 39060 | 0.3827 | 0.8669 |
|
| 74 |
-
| 0.3873 | 9.0337 | 41013 | 0.3797 | 0.8680 |
|
| 75 |
-
| 0.3865 | 9.4639 | 42966 | 0.3775 | 0.8691 |
|
| 76 |
-
| 0.3839 | 9.8941 | 44919 | 0.3742 | 0.8696 |
|
| 77 |
-
| 0.3781 | 10.3242 | 46872 | 0.3757 | 0.8693 |
|
| 78 |
-
| 0.3806 | 10.7544 | 48825 | 0.3698 | 0.8713 |
|
| 79 |
-
| 0.376 | 11.1846 | 50778 | 0.3703 | 0.8712 |
|
| 80 |
-
| 0.3748 | 11.6148 | 52731 | 0.3672 | 0.8720 |
|
| 81 |
-
| 0.3743 | 12.0449 | 54684 | 0.3643 | 0.8733 |
|
| 82 |
-
| 0.3715 | 12.4751 | 56637 | 0.3662 | 0.8727 |
|
| 83 |
-
| 0.3695 | 12.9053 | 58590 | 0.3666 | 0.8724 |
|
| 84 |
|
| 85 |
|
| 86 |
### Framework versions
|
| 87 |
|
| 88 |
-
- Transformers 4.
|
| 89 |
-
- Pytorch 2.8.0.
|
| 90 |
-
- Datasets 3.
|
| 91 |
- Tokenizers 0.21.1
|
|
|
|
| 16 |
|
| 17 |
This model is a fine-tuned version of [](https://huggingface.co/) on an unknown dataset.
|
| 18 |
It achieves the following results on the evaluation set:
|
| 19 |
+
- Loss: 0.2317
|
| 20 |
+
- Accuracy: 0.9237
|
| 21 |
|
| 22 |
## Model description
|
| 23 |
|
|
|
|
| 36 |
### Training hyperparameters
|
| 37 |
|
| 38 |
The following hyperparameters were used during training:
|
| 39 |
+
- learning_rate: 0.005776
|
| 40 |
- train_batch_size: 256
|
| 41 |
- eval_batch_size: 256
|
| 42 |
- seed: 42
|
|
|
|
| 48 |
|
| 49 |
### Training results
|
| 50 |
|
| 51 |
+
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|
| 52 |
+
|:-------------:|:------:|:-----:|:---------------:|:--------:|
|
| 53 |
+
| No log | 0 | 0 | 4.4632 | 0.0019 |
|
| 54 |
+
| 0.4564 | 0.2190 | 1953 | 0.3973 | 0.8732 |
|
| 55 |
+
| 0.3637 | 0.4379 | 3906 | 0.3450 | 0.8886 |
|
| 56 |
+
| 0.3313 | 0.6569 | 5859 | 0.3169 | 0.8970 |
|
| 57 |
+
| 0.3101 | 0.8759 | 7812 | 0.2993 | 0.9026 |
|
| 58 |
+
| 0.3062 | 1.0949 | 9765 | 0.2885 | 0.9058 |
|
| 59 |
+
| 0.2852 | 1.3138 | 11718 | 0.2776 | 0.9092 |
|
| 60 |
+
| 0.2754 | 1.5328 | 13671 | 0.2703 | 0.9117 |
|
| 61 |
+
| 0.2689 | 1.7518 | 15624 | 0.2626 | 0.9142 |
|
| 62 |
+
| 0.2624 | 1.9707 | 17577 | 0.2566 | 0.9159 |
|
| 63 |
+
| 0.2574 | 2.1897 | 19530 | 0.2533 | 0.9169 |
|
| 64 |
+
| 0.2547 | 2.4087 | 21483 | 0.2478 | 0.9187 |
|
| 65 |
+
| 0.2495 | 2.6276 | 23436 | 0.2434 | 0.9201 |
|
| 66 |
+
| 0.2447 | 2.8466 | 25389 | 0.2395 | 0.9213 |
|
| 67 |
+
| 0.2416 | 3.0656 | 27342 | 0.2350 | 0.9227 |
|
| 68 |
+
| 0.2383 | 3.2846 | 29295 | 0.2331 | 0.9233 |
|
| 69 |
+
| 0.2363 | 3.5035 | 31248 | 0.2321 | 0.9236 |
|
| 70 |
+
| 0.0 | 3.7225 | 33201 | 0.2309 | 0.9240 |
|
| 71 |
+
| 0.0 | 3.9415 | 35154 | 0.2320 | 0.9237 |
|
| 72 |
+
| 0.0 | 4.1604 | 37107 | 0.2304 | 0.9241 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 73 |
|
| 74 |
|
| 75 |
### Framework versions
|
| 76 |
|
| 77 |
+
- Transformers 4.52.2
|
| 78 |
+
- Pytorch 2.8.0.dev20250521+cu128
|
| 79 |
+
- Datasets 3.6.0
|
| 80 |
- Tokenizers 0.21.1
|
config.json
CHANGED
|
@@ -37,6 +37,6 @@
|
|
| 37 |
"sparse_pred_ignore_index": -100,
|
| 38 |
"sparse_prediction": false,
|
| 39 |
"torch_dtype": "float32",
|
| 40 |
-
"transformers_version": "4.
|
| 41 |
"vocab_size": 82
|
| 42 |
}
|
|
|
|
| 37 |
"sparse_pred_ignore_index": -100,
|
| 38 |
"sparse_prediction": false,
|
| 39 |
"torch_dtype": "float32",
|
| 40 |
+
"transformers_version": "4.52.2",
|
| 41 |
"vocab_size": 82
|
| 42 |
}
|
model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 60925776
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:dcd833fc7ee5f7d40833d741dabd16a251eb928798344e8b76c9113827562a8a
|
| 3 |
size 60925776
|
training_args.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 5905
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3cc10ffd859f1450b05826ac813f49be0f0747e5248bcbc28c5728d5933007bf
|
| 3 |
size 5905
|