Upload 70 files
Browse filesThis view is limited to 50 files because it contains too many changes. See raw diff
- distilbert-base-cased-ft-BUILD/checkpoint-3120/config.json +36 -0
- distilbert-base-cased-ft-BUILD/checkpoint-3120/model.safetensors +3 -0
- distilbert-base-cased-ft-BUILD/checkpoint-3120/optimizer.pt +3 -0
- distilbert-base-cased-ft-BUILD/checkpoint-3120/rng_state.pth +3 -0
- distilbert-base-cased-ft-BUILD/checkpoint-3120/scheduler.pt +3 -0
- distilbert-base-cased-ft-BUILD/checkpoint-3120/trainer_state.json +166 -0
- distilbert-base-cased-ft-BUILD/checkpoint-3120/training_args.bin +3 -0
- distilbert-base-cased-ft-BUILD/distilbert-base-cased-ft-BUILD-best/config.json +36 -0
- distilbert-base-cased-ft-BUILD/distilbert-base-cased-ft-BUILD-best/model.safetensors +3 -0
- distilbert-base-cased-ft-BUILD/distilbert-base-cased-ft-BUILD-best/special_tokens_map.json +7 -0
- distilbert-base-cased-ft-BUILD/distilbert-base-cased-ft-BUILD-best/tokenizer.json +0 -0
- distilbert-base-cased-ft-BUILD/distilbert-base-cased-ft-BUILD-best/tokenizer_config.json +55 -0
- distilbert-base-cased-ft-BUILD/distilbert-base-cased-ft-BUILD-best/training_args.bin +3 -0
- distilbert-base-cased-ft-BUILD/distilbert-base-cased-ft-BUILD-best/vocab.txt +0 -0
- distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709055246.DESKTOP-VOM7ARG.17955.0 +3 -0
- distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709055584.DESKTOP-VOM7ARG.19083.0 +3 -0
- distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709055671.DESKTOP-VOM7ARG.19446.0 +3 -0
- distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709055714.DESKTOP-VOM7ARG.19697.0 +3 -0
- distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709055822.DESKTOP-VOM7ARG.20135.0 +3 -0
- distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709056245.DESKTOP-VOM7ARG.762.0 +3 -0
- distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709056422.DESKTOP-VOM7ARG.1601.0 +3 -0
- distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709056587.DESKTOP-VOM7ARG.2226.0 +3 -0
- distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709061581.DESKTOP-VOM7ARG.16950.0 +3 -0
- distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709069376.DESKTOP-VOM7ARG.39969.0 +3 -0
- distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709069409.DESKTOP-VOM7ARG.40166.0 +3 -0
- distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709069468.DESKTOP-VOM7ARG.40470.0 +3 -0
- distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709069539.DESKTOP-VOM7ARG.40789.0 +3 -0
- distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709070129.DESKTOP-VOM7ARG.42636.0 +3 -0
- distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709070260.DESKTOP-VOM7ARG.43131.0 +3 -0
- distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709070283.DESKTOP-VOM7ARG.43300.0 +3 -0
- distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709070425.DESKTOP-VOM7ARG.43861.0 +3 -0
- distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709070475.DESKTOP-VOM7ARG.44142.0 +3 -0
- distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1710266349.DESKTOP-VOM7ARG.31210.0 +3 -0
- distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1710266663.DESKTOP-VOM7ARG.34038.0 +3 -0
- ingredients/distilbert-base-cased-ft-BUILD/checkpoint-1872/config.json +36 -0
- ingredients/distilbert-base-cased-ft-BUILD/checkpoint-1872/model.safetensors +3 -0
- ingredients/distilbert-base-cased-ft-BUILD/checkpoint-1872/optimizer.pt +3 -0
- ingredients/distilbert-base-cased-ft-BUILD/checkpoint-1872/rng_state.pth +3 -0
- ingredients/distilbert-base-cased-ft-BUILD/checkpoint-1872/scheduler.pt +3 -0
- ingredients/distilbert-base-cased-ft-BUILD/checkpoint-1872/trainer_state.json +108 -0
- ingredients/distilbert-base-cased-ft-BUILD/checkpoint-1872/training_args.bin +3 -0
- ingredients/distilbert-base-cased-ft-BUILD/distilbert-base-cased-ft-BUILD-best/config.json +36 -0
- ingredients/distilbert-base-cased-ft-BUILD/distilbert-base-cased-ft-BUILD-best/model.safetensors +3 -0
- ingredients/distilbert-base-cased-ft-BUILD/distilbert-base-cased-ft-BUILD-best/special_tokens_map.json +7 -0
- ingredients/distilbert-base-cased-ft-BUILD/distilbert-base-cased-ft-BUILD-best/tokenizer.json +0 -0
- ingredients/distilbert-base-cased-ft-BUILD/distilbert-base-cased-ft-BUILD-best/tokenizer_config.json +55 -0
- ingredients/distilbert-base-cased-ft-BUILD/distilbert-base-cased-ft-BUILD-best/training_args.bin +3 -0
- ingredients/distilbert-base-cased-ft-BUILD/distilbert-base-cased-ft-BUILD-best/vocab.txt +0 -0
- ingredients/distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1710279134.DESKTOP-VOM7ARG.62013.0 +3 -0
- ingredients/distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1710279165.DESKTOP-VOM7ARG.62362.0 +3 -0
distilbert-base-cased-ft-BUILD/checkpoint-3120/config.json
ADDED
|
@@ -0,0 +1,36 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_name_or_path": "distilbert-base-cased",
|
| 3 |
+
"activation": "gelu",
|
| 4 |
+
"architectures": [
|
| 5 |
+
"DistilBertForSequenceClassification"
|
| 6 |
+
],
|
| 7 |
+
"attention_dropout": 0.1,
|
| 8 |
+
"dim": 768,
|
| 9 |
+
"dropout": 0.1,
|
| 10 |
+
"hidden_dim": 3072,
|
| 11 |
+
"id2label": {
|
| 12 |
+
"0": "Entr\u00e9e",
|
| 13 |
+
"1": "Plat principal",
|
| 14 |
+
"2": "Dessert"
|
| 15 |
+
},
|
| 16 |
+
"initializer_range": 0.02,
|
| 17 |
+
"label2id": {
|
| 18 |
+
"Dessert": 2,
|
| 19 |
+
"Entr\u00e9e": 0,
|
| 20 |
+
"Plat principal": 1
|
| 21 |
+
},
|
| 22 |
+
"max_position_embeddings": 512,
|
| 23 |
+
"model_type": "distilbert",
|
| 24 |
+
"n_heads": 12,
|
| 25 |
+
"n_layers": 6,
|
| 26 |
+
"output_past": true,
|
| 27 |
+
"pad_token_id": 0,
|
| 28 |
+
"problem_type": "single_label_classification",
|
| 29 |
+
"qa_dropout": 0.1,
|
| 30 |
+
"seq_classif_dropout": 0.2,
|
| 31 |
+
"sinusoidal_pos_embds": false,
|
| 32 |
+
"tie_weights_": true,
|
| 33 |
+
"torch_dtype": "float32",
|
| 34 |
+
"transformers_version": "4.36.1",
|
| 35 |
+
"vocab_size": 28996
|
| 36 |
+
}
|
distilbert-base-cased-ft-BUILD/checkpoint-3120/model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:45f2614eb71c5ffcc8cdf28a4208a8e552faea6cf4a9d5275bb08bd000f05faf
|
| 3 |
+
size 263147764
|
distilbert-base-cased-ft-BUILD/checkpoint-3120/optimizer.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a2d596eda8e02e80a1f4a8c266deb0c26a604da01525470cfb4974aeea639e72
|
| 3 |
+
size 526357690
|
distilbert-base-cased-ft-BUILD/checkpoint-3120/rng_state.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ea472507a6e777cac31079a42c81583864d50bef5b62cba9d9c123095ece31c9
|
| 3 |
+
size 14244
|
distilbert-base-cased-ft-BUILD/checkpoint-3120/scheduler.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:172a1d25a78d8f5e7423dd4c0a4159c24b48a0ad09fb830eff556e9d62effb56
|
| 3 |
+
size 1064
|
distilbert-base-cased-ft-BUILD/checkpoint-3120/trainer_state.json
ADDED
|
@@ -0,0 +1,166 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"best_metric": 0.8537410092673323,
|
| 3 |
+
"best_model_checkpoint": "./models/distilbertcheffo/distilbert-base-cased-ft-BUILD/checkpoint-2496",
|
| 4 |
+
"epoch": 5.0,
|
| 5 |
+
"eval_steps": 500,
|
| 6 |
+
"global_step": 3120,
|
| 7 |
+
"is_hyper_param_search": false,
|
| 8 |
+
"is_local_process_zero": true,
|
| 9 |
+
"is_world_process_zero": true,
|
| 10 |
+
"log_history": [
|
| 11 |
+
{
|
| 12 |
+
"epoch": 0.32,
|
| 13 |
+
"learning_rate": 2e-05,
|
| 14 |
+
"loss": 0.915,
|
| 15 |
+
"step": 200
|
| 16 |
+
},
|
| 17 |
+
{
|
| 18 |
+
"epoch": 0.64,
|
| 19 |
+
"learning_rate": 4e-05,
|
| 20 |
+
"loss": 0.5652,
|
| 21 |
+
"step": 400
|
| 22 |
+
},
|
| 23 |
+
{
|
| 24 |
+
"epoch": 0.96,
|
| 25 |
+
"learning_rate": 4.845869297163995e-05,
|
| 26 |
+
"loss": 0.481,
|
| 27 |
+
"step": 600
|
| 28 |
+
},
|
| 29 |
+
{
|
| 30 |
+
"epoch": 1.0,
|
| 31 |
+
"eval_f1": 0.8121687008797033,
|
| 32 |
+
"eval_loss": 0.42347970604896545,
|
| 33 |
+
"eval_p": 0.822696467628716,
|
| 34 |
+
"eval_r": 0.8064903846153846,
|
| 35 |
+
"eval_runtime": 10.7242,
|
| 36 |
+
"eval_samples_per_second": 232.744,
|
| 37 |
+
"eval_steps_per_second": 14.546,
|
| 38 |
+
"step": 624
|
| 39 |
+
},
|
| 40 |
+
{
|
| 41 |
+
"epoch": 1.28,
|
| 42 |
+
"learning_rate": 4.5376078914919854e-05,
|
| 43 |
+
"loss": 0.4226,
|
| 44 |
+
"step": 800
|
| 45 |
+
},
|
| 46 |
+
{
|
| 47 |
+
"epoch": 1.6,
|
| 48 |
+
"learning_rate": 4.2293464858199754e-05,
|
| 49 |
+
"loss": 0.3854,
|
| 50 |
+
"step": 1000
|
| 51 |
+
},
|
| 52 |
+
{
|
| 53 |
+
"epoch": 1.92,
|
| 54 |
+
"learning_rate": 3.921085080147966e-05,
|
| 55 |
+
"loss": 0.3815,
|
| 56 |
+
"step": 1200
|
| 57 |
+
},
|
| 58 |
+
{
|
| 59 |
+
"epoch": 2.0,
|
| 60 |
+
"eval_f1": 0.8388272320074119,
|
| 61 |
+
"eval_loss": 0.3714698255062103,
|
| 62 |
+
"eval_p": 0.841085367336758,
|
| 63 |
+
"eval_r": 0.8449519230769231,
|
| 64 |
+
"eval_runtime": 10.8299,
|
| 65 |
+
"eval_samples_per_second": 230.472,
|
| 66 |
+
"eval_steps_per_second": 14.405,
|
| 67 |
+
"step": 1248
|
| 68 |
+
},
|
| 69 |
+
{
|
| 70 |
+
"epoch": 2.24,
|
| 71 |
+
"learning_rate": 3.6128236744759556e-05,
|
| 72 |
+
"loss": 0.3389,
|
| 73 |
+
"step": 1400
|
| 74 |
+
},
|
| 75 |
+
{
|
| 76 |
+
"epoch": 2.56,
|
| 77 |
+
"learning_rate": 3.304562268803946e-05,
|
| 78 |
+
"loss": 0.3105,
|
| 79 |
+
"step": 1600
|
| 80 |
+
},
|
| 81 |
+
{
|
| 82 |
+
"epoch": 2.88,
|
| 83 |
+
"learning_rate": 2.996300863131936e-05,
|
| 84 |
+
"loss": 0.3029,
|
| 85 |
+
"step": 1800
|
| 86 |
+
},
|
| 87 |
+
{
|
| 88 |
+
"epoch": 3.0,
|
| 89 |
+
"eval_f1": 0.8441213935683678,
|
| 90 |
+
"eval_loss": 0.37199097871780396,
|
| 91 |
+
"eval_p": 0.8445166019046795,
|
| 92 |
+
"eval_r": 0.84375,
|
| 93 |
+
"eval_runtime": 10.7362,
|
| 94 |
+
"eval_samples_per_second": 232.484,
|
| 95 |
+
"eval_steps_per_second": 14.53,
|
| 96 |
+
"step": 1872
|
| 97 |
+
},
|
| 98 |
+
{
|
| 99 |
+
"epoch": 3.21,
|
| 100 |
+
"learning_rate": 2.688039457459926e-05,
|
| 101 |
+
"loss": 0.2631,
|
| 102 |
+
"step": 2000
|
| 103 |
+
},
|
| 104 |
+
{
|
| 105 |
+
"epoch": 3.53,
|
| 106 |
+
"learning_rate": 2.3797780517879162e-05,
|
| 107 |
+
"loss": 0.2392,
|
| 108 |
+
"step": 2200
|
| 109 |
+
},
|
| 110 |
+
{
|
| 111 |
+
"epoch": 3.85,
|
| 112 |
+
"learning_rate": 2.0715166461159063e-05,
|
| 113 |
+
"loss": 0.2474,
|
| 114 |
+
"step": 2400
|
| 115 |
+
},
|
| 116 |
+
{
|
| 117 |
+
"epoch": 4.0,
|
| 118 |
+
"eval_f1": 0.8537410092673323,
|
| 119 |
+
"eval_loss": 0.37799614667892456,
|
| 120 |
+
"eval_p": 0.8537701589320283,
|
| 121 |
+
"eval_r": 0.8537660256410257,
|
| 122 |
+
"eval_runtime": 10.7429,
|
| 123 |
+
"eval_samples_per_second": 232.34,
|
| 124 |
+
"eval_steps_per_second": 14.521,
|
| 125 |
+
"step": 2496
|
| 126 |
+
},
|
| 127 |
+
{
|
| 128 |
+
"epoch": 4.17,
|
| 129 |
+
"learning_rate": 1.7632552404438964e-05,
|
| 130 |
+
"loss": 0.2055,
|
| 131 |
+
"step": 2600
|
| 132 |
+
},
|
| 133 |
+
{
|
| 134 |
+
"epoch": 4.49,
|
| 135 |
+
"learning_rate": 1.4549938347718866e-05,
|
| 136 |
+
"loss": 0.1797,
|
| 137 |
+
"step": 2800
|
| 138 |
+
},
|
| 139 |
+
{
|
| 140 |
+
"epoch": 4.81,
|
| 141 |
+
"learning_rate": 1.1467324290998767e-05,
|
| 142 |
+
"loss": 0.1757,
|
| 143 |
+
"step": 3000
|
| 144 |
+
},
|
| 145 |
+
{
|
| 146 |
+
"epoch": 5.0,
|
| 147 |
+
"eval_f1": 0.8502410363258266,
|
| 148 |
+
"eval_loss": 0.5066409707069397,
|
| 149 |
+
"eval_p": 0.8539723028554528,
|
| 150 |
+
"eval_r": 0.8485576923076923,
|
| 151 |
+
"eval_runtime": 10.7461,
|
| 152 |
+
"eval_samples_per_second": 232.269,
|
| 153 |
+
"eval_steps_per_second": 14.517,
|
| 154 |
+
"step": 3120
|
| 155 |
+
}
|
| 156 |
+
],
|
| 157 |
+
"logging_steps": 200,
|
| 158 |
+
"max_steps": 3744,
|
| 159 |
+
"num_input_tokens_seen": 0,
|
| 160 |
+
"num_train_epochs": 6,
|
| 161 |
+
"save_steps": 500,
|
| 162 |
+
"total_flos": 6608254028682240.0,
|
| 163 |
+
"train_batch_size": 16,
|
| 164 |
+
"trial_name": null,
|
| 165 |
+
"trial_params": null
|
| 166 |
+
}
|
distilbert-base-cased-ft-BUILD/checkpoint-3120/training_args.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ae357d8ada26e9e9b973b4b3652af41f8e13ef19e617413bfdef1e140acfd572
|
| 3 |
+
size 4792
|
distilbert-base-cased-ft-BUILD/distilbert-base-cased-ft-BUILD-best/config.json
ADDED
|
@@ -0,0 +1,36 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_name_or_path": "distilbert-base-cased",
|
| 3 |
+
"activation": "gelu",
|
| 4 |
+
"architectures": [
|
| 5 |
+
"DistilBertForSequenceClassification"
|
| 6 |
+
],
|
| 7 |
+
"attention_dropout": 0.1,
|
| 8 |
+
"dim": 768,
|
| 9 |
+
"dropout": 0.1,
|
| 10 |
+
"hidden_dim": 3072,
|
| 11 |
+
"id2label": {
|
| 12 |
+
"0": "Entr\u00e9e",
|
| 13 |
+
"1": "Plat principal",
|
| 14 |
+
"2": "Dessert"
|
| 15 |
+
},
|
| 16 |
+
"initializer_range": 0.02,
|
| 17 |
+
"label2id": {
|
| 18 |
+
"Dessert": 2,
|
| 19 |
+
"Entr\u00e9e": 0,
|
| 20 |
+
"Plat principal": 1
|
| 21 |
+
},
|
| 22 |
+
"max_position_embeddings": 512,
|
| 23 |
+
"model_type": "distilbert",
|
| 24 |
+
"n_heads": 12,
|
| 25 |
+
"n_layers": 6,
|
| 26 |
+
"output_past": true,
|
| 27 |
+
"pad_token_id": 0,
|
| 28 |
+
"problem_type": "single_label_classification",
|
| 29 |
+
"qa_dropout": 0.1,
|
| 30 |
+
"seq_classif_dropout": 0.2,
|
| 31 |
+
"sinusoidal_pos_embds": false,
|
| 32 |
+
"tie_weights_": true,
|
| 33 |
+
"torch_dtype": "float32",
|
| 34 |
+
"transformers_version": "4.36.1",
|
| 35 |
+
"vocab_size": 28996
|
| 36 |
+
}
|
distilbert-base-cased-ft-BUILD/distilbert-base-cased-ft-BUILD-best/model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ebebcfae8ef9207504b86bf25933da79087ff9ea53cf25db77e4c854e7190746
|
| 3 |
+
size 263147764
|
distilbert-base-cased-ft-BUILD/distilbert-base-cased-ft-BUILD-best/special_tokens_map.json
ADDED
|
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"cls_token": "[CLS]",
|
| 3 |
+
"mask_token": "[MASK]",
|
| 4 |
+
"pad_token": "[PAD]",
|
| 5 |
+
"sep_token": "[SEP]",
|
| 6 |
+
"unk_token": "[UNK]"
|
| 7 |
+
}
|
distilbert-base-cased-ft-BUILD/distilbert-base-cased-ft-BUILD-best/tokenizer.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
distilbert-base-cased-ft-BUILD/distilbert-base-cased-ft-BUILD-best/tokenizer_config.json
ADDED
|
@@ -0,0 +1,55 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"added_tokens_decoder": {
|
| 3 |
+
"0": {
|
| 4 |
+
"content": "[PAD]",
|
| 5 |
+
"lstrip": false,
|
| 6 |
+
"normalized": false,
|
| 7 |
+
"rstrip": false,
|
| 8 |
+
"single_word": false,
|
| 9 |
+
"special": true
|
| 10 |
+
},
|
| 11 |
+
"100": {
|
| 12 |
+
"content": "[UNK]",
|
| 13 |
+
"lstrip": false,
|
| 14 |
+
"normalized": false,
|
| 15 |
+
"rstrip": false,
|
| 16 |
+
"single_word": false,
|
| 17 |
+
"special": true
|
| 18 |
+
},
|
| 19 |
+
"101": {
|
| 20 |
+
"content": "[CLS]",
|
| 21 |
+
"lstrip": false,
|
| 22 |
+
"normalized": false,
|
| 23 |
+
"rstrip": false,
|
| 24 |
+
"single_word": false,
|
| 25 |
+
"special": true
|
| 26 |
+
},
|
| 27 |
+
"102": {
|
| 28 |
+
"content": "[SEP]",
|
| 29 |
+
"lstrip": false,
|
| 30 |
+
"normalized": false,
|
| 31 |
+
"rstrip": false,
|
| 32 |
+
"single_word": false,
|
| 33 |
+
"special": true
|
| 34 |
+
},
|
| 35 |
+
"103": {
|
| 36 |
+
"content": "[MASK]",
|
| 37 |
+
"lstrip": false,
|
| 38 |
+
"normalized": false,
|
| 39 |
+
"rstrip": false,
|
| 40 |
+
"single_word": false,
|
| 41 |
+
"special": true
|
| 42 |
+
}
|
| 43 |
+
},
|
| 44 |
+
"clean_up_tokenization_spaces": true,
|
| 45 |
+
"cls_token": "[CLS]",
|
| 46 |
+
"do_lower_case": false,
|
| 47 |
+
"mask_token": "[MASK]",
|
| 48 |
+
"model_max_length": 512,
|
| 49 |
+
"pad_token": "[PAD]",
|
| 50 |
+
"sep_token": "[SEP]",
|
| 51 |
+
"strip_accents": null,
|
| 52 |
+
"tokenize_chinese_chars": true,
|
| 53 |
+
"tokenizer_class": "DistilBertTokenizer",
|
| 54 |
+
"unk_token": "[UNK]"
|
| 55 |
+
}
|
distilbert-base-cased-ft-BUILD/distilbert-base-cased-ft-BUILD-best/training_args.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9a4c98641c0c92377fd9ae69d06b9f4baea0b2802dcb0ab1974894b57bb12353
|
| 3 |
+
size 4792
|
distilbert-base-cased-ft-BUILD/distilbert-base-cased-ft-BUILD-best/vocab.txt
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709055246.DESKTOP-VOM7ARG.17955.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5db5a60ab930869a0380743583bb1f90cf4f93ce105ea1b2629e3b9aa8e36cdb
|
| 3 |
+
size 4939
|
distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709055584.DESKTOP-VOM7ARG.19083.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:23ac74ae8e44e435e383d5b26c5122979fb2a93d83c0982306c4a055342f4228
|
| 3 |
+
size 4939
|
distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709055671.DESKTOP-VOM7ARG.19446.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:921b6c050ae8eeb363fa91cd08f88d52545e20ba896672ed34767fd63d43b48f
|
| 3 |
+
size 4939
|
distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709055714.DESKTOP-VOM7ARG.19697.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0935a20d9ba84cc7d9bf8e35c617923d0f9063fae00ab1bfae44e9d57d914f65
|
| 3 |
+
size 4939
|
distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709055822.DESKTOP-VOM7ARG.20135.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ef0be61d09f03b34d918c75ff88913565c1500abaa093de29cf7d58ab634677c
|
| 3 |
+
size 4939
|
distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709056245.DESKTOP-VOM7ARG.762.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:906c7603626a5e8cc8e9493a36bf674d4d6b623e9bb42b759a7ff3e527e0b7c8
|
| 3 |
+
size 4939
|
distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709056422.DESKTOP-VOM7ARG.1601.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e46c66f4e7e23d928e65cb24c329cb60d67d86732fd8bc0dcdcc2a874c384398
|
| 3 |
+
size 4939
|
distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709056587.DESKTOP-VOM7ARG.2226.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3b3af101cad7ed8a545698ba6169e8cbd4d9001845a7af61bafc0ab4398be483
|
| 3 |
+
size 7927
|
distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709061581.DESKTOP-VOM7ARG.16950.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c1dabde6a72e69d155aa076d1236a9098340c9ea2a14efe5d829876ec47e18e9
|
| 3 |
+
size 10561
|
distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709069376.DESKTOP-VOM7ARG.39969.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3d9574088e90779d5d8bf0001caefa4a1966a15d8fbcfa9e4ae5512266452f71
|
| 3 |
+
size 4545
|
distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709069409.DESKTOP-VOM7ARG.40166.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:76f55194729e28eececd3d7b8277e06b7e0194c0385cd15ac792e30b338cb77f
|
| 3 |
+
size 4545
|
distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709069468.DESKTOP-VOM7ARG.40470.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:257caba1515e99871cb5a101ff9b64d926ae7784d1028abc9d207865c5987c64
|
| 3 |
+
size 4545
|
distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709069539.DESKTOP-VOM7ARG.40789.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:81314967f07de88e7c9ace94f3fca0ec41d702147651936710e82c5d1c5c3a09
|
| 3 |
+
size 4939
|
distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709070129.DESKTOP-VOM7ARG.42636.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1bc58f60ee608b3db87308510611476d4e3dff01d351e566e4a273aaf7872f9c
|
| 3 |
+
size 4559
|
distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709070260.DESKTOP-VOM7ARG.43131.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a0dfede442d3bf163c881f1384c8daea4b72a8989dbb235d813c2b0986e39e84
|
| 3 |
+
size 4565
|
distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709070283.DESKTOP-VOM7ARG.43300.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5e4426ead234a12ef0c91beed610e1bcabcf525c288f8d9ed1ecffd74133678b
|
| 3 |
+
size 4535
|
distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709070425.DESKTOP-VOM7ARG.43861.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:28c5f5dd618291e77bdfdc4f764ac3d2c1583d516d3874a3d6a249cb2a2234a7
|
| 3 |
+
size 4535
|
distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709070475.DESKTOP-VOM7ARG.44142.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9bc317be60ce43259954518fc857e129c055868d15b5a460eedb5c3b1e7b824d
|
| 3 |
+
size 10191
|
distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1710266349.DESKTOP-VOM7ARG.31210.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f31c9ae80d85f653f461c6ca1c100d8d1652d6109c2595f89354b3e23323df3e
|
| 3 |
+
size 4570
|
distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1710266663.DESKTOP-VOM7ARG.34038.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b1198236de54436455f37a9c60ce148c9f4d9e41e05d85dbbb3ac41edf2cef3a
|
| 3 |
+
size 17530
|
ingredients/distilbert-base-cased-ft-BUILD/checkpoint-1872/config.json
ADDED
|
@@ -0,0 +1,36 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_name_or_path": "distilbert-base-cased",
|
| 3 |
+
"activation": "gelu",
|
| 4 |
+
"architectures": [
|
| 5 |
+
"DistilBertForSequenceClassification"
|
| 6 |
+
],
|
| 7 |
+
"attention_dropout": 0.1,
|
| 8 |
+
"dim": 768,
|
| 9 |
+
"dropout": 0.1,
|
| 10 |
+
"hidden_dim": 3072,
|
| 11 |
+
"id2label": {
|
| 12 |
+
"0": "Entr\u00e9e",
|
| 13 |
+
"1": "Plat principal",
|
| 14 |
+
"2": "Dessert"
|
| 15 |
+
},
|
| 16 |
+
"initializer_range": 0.02,
|
| 17 |
+
"label2id": {
|
| 18 |
+
"Dessert": 2,
|
| 19 |
+
"Entr\u00e9e": 0,
|
| 20 |
+
"Plat principal": 1
|
| 21 |
+
},
|
| 22 |
+
"max_position_embeddings": 512,
|
| 23 |
+
"model_type": "distilbert",
|
| 24 |
+
"n_heads": 12,
|
| 25 |
+
"n_layers": 6,
|
| 26 |
+
"output_past": true,
|
| 27 |
+
"pad_token_id": 0,
|
| 28 |
+
"problem_type": "single_label_classification",
|
| 29 |
+
"qa_dropout": 0.1,
|
| 30 |
+
"seq_classif_dropout": 0.2,
|
| 31 |
+
"sinusoidal_pos_embds": false,
|
| 32 |
+
"tie_weights_": true,
|
| 33 |
+
"torch_dtype": "float32",
|
| 34 |
+
"transformers_version": "4.36.1",
|
| 35 |
+
"vocab_size": 28996
|
| 36 |
+
}
|
ingredients/distilbert-base-cased-ft-BUILD/checkpoint-1872/model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f946d23df2edf87b47814770bed590c07aa4b98ced92381ae2bde5c72f31463c
|
| 3 |
+
size 263147764
|
ingredients/distilbert-base-cased-ft-BUILD/checkpoint-1872/optimizer.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:926c308bdbdab946b05bf732ab2e5bb2192fe99ae6476c4fcf8492b11d0871db
|
| 3 |
+
size 526357690
|
ingredients/distilbert-base-cased-ft-BUILD/checkpoint-1872/rng_state.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:899bf7ce686ec08a8b70dd92747507b3b430f9484252777f2df648363b93a1f9
|
| 3 |
+
size 14244
|
ingredients/distilbert-base-cased-ft-BUILD/checkpoint-1872/scheduler.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:96c94514ae676650b0a1c947f585caa3a3b0bb1010bdc06d92e7e19432a456cd
|
| 3 |
+
size 1064
|
ingredients/distilbert-base-cased-ft-BUILD/checkpoint-1872/trainer_state.json
ADDED
|
@@ -0,0 +1,108 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"best_metric": 0.8201436140182747,
|
| 3 |
+
"best_model_checkpoint": "./models/distilbertcheffo/ingredients/distilbert-base-cased-ft-BUILD/checkpoint-1248",
|
| 4 |
+
"epoch": 3.0,
|
| 5 |
+
"eval_steps": 500,
|
| 6 |
+
"global_step": 1872,
|
| 7 |
+
"is_hyper_param_search": false,
|
| 8 |
+
"is_local_process_zero": true,
|
| 9 |
+
"is_world_process_zero": true,
|
| 10 |
+
"log_history": [
|
| 11 |
+
{
|
| 12 |
+
"epoch": 0.32,
|
| 13 |
+
"learning_rate": 2e-05,
|
| 14 |
+
"loss": 0.8064,
|
| 15 |
+
"step": 200
|
| 16 |
+
},
|
| 17 |
+
{
|
| 18 |
+
"epoch": 0.64,
|
| 19 |
+
"learning_rate": 4e-05,
|
| 20 |
+
"loss": 0.5328,
|
| 21 |
+
"step": 400
|
| 22 |
+
},
|
| 23 |
+
{
|
| 24 |
+
"epoch": 0.96,
|
| 25 |
+
"learning_rate": 4.958263772954925e-05,
|
| 26 |
+
"loss": 0.491,
|
| 27 |
+
"step": 600
|
| 28 |
+
},
|
| 29 |
+
{
|
| 30 |
+
"epoch": 1.0,
|
| 31 |
+
"eval_f1": 0.7625179981925605,
|
| 32 |
+
"eval_loss": 0.4677772521972656,
|
| 33 |
+
"eval_p": 0.7920354842384022,
|
| 34 |
+
"eval_r": 0.7564102564102564,
|
| 35 |
+
"eval_runtime": 10.7329,
|
| 36 |
+
"eval_samples_per_second": 232.557,
|
| 37 |
+
"eval_steps_per_second": 14.535,
|
| 38 |
+
"step": 624
|
| 39 |
+
},
|
| 40 |
+
{
|
| 41 |
+
"epoch": 1.28,
|
| 42 |
+
"learning_rate": 4.8747913188647746e-05,
|
| 43 |
+
"loss": 0.4444,
|
| 44 |
+
"step": 800
|
| 45 |
+
},
|
| 46 |
+
{
|
| 47 |
+
"epoch": 1.6,
|
| 48 |
+
"learning_rate": 4.791318864774624e-05,
|
| 49 |
+
"loss": 0.4212,
|
| 50 |
+
"step": 1000
|
| 51 |
+
},
|
| 52 |
+
{
|
| 53 |
+
"epoch": 1.92,
|
| 54 |
+
"learning_rate": 4.7078464106844744e-05,
|
| 55 |
+
"loss": 0.4269,
|
| 56 |
+
"step": 1200
|
| 57 |
+
},
|
| 58 |
+
{
|
| 59 |
+
"epoch": 2.0,
|
| 60 |
+
"eval_f1": 0.8201436140182747,
|
| 61 |
+
"eval_loss": 0.4101373255252838,
|
| 62 |
+
"eval_p": 0.8195197123847887,
|
| 63 |
+
"eval_r": 0.8241185897435898,
|
| 64 |
+
"eval_runtime": 10.7229,
|
| 65 |
+
"eval_samples_per_second": 232.772,
|
| 66 |
+
"eval_steps_per_second": 14.548,
|
| 67 |
+
"step": 1248
|
| 68 |
+
},
|
| 69 |
+
{
|
| 70 |
+
"epoch": 2.24,
|
| 71 |
+
"learning_rate": 4.624373956594324e-05,
|
| 72 |
+
"loss": 0.3915,
|
| 73 |
+
"step": 1400
|
| 74 |
+
},
|
| 75 |
+
{
|
| 76 |
+
"epoch": 2.56,
|
| 77 |
+
"learning_rate": 4.540901502504174e-05,
|
| 78 |
+
"loss": 0.3516,
|
| 79 |
+
"step": 1600
|
| 80 |
+
},
|
| 81 |
+
{
|
| 82 |
+
"epoch": 2.88,
|
| 83 |
+
"learning_rate": 4.457429048414024e-05,
|
| 84 |
+
"loss": 0.3638,
|
| 85 |
+
"step": 1800
|
| 86 |
+
},
|
| 87 |
+
{
|
| 88 |
+
"epoch": 3.0,
|
| 89 |
+
"eval_f1": 0.8160936815512525,
|
| 90 |
+
"eval_loss": 0.42037826776504517,
|
| 91 |
+
"eval_p": 0.8239885163327829,
|
| 92 |
+
"eval_r": 0.8265224358974359,
|
| 93 |
+
"eval_runtime": 10.7428,
|
| 94 |
+
"eval_samples_per_second": 232.341,
|
| 95 |
+
"eval_steps_per_second": 14.521,
|
| 96 |
+
"step": 1872
|
| 97 |
+
}
|
| 98 |
+
],
|
| 99 |
+
"logging_steps": 200,
|
| 100 |
+
"max_steps": 12480,
|
| 101 |
+
"num_input_tokens_seen": 0,
|
| 102 |
+
"num_train_epochs": 20,
|
| 103 |
+
"save_steps": 500,
|
| 104 |
+
"total_flos": 3964952417209344.0,
|
| 105 |
+
"train_batch_size": 16,
|
| 106 |
+
"trial_name": null,
|
| 107 |
+
"trial_params": null
|
| 108 |
+
}
|
ingredients/distilbert-base-cased-ft-BUILD/checkpoint-1872/training_args.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ae5d9e5738c101bc04aa4a0dc3590dc468dd7a97882afbc922e96e1027023bfd
|
| 3 |
+
size 4856
|
ingredients/distilbert-base-cased-ft-BUILD/distilbert-base-cased-ft-BUILD-best/config.json
ADDED
|
@@ -0,0 +1,36 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_name_or_path": "distilbert-base-cased",
|
| 3 |
+
"activation": "gelu",
|
| 4 |
+
"architectures": [
|
| 5 |
+
"DistilBertForSequenceClassification"
|
| 6 |
+
],
|
| 7 |
+
"attention_dropout": 0.1,
|
| 8 |
+
"dim": 768,
|
| 9 |
+
"dropout": 0.1,
|
| 10 |
+
"hidden_dim": 3072,
|
| 11 |
+
"id2label": {
|
| 12 |
+
"0": "Entr\u00e9e",
|
| 13 |
+
"1": "Plat principal",
|
| 14 |
+
"2": "Dessert"
|
| 15 |
+
},
|
| 16 |
+
"initializer_range": 0.02,
|
| 17 |
+
"label2id": {
|
| 18 |
+
"Dessert": 2,
|
| 19 |
+
"Entr\u00e9e": 0,
|
| 20 |
+
"Plat principal": 1
|
| 21 |
+
},
|
| 22 |
+
"max_position_embeddings": 512,
|
| 23 |
+
"model_type": "distilbert",
|
| 24 |
+
"n_heads": 12,
|
| 25 |
+
"n_layers": 6,
|
| 26 |
+
"output_past": true,
|
| 27 |
+
"pad_token_id": 0,
|
| 28 |
+
"problem_type": "single_label_classification",
|
| 29 |
+
"qa_dropout": 0.1,
|
| 30 |
+
"seq_classif_dropout": 0.2,
|
| 31 |
+
"sinusoidal_pos_embds": false,
|
| 32 |
+
"tie_weights_": true,
|
| 33 |
+
"torch_dtype": "float32",
|
| 34 |
+
"transformers_version": "4.36.1",
|
| 35 |
+
"vocab_size": 28996
|
| 36 |
+
}
|
ingredients/distilbert-base-cased-ft-BUILD/distilbert-base-cased-ft-BUILD-best/model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7f8a991280469d06f9edf8203a5b9613afe79fd6aedc23b255c26b5add0bf312
|
| 3 |
+
size 263147764
|
ingredients/distilbert-base-cased-ft-BUILD/distilbert-base-cased-ft-BUILD-best/special_tokens_map.json
ADDED
|
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"cls_token": "[CLS]",
|
| 3 |
+
"mask_token": "[MASK]",
|
| 4 |
+
"pad_token": "[PAD]",
|
| 5 |
+
"sep_token": "[SEP]",
|
| 6 |
+
"unk_token": "[UNK]"
|
| 7 |
+
}
|
ingredients/distilbert-base-cased-ft-BUILD/distilbert-base-cased-ft-BUILD-best/tokenizer.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
ingredients/distilbert-base-cased-ft-BUILD/distilbert-base-cased-ft-BUILD-best/tokenizer_config.json
ADDED
|
@@ -0,0 +1,55 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"added_tokens_decoder": {
|
| 3 |
+
"0": {
|
| 4 |
+
"content": "[PAD]",
|
| 5 |
+
"lstrip": false,
|
| 6 |
+
"normalized": false,
|
| 7 |
+
"rstrip": false,
|
| 8 |
+
"single_word": false,
|
| 9 |
+
"special": true
|
| 10 |
+
},
|
| 11 |
+
"100": {
|
| 12 |
+
"content": "[UNK]",
|
| 13 |
+
"lstrip": false,
|
| 14 |
+
"normalized": false,
|
| 15 |
+
"rstrip": false,
|
| 16 |
+
"single_word": false,
|
| 17 |
+
"special": true
|
| 18 |
+
},
|
| 19 |
+
"101": {
|
| 20 |
+
"content": "[CLS]",
|
| 21 |
+
"lstrip": false,
|
| 22 |
+
"normalized": false,
|
| 23 |
+
"rstrip": false,
|
| 24 |
+
"single_word": false,
|
| 25 |
+
"special": true
|
| 26 |
+
},
|
| 27 |
+
"102": {
|
| 28 |
+
"content": "[SEP]",
|
| 29 |
+
"lstrip": false,
|
| 30 |
+
"normalized": false,
|
| 31 |
+
"rstrip": false,
|
| 32 |
+
"single_word": false,
|
| 33 |
+
"special": true
|
| 34 |
+
},
|
| 35 |
+
"103": {
|
| 36 |
+
"content": "[MASK]",
|
| 37 |
+
"lstrip": false,
|
| 38 |
+
"normalized": false,
|
| 39 |
+
"rstrip": false,
|
| 40 |
+
"single_word": false,
|
| 41 |
+
"special": true
|
| 42 |
+
}
|
| 43 |
+
},
|
| 44 |
+
"clean_up_tokenization_spaces": true,
|
| 45 |
+
"cls_token": "[CLS]",
|
| 46 |
+
"do_lower_case": false,
|
| 47 |
+
"mask_token": "[MASK]",
|
| 48 |
+
"model_max_length": 512,
|
| 49 |
+
"pad_token": "[PAD]",
|
| 50 |
+
"sep_token": "[SEP]",
|
| 51 |
+
"strip_accents": null,
|
| 52 |
+
"tokenize_chinese_chars": true,
|
| 53 |
+
"tokenizer_class": "DistilBertTokenizer",
|
| 54 |
+
"unk_token": "[UNK]"
|
| 55 |
+
}
|
ingredients/distilbert-base-cased-ft-BUILD/distilbert-base-cased-ft-BUILD-best/training_args.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ae5d9e5738c101bc04aa4a0dc3590dc468dd7a97882afbc922e96e1027023bfd
|
| 3 |
+
size 4856
|
ingredients/distilbert-base-cased-ft-BUILD/distilbert-base-cased-ft-BUILD-best/vocab.txt
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
ingredients/distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1710279134.DESKTOP-VOM7ARG.62013.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:52beba5db112e49134740a73aded280dfd5941d1c596f0feca1b4c89ec6fea44
|
| 3 |
+
size 4606
|
ingredients/distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1710279165.DESKTOP-VOM7ARG.62362.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9d3a0170e6b7dab2f08fbd513e1c734583f0f7090fd3c1127f13d8fe2b967908
|
| 3 |
+
size 4606
|