labicquette commited on
Commit
9500443
·
verified ·
1 Parent(s): 508c016

Upload 70 files

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. distilbert-base-cased-ft-BUILD/checkpoint-3120/config.json +36 -0
  2. distilbert-base-cased-ft-BUILD/checkpoint-3120/model.safetensors +3 -0
  3. distilbert-base-cased-ft-BUILD/checkpoint-3120/optimizer.pt +3 -0
  4. distilbert-base-cased-ft-BUILD/checkpoint-3120/rng_state.pth +3 -0
  5. distilbert-base-cased-ft-BUILD/checkpoint-3120/scheduler.pt +3 -0
  6. distilbert-base-cased-ft-BUILD/checkpoint-3120/trainer_state.json +166 -0
  7. distilbert-base-cased-ft-BUILD/checkpoint-3120/training_args.bin +3 -0
  8. distilbert-base-cased-ft-BUILD/distilbert-base-cased-ft-BUILD-best/config.json +36 -0
  9. distilbert-base-cased-ft-BUILD/distilbert-base-cased-ft-BUILD-best/model.safetensors +3 -0
  10. distilbert-base-cased-ft-BUILD/distilbert-base-cased-ft-BUILD-best/special_tokens_map.json +7 -0
  11. distilbert-base-cased-ft-BUILD/distilbert-base-cased-ft-BUILD-best/tokenizer.json +0 -0
  12. distilbert-base-cased-ft-BUILD/distilbert-base-cased-ft-BUILD-best/tokenizer_config.json +55 -0
  13. distilbert-base-cased-ft-BUILD/distilbert-base-cased-ft-BUILD-best/training_args.bin +3 -0
  14. distilbert-base-cased-ft-BUILD/distilbert-base-cased-ft-BUILD-best/vocab.txt +0 -0
  15. distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709055246.DESKTOP-VOM7ARG.17955.0 +3 -0
  16. distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709055584.DESKTOP-VOM7ARG.19083.0 +3 -0
  17. distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709055671.DESKTOP-VOM7ARG.19446.0 +3 -0
  18. distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709055714.DESKTOP-VOM7ARG.19697.0 +3 -0
  19. distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709055822.DESKTOP-VOM7ARG.20135.0 +3 -0
  20. distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709056245.DESKTOP-VOM7ARG.762.0 +3 -0
  21. distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709056422.DESKTOP-VOM7ARG.1601.0 +3 -0
  22. distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709056587.DESKTOP-VOM7ARG.2226.0 +3 -0
  23. distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709061581.DESKTOP-VOM7ARG.16950.0 +3 -0
  24. distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709069376.DESKTOP-VOM7ARG.39969.0 +3 -0
  25. distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709069409.DESKTOP-VOM7ARG.40166.0 +3 -0
  26. distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709069468.DESKTOP-VOM7ARG.40470.0 +3 -0
  27. distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709069539.DESKTOP-VOM7ARG.40789.0 +3 -0
  28. distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709070129.DESKTOP-VOM7ARG.42636.0 +3 -0
  29. distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709070260.DESKTOP-VOM7ARG.43131.0 +3 -0
  30. distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709070283.DESKTOP-VOM7ARG.43300.0 +3 -0
  31. distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709070425.DESKTOP-VOM7ARG.43861.0 +3 -0
  32. distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709070475.DESKTOP-VOM7ARG.44142.0 +3 -0
  33. distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1710266349.DESKTOP-VOM7ARG.31210.0 +3 -0
  34. distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1710266663.DESKTOP-VOM7ARG.34038.0 +3 -0
  35. ingredients/distilbert-base-cased-ft-BUILD/checkpoint-1872/config.json +36 -0
  36. ingredients/distilbert-base-cased-ft-BUILD/checkpoint-1872/model.safetensors +3 -0
  37. ingredients/distilbert-base-cased-ft-BUILD/checkpoint-1872/optimizer.pt +3 -0
  38. ingredients/distilbert-base-cased-ft-BUILD/checkpoint-1872/rng_state.pth +3 -0
  39. ingredients/distilbert-base-cased-ft-BUILD/checkpoint-1872/scheduler.pt +3 -0
  40. ingredients/distilbert-base-cased-ft-BUILD/checkpoint-1872/trainer_state.json +108 -0
  41. ingredients/distilbert-base-cased-ft-BUILD/checkpoint-1872/training_args.bin +3 -0
  42. ingredients/distilbert-base-cased-ft-BUILD/distilbert-base-cased-ft-BUILD-best/config.json +36 -0
  43. ingredients/distilbert-base-cased-ft-BUILD/distilbert-base-cased-ft-BUILD-best/model.safetensors +3 -0
  44. ingredients/distilbert-base-cased-ft-BUILD/distilbert-base-cased-ft-BUILD-best/special_tokens_map.json +7 -0
  45. ingredients/distilbert-base-cased-ft-BUILD/distilbert-base-cased-ft-BUILD-best/tokenizer.json +0 -0
  46. ingredients/distilbert-base-cased-ft-BUILD/distilbert-base-cased-ft-BUILD-best/tokenizer_config.json +55 -0
  47. ingredients/distilbert-base-cased-ft-BUILD/distilbert-base-cased-ft-BUILD-best/training_args.bin +3 -0
  48. ingredients/distilbert-base-cased-ft-BUILD/distilbert-base-cased-ft-BUILD-best/vocab.txt +0 -0
  49. ingredients/distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1710279134.DESKTOP-VOM7ARG.62013.0 +3 -0
  50. ingredients/distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1710279165.DESKTOP-VOM7ARG.62362.0 +3 -0
distilbert-base-cased-ft-BUILD/checkpoint-3120/config.json ADDED
@@ -0,0 +1,36 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "distilbert-base-cased",
3
+ "activation": "gelu",
4
+ "architectures": [
5
+ "DistilBertForSequenceClassification"
6
+ ],
7
+ "attention_dropout": 0.1,
8
+ "dim": 768,
9
+ "dropout": 0.1,
10
+ "hidden_dim": 3072,
11
+ "id2label": {
12
+ "0": "Entr\u00e9e",
13
+ "1": "Plat principal",
14
+ "2": "Dessert"
15
+ },
16
+ "initializer_range": 0.02,
17
+ "label2id": {
18
+ "Dessert": 2,
19
+ "Entr\u00e9e": 0,
20
+ "Plat principal": 1
21
+ },
22
+ "max_position_embeddings": 512,
23
+ "model_type": "distilbert",
24
+ "n_heads": 12,
25
+ "n_layers": 6,
26
+ "output_past": true,
27
+ "pad_token_id": 0,
28
+ "problem_type": "single_label_classification",
29
+ "qa_dropout": 0.1,
30
+ "seq_classif_dropout": 0.2,
31
+ "sinusoidal_pos_embds": false,
32
+ "tie_weights_": true,
33
+ "torch_dtype": "float32",
34
+ "transformers_version": "4.36.1",
35
+ "vocab_size": 28996
36
+ }
distilbert-base-cased-ft-BUILD/checkpoint-3120/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:45f2614eb71c5ffcc8cdf28a4208a8e552faea6cf4a9d5275bb08bd000f05faf
3
+ size 263147764
distilbert-base-cased-ft-BUILD/checkpoint-3120/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a2d596eda8e02e80a1f4a8c266deb0c26a604da01525470cfb4974aeea639e72
3
+ size 526357690
distilbert-base-cased-ft-BUILD/checkpoint-3120/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ea472507a6e777cac31079a42c81583864d50bef5b62cba9d9c123095ece31c9
3
+ size 14244
distilbert-base-cased-ft-BUILD/checkpoint-3120/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:172a1d25a78d8f5e7423dd4c0a4159c24b48a0ad09fb830eff556e9d62effb56
3
+ size 1064
distilbert-base-cased-ft-BUILD/checkpoint-3120/trainer_state.json ADDED
@@ -0,0 +1,166 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.8537410092673323,
3
+ "best_model_checkpoint": "./models/distilbertcheffo/distilbert-base-cased-ft-BUILD/checkpoint-2496",
4
+ "epoch": 5.0,
5
+ "eval_steps": 500,
6
+ "global_step": 3120,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.32,
13
+ "learning_rate": 2e-05,
14
+ "loss": 0.915,
15
+ "step": 200
16
+ },
17
+ {
18
+ "epoch": 0.64,
19
+ "learning_rate": 4e-05,
20
+ "loss": 0.5652,
21
+ "step": 400
22
+ },
23
+ {
24
+ "epoch": 0.96,
25
+ "learning_rate": 4.845869297163995e-05,
26
+ "loss": 0.481,
27
+ "step": 600
28
+ },
29
+ {
30
+ "epoch": 1.0,
31
+ "eval_f1": 0.8121687008797033,
32
+ "eval_loss": 0.42347970604896545,
33
+ "eval_p": 0.822696467628716,
34
+ "eval_r": 0.8064903846153846,
35
+ "eval_runtime": 10.7242,
36
+ "eval_samples_per_second": 232.744,
37
+ "eval_steps_per_second": 14.546,
38
+ "step": 624
39
+ },
40
+ {
41
+ "epoch": 1.28,
42
+ "learning_rate": 4.5376078914919854e-05,
43
+ "loss": 0.4226,
44
+ "step": 800
45
+ },
46
+ {
47
+ "epoch": 1.6,
48
+ "learning_rate": 4.2293464858199754e-05,
49
+ "loss": 0.3854,
50
+ "step": 1000
51
+ },
52
+ {
53
+ "epoch": 1.92,
54
+ "learning_rate": 3.921085080147966e-05,
55
+ "loss": 0.3815,
56
+ "step": 1200
57
+ },
58
+ {
59
+ "epoch": 2.0,
60
+ "eval_f1": 0.8388272320074119,
61
+ "eval_loss": 0.3714698255062103,
62
+ "eval_p": 0.841085367336758,
63
+ "eval_r": 0.8449519230769231,
64
+ "eval_runtime": 10.8299,
65
+ "eval_samples_per_second": 230.472,
66
+ "eval_steps_per_second": 14.405,
67
+ "step": 1248
68
+ },
69
+ {
70
+ "epoch": 2.24,
71
+ "learning_rate": 3.6128236744759556e-05,
72
+ "loss": 0.3389,
73
+ "step": 1400
74
+ },
75
+ {
76
+ "epoch": 2.56,
77
+ "learning_rate": 3.304562268803946e-05,
78
+ "loss": 0.3105,
79
+ "step": 1600
80
+ },
81
+ {
82
+ "epoch": 2.88,
83
+ "learning_rate": 2.996300863131936e-05,
84
+ "loss": 0.3029,
85
+ "step": 1800
86
+ },
87
+ {
88
+ "epoch": 3.0,
89
+ "eval_f1": 0.8441213935683678,
90
+ "eval_loss": 0.37199097871780396,
91
+ "eval_p": 0.8445166019046795,
92
+ "eval_r": 0.84375,
93
+ "eval_runtime": 10.7362,
94
+ "eval_samples_per_second": 232.484,
95
+ "eval_steps_per_second": 14.53,
96
+ "step": 1872
97
+ },
98
+ {
99
+ "epoch": 3.21,
100
+ "learning_rate": 2.688039457459926e-05,
101
+ "loss": 0.2631,
102
+ "step": 2000
103
+ },
104
+ {
105
+ "epoch": 3.53,
106
+ "learning_rate": 2.3797780517879162e-05,
107
+ "loss": 0.2392,
108
+ "step": 2200
109
+ },
110
+ {
111
+ "epoch": 3.85,
112
+ "learning_rate": 2.0715166461159063e-05,
113
+ "loss": 0.2474,
114
+ "step": 2400
115
+ },
116
+ {
117
+ "epoch": 4.0,
118
+ "eval_f1": 0.8537410092673323,
119
+ "eval_loss": 0.37799614667892456,
120
+ "eval_p": 0.8537701589320283,
121
+ "eval_r": 0.8537660256410257,
122
+ "eval_runtime": 10.7429,
123
+ "eval_samples_per_second": 232.34,
124
+ "eval_steps_per_second": 14.521,
125
+ "step": 2496
126
+ },
127
+ {
128
+ "epoch": 4.17,
129
+ "learning_rate": 1.7632552404438964e-05,
130
+ "loss": 0.2055,
131
+ "step": 2600
132
+ },
133
+ {
134
+ "epoch": 4.49,
135
+ "learning_rate": 1.4549938347718866e-05,
136
+ "loss": 0.1797,
137
+ "step": 2800
138
+ },
139
+ {
140
+ "epoch": 4.81,
141
+ "learning_rate": 1.1467324290998767e-05,
142
+ "loss": 0.1757,
143
+ "step": 3000
144
+ },
145
+ {
146
+ "epoch": 5.0,
147
+ "eval_f1": 0.8502410363258266,
148
+ "eval_loss": 0.5066409707069397,
149
+ "eval_p": 0.8539723028554528,
150
+ "eval_r": 0.8485576923076923,
151
+ "eval_runtime": 10.7461,
152
+ "eval_samples_per_second": 232.269,
153
+ "eval_steps_per_second": 14.517,
154
+ "step": 3120
155
+ }
156
+ ],
157
+ "logging_steps": 200,
158
+ "max_steps": 3744,
159
+ "num_input_tokens_seen": 0,
160
+ "num_train_epochs": 6,
161
+ "save_steps": 500,
162
+ "total_flos": 6608254028682240.0,
163
+ "train_batch_size": 16,
164
+ "trial_name": null,
165
+ "trial_params": null
166
+ }
distilbert-base-cased-ft-BUILD/checkpoint-3120/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ae357d8ada26e9e9b973b4b3652af41f8e13ef19e617413bfdef1e140acfd572
3
+ size 4792
distilbert-base-cased-ft-BUILD/distilbert-base-cased-ft-BUILD-best/config.json ADDED
@@ -0,0 +1,36 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "distilbert-base-cased",
3
+ "activation": "gelu",
4
+ "architectures": [
5
+ "DistilBertForSequenceClassification"
6
+ ],
7
+ "attention_dropout": 0.1,
8
+ "dim": 768,
9
+ "dropout": 0.1,
10
+ "hidden_dim": 3072,
11
+ "id2label": {
12
+ "0": "Entr\u00e9e",
13
+ "1": "Plat principal",
14
+ "2": "Dessert"
15
+ },
16
+ "initializer_range": 0.02,
17
+ "label2id": {
18
+ "Dessert": 2,
19
+ "Entr\u00e9e": 0,
20
+ "Plat principal": 1
21
+ },
22
+ "max_position_embeddings": 512,
23
+ "model_type": "distilbert",
24
+ "n_heads": 12,
25
+ "n_layers": 6,
26
+ "output_past": true,
27
+ "pad_token_id": 0,
28
+ "problem_type": "single_label_classification",
29
+ "qa_dropout": 0.1,
30
+ "seq_classif_dropout": 0.2,
31
+ "sinusoidal_pos_embds": false,
32
+ "tie_weights_": true,
33
+ "torch_dtype": "float32",
34
+ "transformers_version": "4.36.1",
35
+ "vocab_size": 28996
36
+ }
distilbert-base-cased-ft-BUILD/distilbert-base-cased-ft-BUILD-best/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ebebcfae8ef9207504b86bf25933da79087ff9ea53cf25db77e4c854e7190746
3
+ size 263147764
distilbert-base-cased-ft-BUILD/distilbert-base-cased-ft-BUILD-best/special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
distilbert-base-cased-ft-BUILD/distilbert-base-cased-ft-BUILD-best/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
distilbert-base-cased-ft-BUILD/distilbert-base-cased-ft-BUILD-best/tokenizer_config.json ADDED
@@ -0,0 +1,55 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "100": {
12
+ "content": "[UNK]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "101": {
20
+ "content": "[CLS]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "102": {
28
+ "content": "[SEP]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "103": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "clean_up_tokenization_spaces": true,
45
+ "cls_token": "[CLS]",
46
+ "do_lower_case": false,
47
+ "mask_token": "[MASK]",
48
+ "model_max_length": 512,
49
+ "pad_token": "[PAD]",
50
+ "sep_token": "[SEP]",
51
+ "strip_accents": null,
52
+ "tokenize_chinese_chars": true,
53
+ "tokenizer_class": "DistilBertTokenizer",
54
+ "unk_token": "[UNK]"
55
+ }
distilbert-base-cased-ft-BUILD/distilbert-base-cased-ft-BUILD-best/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9a4c98641c0c92377fd9ae69d06b9f4baea0b2802dcb0ab1974894b57bb12353
3
+ size 4792
distilbert-base-cased-ft-BUILD/distilbert-base-cased-ft-BUILD-best/vocab.txt ADDED
The diff for this file is too large to render. See raw diff
 
distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709055246.DESKTOP-VOM7ARG.17955.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5db5a60ab930869a0380743583bb1f90cf4f93ce105ea1b2629e3b9aa8e36cdb
3
+ size 4939
distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709055584.DESKTOP-VOM7ARG.19083.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:23ac74ae8e44e435e383d5b26c5122979fb2a93d83c0982306c4a055342f4228
3
+ size 4939
distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709055671.DESKTOP-VOM7ARG.19446.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:921b6c050ae8eeb363fa91cd08f88d52545e20ba896672ed34767fd63d43b48f
3
+ size 4939
distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709055714.DESKTOP-VOM7ARG.19697.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0935a20d9ba84cc7d9bf8e35c617923d0f9063fae00ab1bfae44e9d57d914f65
3
+ size 4939
distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709055822.DESKTOP-VOM7ARG.20135.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ef0be61d09f03b34d918c75ff88913565c1500abaa093de29cf7d58ab634677c
3
+ size 4939
distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709056245.DESKTOP-VOM7ARG.762.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:906c7603626a5e8cc8e9493a36bf674d4d6b623e9bb42b759a7ff3e527e0b7c8
3
+ size 4939
distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709056422.DESKTOP-VOM7ARG.1601.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e46c66f4e7e23d928e65cb24c329cb60d67d86732fd8bc0dcdcc2a874c384398
3
+ size 4939
distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709056587.DESKTOP-VOM7ARG.2226.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3b3af101cad7ed8a545698ba6169e8cbd4d9001845a7af61bafc0ab4398be483
3
+ size 7927
distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709061581.DESKTOP-VOM7ARG.16950.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c1dabde6a72e69d155aa076d1236a9098340c9ea2a14efe5d829876ec47e18e9
3
+ size 10561
distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709069376.DESKTOP-VOM7ARG.39969.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3d9574088e90779d5d8bf0001caefa4a1966a15d8fbcfa9e4ae5512266452f71
3
+ size 4545
distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709069409.DESKTOP-VOM7ARG.40166.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:76f55194729e28eececd3d7b8277e06b7e0194c0385cd15ac792e30b338cb77f
3
+ size 4545
distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709069468.DESKTOP-VOM7ARG.40470.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:257caba1515e99871cb5a101ff9b64d926ae7784d1028abc9d207865c5987c64
3
+ size 4545
distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709069539.DESKTOP-VOM7ARG.40789.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:81314967f07de88e7c9ace94f3fca0ec41d702147651936710e82c5d1c5c3a09
3
+ size 4939
distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709070129.DESKTOP-VOM7ARG.42636.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1bc58f60ee608b3db87308510611476d4e3dff01d351e566e4a273aaf7872f9c
3
+ size 4559
distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709070260.DESKTOP-VOM7ARG.43131.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a0dfede442d3bf163c881f1384c8daea4b72a8989dbb235d813c2b0986e39e84
3
+ size 4565
distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709070283.DESKTOP-VOM7ARG.43300.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5e4426ead234a12ef0c91beed610e1bcabcf525c288f8d9ed1ecffd74133678b
3
+ size 4535
distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709070425.DESKTOP-VOM7ARG.43861.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:28c5f5dd618291e77bdfdc4f764ac3d2c1583d516d3874a3d6a249cb2a2234a7
3
+ size 4535
distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1709070475.DESKTOP-VOM7ARG.44142.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9bc317be60ce43259954518fc857e129c055868d15b5a460eedb5c3b1e7b824d
3
+ size 10191
distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1710266349.DESKTOP-VOM7ARG.31210.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f31c9ae80d85f653f461c6ca1c100d8d1652d6109c2595f89354b3e23323df3e
3
+ size 4570
distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1710266663.DESKTOP-VOM7ARG.34038.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b1198236de54436455f37a9c60ce148c9f4d9e41e05d85dbbb3ac41edf2cef3a
3
+ size 17530
ingredients/distilbert-base-cased-ft-BUILD/checkpoint-1872/config.json ADDED
@@ -0,0 +1,36 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "distilbert-base-cased",
3
+ "activation": "gelu",
4
+ "architectures": [
5
+ "DistilBertForSequenceClassification"
6
+ ],
7
+ "attention_dropout": 0.1,
8
+ "dim": 768,
9
+ "dropout": 0.1,
10
+ "hidden_dim": 3072,
11
+ "id2label": {
12
+ "0": "Entr\u00e9e",
13
+ "1": "Plat principal",
14
+ "2": "Dessert"
15
+ },
16
+ "initializer_range": 0.02,
17
+ "label2id": {
18
+ "Dessert": 2,
19
+ "Entr\u00e9e": 0,
20
+ "Plat principal": 1
21
+ },
22
+ "max_position_embeddings": 512,
23
+ "model_type": "distilbert",
24
+ "n_heads": 12,
25
+ "n_layers": 6,
26
+ "output_past": true,
27
+ "pad_token_id": 0,
28
+ "problem_type": "single_label_classification",
29
+ "qa_dropout": 0.1,
30
+ "seq_classif_dropout": 0.2,
31
+ "sinusoidal_pos_embds": false,
32
+ "tie_weights_": true,
33
+ "torch_dtype": "float32",
34
+ "transformers_version": "4.36.1",
35
+ "vocab_size": 28996
36
+ }
ingredients/distilbert-base-cased-ft-BUILD/checkpoint-1872/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f946d23df2edf87b47814770bed590c07aa4b98ced92381ae2bde5c72f31463c
3
+ size 263147764
ingredients/distilbert-base-cased-ft-BUILD/checkpoint-1872/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:926c308bdbdab946b05bf732ab2e5bb2192fe99ae6476c4fcf8492b11d0871db
3
+ size 526357690
ingredients/distilbert-base-cased-ft-BUILD/checkpoint-1872/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:899bf7ce686ec08a8b70dd92747507b3b430f9484252777f2df648363b93a1f9
3
+ size 14244
ingredients/distilbert-base-cased-ft-BUILD/checkpoint-1872/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:96c94514ae676650b0a1c947f585caa3a3b0bb1010bdc06d92e7e19432a456cd
3
+ size 1064
ingredients/distilbert-base-cased-ft-BUILD/checkpoint-1872/trainer_state.json ADDED
@@ -0,0 +1,108 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.8201436140182747,
3
+ "best_model_checkpoint": "./models/distilbertcheffo/ingredients/distilbert-base-cased-ft-BUILD/checkpoint-1248",
4
+ "epoch": 3.0,
5
+ "eval_steps": 500,
6
+ "global_step": 1872,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.32,
13
+ "learning_rate": 2e-05,
14
+ "loss": 0.8064,
15
+ "step": 200
16
+ },
17
+ {
18
+ "epoch": 0.64,
19
+ "learning_rate": 4e-05,
20
+ "loss": 0.5328,
21
+ "step": 400
22
+ },
23
+ {
24
+ "epoch": 0.96,
25
+ "learning_rate": 4.958263772954925e-05,
26
+ "loss": 0.491,
27
+ "step": 600
28
+ },
29
+ {
30
+ "epoch": 1.0,
31
+ "eval_f1": 0.7625179981925605,
32
+ "eval_loss": 0.4677772521972656,
33
+ "eval_p": 0.7920354842384022,
34
+ "eval_r": 0.7564102564102564,
35
+ "eval_runtime": 10.7329,
36
+ "eval_samples_per_second": 232.557,
37
+ "eval_steps_per_second": 14.535,
38
+ "step": 624
39
+ },
40
+ {
41
+ "epoch": 1.28,
42
+ "learning_rate": 4.8747913188647746e-05,
43
+ "loss": 0.4444,
44
+ "step": 800
45
+ },
46
+ {
47
+ "epoch": 1.6,
48
+ "learning_rate": 4.791318864774624e-05,
49
+ "loss": 0.4212,
50
+ "step": 1000
51
+ },
52
+ {
53
+ "epoch": 1.92,
54
+ "learning_rate": 4.7078464106844744e-05,
55
+ "loss": 0.4269,
56
+ "step": 1200
57
+ },
58
+ {
59
+ "epoch": 2.0,
60
+ "eval_f1": 0.8201436140182747,
61
+ "eval_loss": 0.4101373255252838,
62
+ "eval_p": 0.8195197123847887,
63
+ "eval_r": 0.8241185897435898,
64
+ "eval_runtime": 10.7229,
65
+ "eval_samples_per_second": 232.772,
66
+ "eval_steps_per_second": 14.548,
67
+ "step": 1248
68
+ },
69
+ {
70
+ "epoch": 2.24,
71
+ "learning_rate": 4.624373956594324e-05,
72
+ "loss": 0.3915,
73
+ "step": 1400
74
+ },
75
+ {
76
+ "epoch": 2.56,
77
+ "learning_rate": 4.540901502504174e-05,
78
+ "loss": 0.3516,
79
+ "step": 1600
80
+ },
81
+ {
82
+ "epoch": 2.88,
83
+ "learning_rate": 4.457429048414024e-05,
84
+ "loss": 0.3638,
85
+ "step": 1800
86
+ },
87
+ {
88
+ "epoch": 3.0,
89
+ "eval_f1": 0.8160936815512525,
90
+ "eval_loss": 0.42037826776504517,
91
+ "eval_p": 0.8239885163327829,
92
+ "eval_r": 0.8265224358974359,
93
+ "eval_runtime": 10.7428,
94
+ "eval_samples_per_second": 232.341,
95
+ "eval_steps_per_second": 14.521,
96
+ "step": 1872
97
+ }
98
+ ],
99
+ "logging_steps": 200,
100
+ "max_steps": 12480,
101
+ "num_input_tokens_seen": 0,
102
+ "num_train_epochs": 20,
103
+ "save_steps": 500,
104
+ "total_flos": 3964952417209344.0,
105
+ "train_batch_size": 16,
106
+ "trial_name": null,
107
+ "trial_params": null
108
+ }
ingredients/distilbert-base-cased-ft-BUILD/checkpoint-1872/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ae5d9e5738c101bc04aa4a0dc3590dc468dd7a97882afbc922e96e1027023bfd
3
+ size 4856
ingredients/distilbert-base-cased-ft-BUILD/distilbert-base-cased-ft-BUILD-best/config.json ADDED
@@ -0,0 +1,36 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "distilbert-base-cased",
3
+ "activation": "gelu",
4
+ "architectures": [
5
+ "DistilBertForSequenceClassification"
6
+ ],
7
+ "attention_dropout": 0.1,
8
+ "dim": 768,
9
+ "dropout": 0.1,
10
+ "hidden_dim": 3072,
11
+ "id2label": {
12
+ "0": "Entr\u00e9e",
13
+ "1": "Plat principal",
14
+ "2": "Dessert"
15
+ },
16
+ "initializer_range": 0.02,
17
+ "label2id": {
18
+ "Dessert": 2,
19
+ "Entr\u00e9e": 0,
20
+ "Plat principal": 1
21
+ },
22
+ "max_position_embeddings": 512,
23
+ "model_type": "distilbert",
24
+ "n_heads": 12,
25
+ "n_layers": 6,
26
+ "output_past": true,
27
+ "pad_token_id": 0,
28
+ "problem_type": "single_label_classification",
29
+ "qa_dropout": 0.1,
30
+ "seq_classif_dropout": 0.2,
31
+ "sinusoidal_pos_embds": false,
32
+ "tie_weights_": true,
33
+ "torch_dtype": "float32",
34
+ "transformers_version": "4.36.1",
35
+ "vocab_size": 28996
36
+ }
ingredients/distilbert-base-cased-ft-BUILD/distilbert-base-cased-ft-BUILD-best/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7f8a991280469d06f9edf8203a5b9613afe79fd6aedc23b255c26b5add0bf312
3
+ size 263147764
ingredients/distilbert-base-cased-ft-BUILD/distilbert-base-cased-ft-BUILD-best/special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
ingredients/distilbert-base-cased-ft-BUILD/distilbert-base-cased-ft-BUILD-best/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
ingredients/distilbert-base-cased-ft-BUILD/distilbert-base-cased-ft-BUILD-best/tokenizer_config.json ADDED
@@ -0,0 +1,55 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "100": {
12
+ "content": "[UNK]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "101": {
20
+ "content": "[CLS]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "102": {
28
+ "content": "[SEP]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "103": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "clean_up_tokenization_spaces": true,
45
+ "cls_token": "[CLS]",
46
+ "do_lower_case": false,
47
+ "mask_token": "[MASK]",
48
+ "model_max_length": 512,
49
+ "pad_token": "[PAD]",
50
+ "sep_token": "[SEP]",
51
+ "strip_accents": null,
52
+ "tokenize_chinese_chars": true,
53
+ "tokenizer_class": "DistilBertTokenizer",
54
+ "unk_token": "[UNK]"
55
+ }
ingredients/distilbert-base-cased-ft-BUILD/distilbert-base-cased-ft-BUILD-best/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ae5d9e5738c101bc04aa4a0dc3590dc468dd7a97882afbc922e96e1027023bfd
3
+ size 4856
ingredients/distilbert-base-cased-ft-BUILD/distilbert-base-cased-ft-BUILD-best/vocab.txt ADDED
The diff for this file is too large to render. See raw diff
 
ingredients/distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1710279134.DESKTOP-VOM7ARG.62013.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:52beba5db112e49134740a73aded280dfd5941d1c596f0feca1b4c89ec6fea44
3
+ size 4606
ingredients/distilbert-base-cased-ft-BUILD/logs/events.out.tfevents.1710279165.DESKTOP-VOM7ARG.62362.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9d3a0170e6b7dab2f08fbd513e1c734583f0f7090fd3c1127f13d8fe2b967908
3
+ size 4606