nicolasrl commited on
Commit
0ca651d
·
verified ·
1 Parent(s): 6ac7c5d

Añadir modelo y artefactos de la ejecución: n_2025_W41_8v45i4

Browse files
n_2025_W41_8v45i4/checkpoint-6376/config.json ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "ViTForImageClassification"
4
+ ],
5
+ "attention_probs_dropout_prob": 0.0,
6
+ "dtype": "float32",
7
+ "encoder_stride": 16,
8
+ "hidden_act": "gelu",
9
+ "hidden_dropout_prob": 0.0,
10
+ "hidden_size": 768,
11
+ "id2label": {
12
+ "0": "real",
13
+ "1": "semisynthetic",
14
+ "2": "synthetic"
15
+ },
16
+ "image_size": 224,
17
+ "initializer_range": 0.02,
18
+ "intermediate_size": 3072,
19
+ "label2id": {
20
+ "real": 0,
21
+ "semisynthetic": 1,
22
+ "synthetic": 2
23
+ },
24
+ "layer_norm_eps": 1e-12,
25
+ "model_type": "vit",
26
+ "num_attention_heads": 12,
27
+ "num_channels": 3,
28
+ "num_hidden_layers": 12,
29
+ "patch_size": 16,
30
+ "pooler_act": "tanh",
31
+ "pooler_output_size": 768,
32
+ "problem_type": "single_label_classification",
33
+ "qkv_bias": true,
34
+ "transformers_version": "4.57.0"
35
+ }
n_2025_W41_8v45i4/checkpoint-6376/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:efae89443336cf52f66b43a7fe5c1ad1d2fa6bf4727bed11442ca2ade3252b5b
3
+ size 343227052
n_2025_W41_8v45i4/checkpoint-6376/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:11f10fb7bde6e88e4158ece009f7d40c3cf01f7dbbfe746004153fc64bd5bd47
3
+ size 340332474
n_2025_W41_8v45i4/checkpoint-6376/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2ce51d3cff3db58624fe44160f319bbf674f3e9ac7a037e9b2e2af975902220a
3
+ size 14244
n_2025_W41_8v45i4/checkpoint-6376/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8852261b3dcdb00324bbc72e1edd5c2600b590675be3379ac274ef0fa9d37af6
3
+ size 1064
n_2025_W41_8v45i4/checkpoint-6376/trainer_state.json ADDED
@@ -0,0 +1,147 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": 6376,
3
+ "best_metric": 0.1828499585390091,
4
+ "best_model_checkpoint": "df-vs-real/checkpoint-6376",
5
+ "epoch": 2.0,
6
+ "eval_steps": 500,
7
+ "global_step": 6376,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.15683814303638646,
14
+ "grad_norm": 2.2149393558502197,
15
+ "learning_rate": 1.972960752396455e-05,
16
+ "loss": 0.7685,
17
+ "step": 500
18
+ },
19
+ {
20
+ "epoch": 0.3136762860727729,
21
+ "grad_norm": 3.101264476776123,
22
+ "learning_rate": 1.927744619280159e-05,
23
+ "loss": 0.3474,
24
+ "step": 1000
25
+ },
26
+ {
27
+ "epoch": 0.47051442910915936,
28
+ "grad_norm": 1.8101170063018799,
29
+ "learning_rate": 1.8825284861638635e-05,
30
+ "loss": 0.2921,
31
+ "step": 1500
32
+ },
33
+ {
34
+ "epoch": 0.6273525721455459,
35
+ "grad_norm": 2.5019593238830566,
36
+ "learning_rate": 1.8373123530475676e-05,
37
+ "loss": 0.2702,
38
+ "step": 2000
39
+ },
40
+ {
41
+ "epoch": 0.7841907151819323,
42
+ "grad_norm": 1.3309218883514404,
43
+ "learning_rate": 1.7920962199312717e-05,
44
+ "loss": 0.2366,
45
+ "step": 2500
46
+ },
47
+ {
48
+ "epoch": 0.9410288582183187,
49
+ "grad_norm": 1.73959481716156,
50
+ "learning_rate": 1.7468800868149758e-05,
51
+ "loss": 0.2292,
52
+ "step": 3000
53
+ },
54
+ {
55
+ "epoch": 1.0,
56
+ "eval_accuracy": 0.9268888888888889,
57
+ "eval_loss": 0.20106515288352966,
58
+ "eval_model_preparation_time": 0.0027,
59
+ "eval_runtime": 149.0356,
60
+ "eval_samples_per_second": 60.388,
61
+ "eval_steps_per_second": 3.778,
62
+ "step": 3188
63
+ },
64
+ {
65
+ "epoch": 1.0978670012547052,
66
+ "grad_norm": 3.891098737716675,
67
+ "learning_rate": 1.70166395369868e-05,
68
+ "loss": 0.2023,
69
+ "step": 3500
70
+ },
71
+ {
72
+ "epoch": 1.2547051442910915,
73
+ "grad_norm": 2.0172224044799805,
74
+ "learning_rate": 1.656447820582384e-05,
75
+ "loss": 0.1721,
76
+ "step": 4000
77
+ },
78
+ {
79
+ "epoch": 1.411543287327478,
80
+ "grad_norm": 5.1710944175720215,
81
+ "learning_rate": 1.611231687466088e-05,
82
+ "loss": 0.1662,
83
+ "step": 4500
84
+ },
85
+ {
86
+ "epoch": 1.5683814303638646,
87
+ "grad_norm": 0.4657936096191406,
88
+ "learning_rate": 1.566015554349792e-05,
89
+ "loss": 0.1758,
90
+ "step": 5000
91
+ },
92
+ {
93
+ "epoch": 1.725219573400251,
94
+ "grad_norm": 2.6389691829681396,
95
+ "learning_rate": 1.5207994212334962e-05,
96
+ "loss": 0.1716,
97
+ "step": 5500
98
+ },
99
+ {
100
+ "epoch": 1.8820577164366374,
101
+ "grad_norm": 1.282594919204712,
102
+ "learning_rate": 1.4755832881172003e-05,
103
+ "loss": 0.1708,
104
+ "step": 6000
105
+ },
106
+ {
107
+ "epoch": 2.0,
108
+ "eval_accuracy": 0.935,
109
+ "eval_loss": 0.1828499585390091,
110
+ "eval_model_preparation_time": 0.0027,
111
+ "eval_runtime": 149.5266,
112
+ "eval_samples_per_second": 60.19,
113
+ "eval_steps_per_second": 3.765,
114
+ "step": 6376
115
+ }
116
+ ],
117
+ "logging_steps": 500,
118
+ "max_steps": 22316,
119
+ "num_input_tokens_seen": 0,
120
+ "num_train_epochs": 7,
121
+ "save_steps": 500,
122
+ "stateful_callbacks": {
123
+ "EarlyStoppingCallback": {
124
+ "args": {
125
+ "early_stopping_patience": 3,
126
+ "early_stopping_threshold": 0.0
127
+ },
128
+ "attributes": {
129
+ "early_stopping_patience_counter": 0
130
+ }
131
+ },
132
+ "TrainerControl": {
133
+ "args": {
134
+ "should_epoch_stop": false,
135
+ "should_evaluate": false,
136
+ "should_log": false,
137
+ "should_save": true,
138
+ "should_training_stop": false
139
+ },
140
+ "attributes": {}
141
+ }
142
+ },
143
+ "total_flos": 7.904253783379968e+18,
144
+ "train_batch_size": 16,
145
+ "trial_name": null,
146
+ "trial_params": null
147
+ }
n_2025_W41_8v45i4/checkpoint-6376/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:05a408eb883a1b6e69f1edf724cb964521c5d499049fd4feed9e4bf218eaa59a
3
+ size 5368
n_2025_W41_8v45i4/confusion_matrix.png ADDED
n_2025_W41_8v45i4/model.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fb5b57951a9db934a12ebc9938e0af913a8fea0abb848cdeded446b31acb7833
3
+ size 343398125
n_2025_W41_8v45i4/onnx_validation_report.txt ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ === Reporte de Validación Cruzada (PyTorch vs ONNX) ===
2
+ Modelo: n_2025_W41_8v45i4
3
+ Fecha: 2025-10-13 00:33:13
4
+ =======================================================
5
+ Precisión del modelo PyTorch (Trainer): 0.9350
6
+ Precisión del modelo ONNX: 0.9350
7
+ Diferencia: 0.0000
8
+ =======================================================
9
+ Total de imágenes evaluadas: 9000
10
+ ✅ La precisión se mantiene consistente.
n_2025_W41_8v45i4/train_report.txt ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ === Reporte de Entrenamiento ===
2
+ Fecha y hora: 2025-10-13T00:01:46.437924
3
+
4
+ >> Métricas iniciales (baseline)
5
+ {
6
+ "eval_loss": 1.107085943222046,
7
+ "eval_model_preparation_time": 0.0027,
8
+ "eval_accuracy": 0.31166666666666665,
9
+ "eval_runtime": 193.5134,
10
+ "eval_samples_per_second": 46.508,
11
+ "eval_steps_per_second": 2.909
12
+ }
13
+
14
+ >> Métricas finales
15
+ {
16
+ "eval_loss": 0.1828499585390091,
17
+ "eval_model_preparation_time": 0.0027,
18
+ "eval_accuracy": 0.935,
19
+ "eval_runtime": 135.0182,
20
+ "eval_samples_per_second": 66.658,
21
+ "eval_steps_per_second": 4.17,
22
+ "epoch": 5.0
23
+ }
n_2025_W41_8v45i4/training_summary.txt ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ === Resumen de Entrenamiento ===
2
+ ID de ejecución: n_2025_W41_8v45i4
3
+ Modelo base: google/vit-base-patch16-224-in21k
4
+ Fecha: 2025-10-13 00:33:14
5
+
6
+ Total de imágenes en dataset: 60000
7
+ - Entrenamiento: 51000
8
+ - Prueba: 9000
9
+
10
+ Distribución de etiquetas:
11
+ - real: 20000
12
+ - semisynthetic: 20000
13
+ - synthetic: 20000