dima806 commited on
Commit
c7e3ed8
·
verified ·
1 Parent(s): 7cdaa37

Upload folder using huggingface_hub

Browse files
checkpoint-912/config.json ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "dima806/faces_age_detection",
3
+ "architectures": [
4
+ "ViTForImageClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.0,
7
+ "encoder_stride": 16,
8
+ "hidden_act": "gelu",
9
+ "hidden_dropout_prob": 0.0,
10
+ "hidden_size": 768,
11
+ "id2label": {
12
+ "0": "MIDDLE",
13
+ "1": "YOUNG",
14
+ "2": "OLD"
15
+ },
16
+ "image_size": 224,
17
+ "initializer_range": 0.02,
18
+ "intermediate_size": 3072,
19
+ "label2id": {
20
+ "MIDDLE": 0,
21
+ "OLD": 2,
22
+ "YOUNG": 1
23
+ },
24
+ "layer_norm_eps": 1e-12,
25
+ "model_type": "vit",
26
+ "num_attention_heads": 12,
27
+ "num_channels": 3,
28
+ "num_hidden_layers": 12,
29
+ "patch_size": 16,
30
+ "problem_type": "single_label_classification",
31
+ "qkv_bias": true,
32
+ "torch_dtype": "float32",
33
+ "transformers_version": "4.38.2"
34
+ }
checkpoint-912/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:daecfa46675787b1707ed8ee020c8ada43c783d760c6e8ac8b638e873803193b
3
+ size 343227052
checkpoint-912/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:deafc652d1c2096eee8fc29971ce1c82924608a80fb1f934a7218ccb4f82e4ac
3
+ size 686574597
checkpoint-912/preprocessor_config.json ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "do_normalize": true,
3
+ "do_rescale": true,
4
+ "do_resize": true,
5
+ "image_mean": [
6
+ 0.5,
7
+ 0.5,
8
+ 0.5
9
+ ],
10
+ "image_processor_type": "ViTImageProcessor",
11
+ "image_std": [
12
+ 0.5,
13
+ 0.5,
14
+ 0.5
15
+ ],
16
+ "resample": 2,
17
+ "rescale_factor": 0.00392156862745098,
18
+ "size": {
19
+ "height": 224,
20
+ "width": 224
21
+ }
22
+ }
checkpoint-912/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7ab7a476f0032958e9c7e8d34df7dde53a0b9a115ea5b228f44725ab43d5d475
3
+ size 14575
checkpoint-912/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4f0de2bbeecf7bd3293549b52bdcf87c2f95e29370683e41275b9b480ad91aba
3
+ size 627
checkpoint-912/trainer_state.json ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.25590604543685913,
3
+ "best_model_checkpoint": "faces_age_detection/checkpoint-912",
4
+ "epoch": 2.0,
5
+ "eval_steps": 500,
6
+ "global_step": 912,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 1.0,
13
+ "eval_accuracy": 0.9012954966070327,
14
+ "eval_loss": 0.26023536920547485,
15
+ "eval_runtime": 28.2604,
16
+ "eval_samples_per_second": 114.719,
17
+ "eval_steps_per_second": 3.609,
18
+ "step": 456
19
+ },
20
+ {
21
+ "epoch": 1.1,
22
+ "grad_norm": 6.117819786071777,
23
+ "learning_rate": 4.779582366589327e-07,
24
+ "loss": 0.1137,
25
+ "step": 500
26
+ },
27
+ {
28
+ "epoch": 2.0,
29
+ "eval_accuracy": 0.9031462060456509,
30
+ "eval_loss": 0.25590604543685913,
31
+ "eval_runtime": 28.5203,
32
+ "eval_samples_per_second": 113.673,
33
+ "eval_steps_per_second": 3.576,
34
+ "step": 912
35
+ }
36
+ ],
37
+ "logging_steps": 500,
38
+ "max_steps": 912,
39
+ "num_input_tokens_seen": 0,
40
+ "num_train_epochs": 2,
41
+ "save_steps": 500,
42
+ "total_flos": 4.5209231933567386e+18,
43
+ "train_batch_size": 64,
44
+ "trial_name": null,
45
+ "trial_params": null
46
+ }
checkpoint-912/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3c29334dc4de288a7c3f093e596c8e19c0643c2ab9b32aa95c2c7e1069ebc8f5
3
+ size 4411
config.json CHANGED
@@ -30,5 +30,5 @@
30
  "problem_type": "single_label_classification",
31
  "qkv_bias": true,
32
  "torch_dtype": "float32",
33
- "transformers_version": "4.36.1"
34
  }
 
30
  "problem_type": "single_label_classification",
31
  "qkv_bias": true,
32
  "torch_dtype": "float32",
33
+ "transformers_version": "4.38.2"
34
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5d3a7c9d3c224dfc90448da0b1a9555f833fa9c1980b2cc388f42adc627da263
3
  size 343227052
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:daecfa46675787b1707ed8ee020c8ada43c783d760c6e8ac8b638e873803193b
3
  size 343227052
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9bd72a040a5273c5a59914a678aa821a7674b174f7cddfede528074dbb317864
3
- size 4219
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3c29334dc4de288a7c3f093e596c8e19c0643c2ab9b32aa95c2c7e1069ebc8f5
3
+ size 4411