thaonguyen274 commited on
Commit
9119b4b
·
1 Parent(s): 86f928f

Training in progress, epoch 1

Browse files
all_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 10.0,
3
+ "total_flos": 4.769009900476416e+17,
4
+ "train_loss": 1.3194919480217828,
5
+ "train_runtime": 7630.326,
6
+ "train_samples_per_second": 2.942,
7
+ "train_steps_per_second": 0.012
8
+ }
config.json CHANGED
@@ -1,26 +1,13 @@
1
  {
2
  "_name_or_path": "google/vit-base-patch16-224",
3
  "architectures": [
4
- "ResNetForImageClassification"
5
  ],
6
  "attention_probs_dropout_prob": 0.0,
7
- "depths": [
8
- 3,
9
- 4,
10
- 6,
11
- 3
12
- ],
13
- "downsample_in_first_stage": false,
14
- "embedding_size": 64,
15
  "hidden_act": "gelu",
16
  "hidden_dropout_prob": 0.0,
17
  "hidden_size": 768,
18
- "hidden_sizes": [
19
- 256,
20
- 512,
21
- 1024,
22
- 2048
23
- ],
24
  "id2label": {
25
  "0": "buildings",
26
  "1": "forest",
@@ -41,8 +28,7 @@
41
  "street": 5
42
  },
43
  "layer_norm_eps": 1e-12,
44
- "layer_type": "bottleneck",
45
- "model_type": "resnet",
46
  "num_attention_heads": 12,
47
  "num_channels": 3,
48
  "num_hidden_layers": 12,
 
1
  {
2
  "_name_or_path": "google/vit-base-patch16-224",
3
  "architectures": [
4
+ "ViTForImageClassification"
5
  ],
6
  "attention_probs_dropout_prob": 0.0,
7
+ "encoder_stride": 16,
 
 
 
 
 
 
 
8
  "hidden_act": "gelu",
9
  "hidden_dropout_prob": 0.0,
10
  "hidden_size": 768,
 
 
 
 
 
 
11
  "id2label": {
12
  "0": "buildings",
13
  "1": "forest",
 
28
  "street": 5
29
  },
30
  "layer_norm_eps": 1e-12,
31
+ "model_type": "vit",
 
32
  "num_attention_heads": 12,
33
  "num_channels": 3,
34
  "num_hidden_layers": 12,
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4abac6db2cf7f131e44462640ba38f0af9da739cf0b62b02fca2cc5af7a773e0
3
- size 94405441
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c7ebf6795980ce7cc67e7dc73455c42937d82ed37f161d63e6d87f793f7b8efb
3
+ size 343279281
runs/Nov06_19-15-31_a77448885f3d/1667762138.3268747/events.out.tfevents.1667762138.a77448885f3d.77.3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c062c99830492c329dc2ffb42afec79225477fc4cec6d0f622b3f779114ba619
3
+ size 5571
runs/Nov06_19-15-31_a77448885f3d/events.out.tfevents.1667762138.a77448885f3d.77.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2e898b46ff7970509ecad2683ec73a68beca972ff66e5b0afd3598229fa5df88
3
+ size 4274
train_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 10.0,
3
+ "total_flos": 4.769009900476416e+17,
4
+ "train_loss": 1.3194919480217828,
5
+ "train_runtime": 7630.326,
6
+ "train_samples_per_second": 2.942,
7
+ "train_steps_per_second": 0.012
8
+ }
trainer_state.json ADDED
@@ -0,0 +1,169 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.5742596303718548,
3
+ "best_model_checkpoint": "vit-base-patch16-224-finetuned-imageclassification/checkpoint-90",
4
+ "epoch": 10.0,
5
+ "global_step": 90,
6
+ "is_hyper_param_search": false,
7
+ "is_local_process_zero": true,
8
+ "is_world_process_zero": true,
9
+ "log_history": [
10
+ {
11
+ "epoch": 1.0,
12
+ "eval_accuracy": 0.18147405922957024,
13
+ "eval_loss": 2.0663375854492188,
14
+ "eval_runtime": 5315.8097,
15
+ "eval_samples_per_second": 1.69,
16
+ "eval_steps_per_second": 0.027,
17
+ "step": 9
18
+ },
19
+ {
20
+ "epoch": 1.11,
21
+ "learning_rate": 9.876543209876543e-05,
22
+ "loss": 1.7049,
23
+ "step": 10
24
+ },
25
+ {
26
+ "epoch": 2.0,
27
+ "eval_accuracy": 0.34769539078156314,
28
+ "eval_loss": 1.633109211921692,
29
+ "eval_runtime": 62.6786,
30
+ "eval_samples_per_second": 143.302,
31
+ "eval_steps_per_second": 2.25,
32
+ "step": 18
33
+ },
34
+ {
35
+ "epoch": 2.22,
36
+ "learning_rate": 8.641975308641975e-05,
37
+ "loss": 1.5848,
38
+ "step": 20
39
+ },
40
+ {
41
+ "epoch": 3.0,
42
+ "eval_accuracy": 0.3907815631262525,
43
+ "eval_loss": 1.6548928022384644,
44
+ "eval_runtime": 59.6679,
45
+ "eval_samples_per_second": 150.533,
46
+ "eval_steps_per_second": 2.363,
47
+ "step": 27
48
+ },
49
+ {
50
+ "epoch": 3.33,
51
+ "learning_rate": 7.407407407407407e-05,
52
+ "loss": 1.4471,
53
+ "step": 30
54
+ },
55
+ {
56
+ "epoch": 4.0,
57
+ "eval_accuracy": 0.47873524827432645,
58
+ "eval_loss": 1.3122586011886597,
59
+ "eval_runtime": 64.0678,
60
+ "eval_samples_per_second": 140.195,
61
+ "eval_steps_per_second": 2.201,
62
+ "step": 36
63
+ },
64
+ {
65
+ "epoch": 4.44,
66
+ "learning_rate": 6.17283950617284e-05,
67
+ "loss": 1.3214,
68
+ "step": 40
69
+ },
70
+ {
71
+ "epoch": 5.0,
72
+ "eval_accuracy": 0.5012246715653529,
73
+ "eval_loss": 1.2882213592529297,
74
+ "eval_runtime": 63.7848,
75
+ "eval_samples_per_second": 140.817,
76
+ "eval_steps_per_second": 2.211,
77
+ "step": 45
78
+ },
79
+ {
80
+ "epoch": 5.56,
81
+ "learning_rate": 4.938271604938271e-05,
82
+ "loss": 1.2432,
83
+ "step": 50
84
+ },
85
+ {
86
+ "epoch": 6.0,
87
+ "eval_accuracy": 0.5526608773101759,
88
+ "eval_loss": 1.1481624841690063,
89
+ "eval_runtime": 63.6972,
90
+ "eval_samples_per_second": 141.011,
91
+ "eval_steps_per_second": 2.214,
92
+ "step": 54
93
+ },
94
+ {
95
+ "epoch": 6.67,
96
+ "learning_rate": 3.7037037037037037e-05,
97
+ "loss": 1.1932,
98
+ "step": 60
99
+ },
100
+ {
101
+ "epoch": 7.0,
102
+ "eval_accuracy": 0.5577822311289246,
103
+ "eval_loss": 1.1259204149246216,
104
+ "eval_runtime": 64.3997,
105
+ "eval_samples_per_second": 139.473,
106
+ "eval_steps_per_second": 2.189,
107
+ "step": 63
108
+ },
109
+ {
110
+ "epoch": 7.78,
111
+ "learning_rate": 2.4691358024691357e-05,
112
+ "loss": 1.1336,
113
+ "step": 70
114
+ },
115
+ {
116
+ "epoch": 8.0,
117
+ "eval_accuracy": 0.56312625250501,
118
+ "eval_loss": 1.1054401397705078,
119
+ "eval_runtime": 60.6277,
120
+ "eval_samples_per_second": 148.15,
121
+ "eval_steps_per_second": 2.326,
122
+ "step": 72
123
+ },
124
+ {
125
+ "epoch": 8.89,
126
+ "learning_rate": 1.2345679012345678e-05,
127
+ "loss": 1.1301,
128
+ "step": 80
129
+ },
130
+ {
131
+ "epoch": 9.0,
132
+ "eval_accuracy": 0.5655755956357159,
133
+ "eval_loss": 1.0953378677368164,
134
+ "eval_runtime": 63.7804,
135
+ "eval_samples_per_second": 140.827,
136
+ "eval_steps_per_second": 2.211,
137
+ "step": 81
138
+ },
139
+ {
140
+ "epoch": 10.0,
141
+ "learning_rate": 0.0,
142
+ "loss": 1.1172,
143
+ "step": 90
144
+ },
145
+ {
146
+ "epoch": 10.0,
147
+ "eval_accuracy": 0.5742596303718548,
148
+ "eval_loss": 1.0771620273590088,
149
+ "eval_runtime": 64.1958,
150
+ "eval_samples_per_second": 139.916,
151
+ "eval_steps_per_second": 2.196,
152
+ "step": 90
153
+ },
154
+ {
155
+ "epoch": 10.0,
156
+ "step": 90,
157
+ "total_flos": 4.769009900476416e+17,
158
+ "train_loss": 1.3194919480217828,
159
+ "train_runtime": 7630.326,
160
+ "train_samples_per_second": 2.942,
161
+ "train_steps_per_second": 0.012
162
+ }
163
+ ],
164
+ "max_steps": 90,
165
+ "num_train_epochs": 10,
166
+ "total_flos": 4.769009900476416e+17,
167
+ "trial_name": null,
168
+ "trial_params": null
169
+ }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ffd89709a7eb60b1ea5efb67fa48dac5cf8beb3bf9f47f61506ffc0eb51632d9
3
  size 3439
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2d3e4e86369a939107e7ddeda00ff61e05c6d3a5dee77cf209e7104ea10ea3d4
3
  size 3439