AbstractPhil commited on
Commit
2698e4b
·
verified ·
1 Parent(s): 3676c1f

Update training_history.json - Run 20251012_161107

Browse files
weights/David-partial_shared-deep_efficiency/20251012_161107/training_history.json CHANGED
@@ -6,7 +6,9 @@
6
  4,
7
  5,
8
  6,
9
- 7
 
 
10
  ],
11
  "train_loss": [
12
  2.090034657678665,
@@ -15,7 +17,9 @@
15
  0.9437485171583133,
16
  0.8316704105740538,
17
  0.7262678437244398,
18
- 0.6332142021233281
 
 
19
  ],
20
  "train_acc": [
21
  81.03081019102116,
@@ -24,7 +28,9 @@
24
  89.53789786967663,
25
  90.7392244726878,
26
  91.92845273098668,
27
- 93.03353895315755
 
 
28
  ],
29
  "val_acc": [
30
  83.714,
@@ -33,7 +39,9 @@
33
  84.414,
34
  84.578,
35
  84.512,
36
- 84.586
 
 
37
  ],
38
  "scale_accs": {
39
  "384": [
@@ -43,7 +51,9 @@
43
  84.314,
44
  84.636,
45
  84.464,
46
- 84.672
 
 
47
  ],
48
  "512": [
49
  83.54,
@@ -52,7 +62,9 @@
52
  84.448,
53
  84.67,
54
  84.622,
55
- 84.618
 
 
56
  ],
57
  "768": [
58
  83.714,
@@ -61,7 +73,9 @@
61
  84.414,
62
  84.578,
63
  84.512,
64
- 84.586
 
 
65
  ],
66
  "1024": [
67
  83.484,
@@ -70,7 +84,9 @@
70
  84.44,
71
  84.402,
72
  84.508,
73
- 84.368
 
 
74
  ],
75
  "1280": [
76
  83.63,
@@ -79,7 +95,9 @@
79
  84.276,
80
  84.516,
81
  84.346,
82
- 84.366
 
 
83
  ],
84
  "1536": [
85
  83.592,
@@ -88,7 +106,9 @@
88
  84.37,
89
  84.41,
90
  84.436,
91
- 84.46
 
 
92
  ],
93
  "1792": [
94
  83.622,
@@ -97,7 +117,9 @@
97
  84.326,
98
  84.406,
99
  84.276,
100
- 84.254
 
 
101
  ],
102
  "2048": [
103
  83.664,
@@ -106,7 +128,9 @@
106
  84.416,
107
  84.38,
108
  84.396,
109
- 84.364
 
 
110
  ]
111
  },
112
  "lr": [
@@ -116,6 +140,8 @@
116
  0.0006548539886902864,
117
  0.0005005000000000001,
118
  0.0003461460113097139,
119
- 0.00020690126647990973
 
 
120
  ]
121
  }
 
6
  4,
7
  5,
8
  6,
9
+ 7,
10
+ 8,
11
+ 9
12
  ],
13
  "train_loss": [
14
  2.090034657678665,
 
17
  0.9437485171583133,
18
  0.8316704105740538,
19
  0.7262678437244398,
20
+ 0.6332142021233281,
21
+ 0.5566336322849551,
22
+ 0.5015984917648684
23
  ],
24
  "train_acc": [
25
  81.03081019102116,
 
28
  89.53789786967663,
29
  90.7392244726878,
30
  91.92845273098668,
31
+ 93.03353895315755,
32
+ 93.96557981902437,
33
+ 94.6770405419434
34
  ],
35
  "val_acc": [
36
  83.714,
 
39
  84.414,
40
  84.578,
41
  84.512,
42
+ 84.586,
43
+ 84.56,
44
+ 84.608
45
  ],
46
  "scale_accs": {
47
  "384": [
 
51
  84.314,
52
  84.636,
53
  84.464,
54
+ 84.672,
55
+ 84.65,
56
+ 84.744
57
  ],
58
  "512": [
59
  83.54,
 
62
  84.448,
63
  84.67,
64
  84.622,
65
+ 84.618,
66
+ 84.654,
67
+ 84.71
68
  ],
69
  "768": [
70
  83.714,
 
73
  84.414,
74
  84.578,
75
  84.512,
76
+ 84.586,
77
+ 84.56,
78
+ 84.608
79
  ],
80
  "1024": [
81
  83.484,
 
84
  84.44,
85
  84.402,
86
  84.508,
87
+ 84.368,
88
+ 84.352,
89
+ 84.366
90
  ],
91
  "1280": [
92
  83.63,
 
95
  84.276,
96
  84.516,
97
  84.346,
98
+ 84.366,
99
+ 84.244,
100
+ 84.352
101
  ],
102
  "1536": [
103
  83.592,
 
106
  84.37,
107
  84.41,
108
  84.436,
109
+ 84.46,
110
+ 84.292,
111
+ 84.274
112
  ],
113
  "1792": [
114
  83.622,
 
117
  84.326,
118
  84.406,
119
  84.276,
120
+ 84.254,
121
+ 84.216,
122
+ 84.186
123
  ],
124
  "2048": [
125
  83.664,
 
128
  84.416,
129
  84.38,
130
  84.396,
131
+ 84.364,
132
+ 84.306,
133
+ 84.286
134
  ]
135
  },
136
  "lr": [
 
140
  0.0006548539886902864,
141
  0.0005005000000000001,
142
  0.0003461460113097139,
143
+ 0.00020690126647990973,
144
+ 9.639601130971382e-05,
145
+ 2.5447270110570814e-05
146
  ]
147
  }