AbstractPhil commited on
Commit
4e2b11d
·
verified ·
1 Parent(s): 6c1ee55

Update training_history.json - Run 20251012_161107

Browse files
weights/David-partial_shared-deep_efficiency/20251012_161107/training_history.json CHANGED
@@ -4,28 +4,36 @@
4
  2,
5
  3,
6
  4,
7
- 5
 
 
8
  ],
9
  "train_loss": [
10
  2.090034657678665,
11
  1.2265579131083748,
12
  1.0678428365304447,
13
  0.9437485171583133,
14
- 0.8316704105740538
 
 
15
  ],
16
  "train_acc": [
17
  81.03081019102116,
18
  86.79532020415762,
19
  88.322443522195,
20
  89.53789786967663,
21
- 90.7392244726878
 
 
22
  ],
23
  "val_acc": [
24
  83.714,
25
  84.042,
26
  84.372,
27
  84.414,
28
- 84.578
 
 
29
  ],
30
  "scale_accs": {
31
  "384": [
@@ -33,56 +41,72 @@
33
  83.85,
34
  84.262,
35
  84.314,
36
- 84.636
 
 
37
  ],
38
  "512": [
39
  83.54,
40
  83.942,
41
  84.382,
42
  84.448,
43
- 84.67
 
 
44
  ],
45
  "768": [
46
  83.714,
47
  84.042,
48
  84.372,
49
  84.414,
50
- 84.578
 
 
51
  ],
52
  "1024": [
53
  83.484,
54
  84.12,
55
  84.464,
56
  84.44,
57
- 84.402
 
 
58
  ],
59
  "1280": [
60
  83.63,
61
  83.998,
62
  84.362,
63
  84.276,
64
- 84.516
 
 
65
  ],
66
  "1536": [
67
  83.592,
68
  84.212,
69
  84.324,
70
  84.37,
71
- 84.41
 
 
72
  ],
73
  "1792": [
74
  83.622,
75
  83.96,
76
  84.282,
77
  84.326,
78
- 84.406
 
 
79
  ],
80
  "2048": [
81
  83.664,
82
  84.082,
83
  84.258,
84
  84.416,
85
- 84.38
 
 
86
  ]
87
  },
88
  "lr": [
@@ -90,6 +114,8 @@
90
  0.0009046039886902864,
91
  0.0007940987335200904,
92
  0.0006548539886902864,
93
- 0.0005005000000000001
 
 
94
  ]
95
  }
 
4
  2,
5
  3,
6
  4,
7
+ 5,
8
+ 6,
9
+ 7
10
  ],
11
  "train_loss": [
12
  2.090034657678665,
13
  1.2265579131083748,
14
  1.0678428365304447,
15
  0.9437485171583133,
16
+ 0.8316704105740538,
17
+ 0.7262678437244398,
18
+ 0.6332142021233281
19
  ],
20
  "train_acc": [
21
  81.03081019102116,
22
  86.79532020415762,
23
  88.322443522195,
24
  89.53789786967663,
25
+ 90.7392244726878,
26
+ 91.92845273098668,
27
+ 93.03353895315755
28
  ],
29
  "val_acc": [
30
  83.714,
31
  84.042,
32
  84.372,
33
  84.414,
34
+ 84.578,
35
+ 84.512,
36
+ 84.586
37
  ],
38
  "scale_accs": {
39
  "384": [
 
41
  83.85,
42
  84.262,
43
  84.314,
44
+ 84.636,
45
+ 84.464,
46
+ 84.672
47
  ],
48
  "512": [
49
  83.54,
50
  83.942,
51
  84.382,
52
  84.448,
53
+ 84.67,
54
+ 84.622,
55
+ 84.618
56
  ],
57
  "768": [
58
  83.714,
59
  84.042,
60
  84.372,
61
  84.414,
62
+ 84.578,
63
+ 84.512,
64
+ 84.586
65
  ],
66
  "1024": [
67
  83.484,
68
  84.12,
69
  84.464,
70
  84.44,
71
+ 84.402,
72
+ 84.508,
73
+ 84.368
74
  ],
75
  "1280": [
76
  83.63,
77
  83.998,
78
  84.362,
79
  84.276,
80
+ 84.516,
81
+ 84.346,
82
+ 84.366
83
  ],
84
  "1536": [
85
  83.592,
86
  84.212,
87
  84.324,
88
  84.37,
89
+ 84.41,
90
+ 84.436,
91
+ 84.46
92
  ],
93
  "1792": [
94
  83.622,
95
  83.96,
96
  84.282,
97
  84.326,
98
+ 84.406,
99
+ 84.276,
100
+ 84.254
101
  ],
102
  "2048": [
103
  83.664,
104
  84.082,
105
  84.258,
106
  84.416,
107
+ 84.38,
108
+ 84.396,
109
+ 84.364
110
  ]
111
  },
112
  "lr": [
 
114
  0.0009046039886902864,
115
  0.0007940987335200904,
116
  0.0006548539886902864,
117
+ 0.0005005000000000001,
118
+ 0.0003461460113097139,
119
+ 0.00020690126647990973
120
  ]
121
  }