File size: 7,482 Bytes
5d2288c
b74334b
5d2288c
 
b74334b
f8e8e62
5d2288c
8fd6486
c246cf9
520ec9b
c246cf9
8fd6486
e9eff31
85c225f
b74334b
e9eff31
d192d2f
b74334b
 
8fd6486
0147bb9
d0687eb
8fd6486
 
 
b74334b
8fd6486
c246cf9
8fd6486
b74334b
 
8fd6486
 
e9eff31
8fd6486
9c6ee37
e9eff31
b74334b
c23e229
b74334b
5d2288c
930b4ae
af34624
8fd6486
 
b74334b
a53ea9e
8fd6486
e9eff31
b74334b
8fd6486
 
4b59816
b74334b
d0687eb
8fd6486
c246cf9
8fd6486
af34624
b74334b
e9eff31
b74334b
5d2288c
d483c1d
e9eff31
b74334b
 
c23e229
930b4ae
c246cf9
223465a
b74334b
 
 
 
9c6ee37
8fd6486
0147bb9
b74334b
947ef92
e9eff31
617fe9e
9c6ee37
8fd6486
5d2288c
b09d65f
 
8fd6486
85c225f
5d2288c
091b378
520ec9b
f8e8e62
9c6ee37
223465a
ddfb4f9
4c66597
8fd6486
d0687eb
9a39110
c246cf9
b74334b
8fd6486
5d2288c
8fd6486
 
 
e9eff31
c1d154c
b74334b
b09d65f
b74334b
8fd6486
947ef92
e9eff31
8fd6486
9a39110
7802bb4
b74334b
5a6f670
 
8fd6486
b74334b
af34624
5d2288c
b74334b
22fec13
ddfb4f9
b74334b
8fd6486
 
 
b74334b
 
c246cf9
8fd6486
 
22fec13
8fd6486
b74334b
 
091b378
9a39110
5d2288c
04707b8
5d2288c
c246cf9
2365584
b74334b
9a39110
8fd6486
551f1d4
27daaa6
0a911e4
22fec13
223465a
091b378
af34624
 
9c6ee37
930b4ae
8fd6486
b09d65f
b74334b
8fd6486
543cb39
8fd6486
b74334b
d0687eb
b74334b
 
d0687eb
e9eff31
b74334b
 
9c6ee37
8fd6486
930b4ae
8fd6486
930b4ae
22fec13
0147bb9
8fd6486
 
b74334b
0147bb9
5d2288c
930b4ae
af34624
b74334b
4c66597
b74334b
0147bb9
b09d65f
8fd6486
b74334b
648f8cc
520ec9b
8fd6486
27daaa6
ddfb4f9
0147bb9
af34624
8fd6486
b74334b
22fec13
520ec9b
8fd6486
b74334b
a53ea9e
f161c6a
b74334b
ae627ee
af34624
d0687eb
e9eff31
b09d65f
8fd6486
 
 
22fec13
b74334b
f8e8e62
85c225f
930b4ae
22fec13
8fd6486
930b4ae
8fd6486
 
b74334b
d0687eb
b74334b
e9eff31
8fd6486
f8e8e62
e9eff31
f8e8e62
8fd6486
 
0a911e4
8fd6486
b74334b
 
af34624
b74334b
0147bb9
ad528ff
b74334b
b09d65f
4b59816
0147bb9
2365584
e9eff31
85c225f
8fd6486
b74334b
947ef92
4c66597
5d2288c
 
 
 
 
 
 
39c9403
 
5d2288c
 
 
 
 
 
39c9403
5d2288c
8fd6486
5d2288c
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
{
  "_name_or_path": "distributed/llama-1b",
  "all_reduce_scores": {
    "0": "NOT_ALIVE",
    "1": "NOT_ALIVE",
    "10": "NOT_ALIVE",
    "100": "NOT_ALIVE",
    "101": "NON_PARTICIPATING",
    "102": "NON_PARTICIPATING",
    "103": "NON_PARTICIPATING",
    "104": "NOT_ALIVE",
    "105": "NOT_ALIVE",
    "106": "NOT_ALIVE",
    "107": "NOT_ALIVE",
    "108": "NON_PARTICIPATING",
    "109": "NOT_ALIVE",
    "11": "NOT_ALIVE",
    "110": "NOT_ALIVE",
    "111": "NOT_ALIVE",
    "112": "NON_PARTICIPATING",
    "113": "SUCCESS",
    "114": "NON_PARTICIPATING",
    "115": "NOT_ALIVE",
    "116": "NON_PARTICIPATING",
    "117": "SUCCESS",
    "118": "NOT_ALIVE",
    "119": "NON_PARTICIPATING",
    "12": "NOT_ALIVE",
    "120": "NON_PARTICIPATING",
    "121": "NOT_ALIVE",
    "122": "NOT_ALIVE",
    "123": "NOT_ALIVE",
    "124": "SUCCESS",
    "125": "NON_PARTICIPATING",
    "126": "NON_PARTICIPATING",
    "127": "NOT_ALIVE",
    "128": "NON_PARTICIPATING",
    "129": "NOT_ALIVE",
    "13": "NOT_ALIVE",
    "130": "NOT_ALIVE",
    "131": "NOT_ALIVE",
    "132": "NOT_ALIVE",
    "133": "NON_PARTICIPATING",
    "134": "NON_PARTICIPATING",
    "135": "NON_PARTICIPATING",
    "136": "NON_PARTICIPATING",
    "137": "NON_PARTICIPATING",
    "138": "SUCCESS",
    "139": "NON_PARTICIPATING",
    "14": "SUCCESS",
    "140": "NOT_ALIVE",
    "141": "NOT_ALIVE",
    "142": "NON_PARTICIPATING",
    "143": "NOT_ALIVE",
    "144": "NOT_ALIVE",
    "145": "NOT_ALIVE",
    "146": "NOT_ALIVE",
    "147": "NOT_ALIVE",
    "148": "SUCCESS",
    "149": "NOT_ALIVE",
    "15": "SUCCESS",
    "150": "NOT_ALIVE",
    "151": "NOT_ALIVE",
    "152": "NOT_ALIVE",
    "153": "NON_PARTICIPATING",
    "154": "SUCCESS",
    "155": "NOT_ALIVE",
    "156": "NOT_ALIVE",
    "157": "NOT_ALIVE",
    "158": "NOT_ALIVE",
    "159": "NOT_ALIVE",
    "16": "NON_PARTICIPATING",
    "160": "NOT_ALIVE",
    "161": "NON_PARTICIPATING",
    "162": "NON_PARTICIPATING",
    "163": "NOT_ALIVE",
    "164": "SUCCESS",
    "165": "SUCCESS",
    "166": "NOT_ALIVE",
    "167": "NON_PARTICIPATING",
    "168": "SUCCESS",
    "169": "NOT_ALIVE",
    "17": "NOT_ALIVE",
    "170": "NOT_ALIVE",
    "171": "NOT_ALIVE",
    "172": "NON_PARTICIPATING",
    "173": "NON_PARTICIPATING",
    "174": "NOT_ALIVE",
    "175": "NOT_ALIVE",
    "176": "NOT_ALIVE",
    "177": "NOT_ALIVE",
    "178": "NOT_ALIVE",
    "179": "NOT_ALIVE",
    "18": "NOT_ALIVE",
    "180": "NOT_ALIVE",
    "181": "NOT_ALIVE",
    "182": "NOT_ALIVE",
    "183": "NON_PARTICIPATING",
    "184": "NON_PARTICIPATING",
    "185": "NOT_ALIVE",
    "186": "NOT_ALIVE",
    "187": "NOT_ALIVE",
    "188": "NOT_ALIVE",
    "189": "NOT_ALIVE",
    "19": "NON_PARTICIPATING",
    "190": "NON_PARTICIPATING",
    "191": "NON_PARTICIPATING",
    "192": "NOT_ALIVE",
    "193": "NOT_ALIVE",
    "194": "NON_PARTICIPATING",
    "195": "NON_PARTICIPATING",
    "196": "NON_PARTICIPATING",
    "197": "NOT_ALIVE",
    "198": "NON_PARTICIPATING",
    "199": "NON_PARTICIPATING",
    "2": "NON_PARTICIPATING",
    "20": "NOT_ALIVE",
    "200": "NOT_ALIVE",
    "201": "NON_PARTICIPATING",
    "202": "NOT_ALIVE",
    "203": "NOT_ALIVE",
    "204": "NON_PARTICIPATING",
    "205": "NOT_ALIVE",
    "206": "SUCCESS",
    "207": "NOT_ALIVE",
    "208": "NOT_ALIVE",
    "209": "NOT_ALIVE",
    "21": "NOT_ALIVE",
    "210": "NOT_ALIVE",
    "211": "NOT_ALIVE",
    "212": "SUCCESS",
    "213": "SUCCESS",
    "214": "NOT_ALIVE",
    "215": "SUCCESS",
    "216": "NOT_ALIVE",
    "217": "NON_PARTICIPATING",
    "218": "NON_PARTICIPATING",
    "219": "NOT_ALIVE",
    "22": "NON_PARTICIPATING",
    "220": "NOT_ALIVE",
    "221": "NOT_ALIVE",
    "222": "NOT_ALIVE",
    "223": "NOT_ALIVE",
    "224": "NOT_ALIVE",
    "225": "NOT_ALIVE",
    "226": "NOT_ALIVE",
    "227": "NOT_ALIVE",
    "228": "NON_PARTICIPATING",
    "229": "NON_PARTICIPATING",
    "23": "NON_PARTICIPATING",
    "230": "NOT_ALIVE",
    "231": "NOT_ALIVE",
    "232": "NOT_ALIVE",
    "233": "NOT_ALIVE",
    "234": "NOT_ALIVE",
    "235": "NON_PARTICIPATING",
    "236": "NOT_ALIVE",
    "237": "NON_PARTICIPATING",
    "238": "NON_PARTICIPATING",
    "239": "NOT_ALIVE",
    "24": "NOT_ALIVE",
    "240": "NON_PARTICIPATING",
    "241": "NON_PARTICIPATING",
    "242": "NOT_ALIVE",
    "243": "NOT_ALIVE",
    "244": "NOT_ALIVE",
    "245": "NOT_ALIVE",
    "246": "NON_PARTICIPATING",
    "247": "NON_PARTICIPATING",
    "248": "NOT_ALIVE",
    "249": "NOT_ALIVE",
    "25": "SUCCESS",
    "250": "NON_PARTICIPATING",
    "251": "NOT_ALIVE",
    "252": "NON_PARTICIPATING",
    "253": "NOT_ALIVE",
    "254": "NON_PARTICIPATING",
    "255": "NOT_ALIVE",
    "26": "NOT_ALIVE",
    "27": "NOT_ALIVE",
    "28": "NOT_ALIVE",
    "29": "NON_PARTICIPATING",
    "3": "NON_PARTICIPATING",
    "30": "NON_PARTICIPATING",
    "31": "NON_PARTICIPATING",
    "32": "SUCCESS",
    "33": "NOT_ALIVE",
    "34": "NOT_ALIVE",
    "35": "NON_PARTICIPATING",
    "36": "NOT_ALIVE",
    "37": "NOT_ALIVE",
    "38": "NON_PARTICIPATING",
    "39": "SUCCESS",
    "4": "NON_PARTICIPATING",
    "40": "NON_PARTICIPATING",
    "41": "NOT_ALIVE",
    "42": "NOT_ALIVE",
    "43": "NOT_ALIVE",
    "44": "NOT_ALIVE",
    "45": "NOT_ALIVE",
    "46": "NOT_ALIVE",
    "47": "NON_PARTICIPATING",
    "48": "NON_PARTICIPATING",
    "49": "NOT_ALIVE",
    "5": "SUCCESS",
    "50": "NOT_ALIVE",
    "51": "NOT_ALIVE",
    "52": "SUCCESS",
    "53": "NOT_ALIVE",
    "54": "NOT_ALIVE",
    "55": "NOT_ALIVE",
    "56": "NOT_ALIVE",
    "57": "NON_PARTICIPATING",
    "58": "NON_PARTICIPATING",
    "59": "NON_PARTICIPATING",
    "6": "NON_PARTICIPATING",
    "60": "NOT_ALIVE",
    "61": "NOT_ALIVE",
    "62": "NON_PARTICIPATING",
    "63": "NON_PARTICIPATING",
    "64": "NOT_ALIVE",
    "65": "NOT_ALIVE",
    "66": "NOT_ALIVE",
    "67": "NOT_ALIVE",
    "68": "NOT_ALIVE",
    "69": "NOT_ALIVE",
    "7": "NON_PARTICIPATING",
    "70": "NOT_ALIVE",
    "71": "NON_PARTICIPATING",
    "72": "NOT_ALIVE",
    "73": "NOT_ALIVE",
    "74": "NOT_ALIVE",
    "75": "NON_PARTICIPATING",
    "76": "NOT_ALIVE",
    "77": "NON_PARTICIPATING",
    "78": "NOT_ALIVE",
    "79": "NON_PARTICIPATING",
    "8": "NOT_ALIVE",
    "80": "NOT_ALIVE",
    "81": "NOT_ALIVE",
    "82": "NON_PARTICIPATING",
    "83": "NON_PARTICIPATING",
    "84": "NOT_ALIVE",
    "85": "NOT_ALIVE",
    "86": "NON_PARTICIPATING",
    "87": "NOT_ALIVE",
    "88": "NOT_ALIVE",
    "89": "NOT_ALIVE",
    "9": "NOT_ALIVE",
    "90": "NOT_ALIVE",
    "91": "NOT_ALIVE",
    "92": "NON_PARTICIPATING",
    "93": "NOT_ALIVE",
    "94": "NON_PARTICIPATING",
    "95": "NOT_ALIVE",
    "96": "NON_PARTICIPATING",
    "97": "NOT_ALIVE",
    "98": "NOT_ALIVE",
    "99": "NOT_ALIVE"
  },
  "architectures": [
    "LlamaForCausalLM"
  ],
  "attention_bias": false,
  "attention_dropout": 0.0,
  "block_list": [
    6008401,
    6008441
  ],
  "bos_token_id": 1,
  "eos_token_id": 2,
  "hidden_act": "silu",
  "hidden_size": 2048,
  "initializer_range": 0.02,
  "inner_step": 40,
  "intermediate_size": 5632,
  "last_allreduce_block": 6005880,
  "max_position_embeddings": 2048,
  "mlp_bias": false,
  "model_type": "llama",
  "num_attention_heads": 32,
  "num_hidden_layers": 22,
  "num_key_value_heads": 4,
  "pretraining_tp": 1,
  "rms_norm_eps": 1e-05,
  "rope_scaling": null,
  "rope_theta": 10000.0,
  "tie_word_embeddings": false,
  "torch_dtype": "float32",
  "transformers_version": "4.39.3",
  "use_cache": false,
  "vocab_size": 32000
}