hf-transformers-bot commited on
Commit
34a3e30
·
verified ·
1 Parent(s): 8175ec7

Upload 2026-01-12/runs/26534-20918795329/ci_results_run_models_gpu/model_results.json with huggingface_hub

Browse files
2026-01-12/runs/26534-20918795329/ci_results_run_models_gpu/model_results.json ADDED
@@ -0,0 +1,1183 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "models_afmoe": {
3
+ "failed": {
4
+ "PyTorch": {
5
+ "unclassified": 0,
6
+ "single": 2,
7
+ "multi": 2
8
+ },
9
+ "Tokenizers": {
10
+ "unclassified": 0,
11
+ "single": 0,
12
+ "multi": 0
13
+ },
14
+ "Pipelines": {
15
+ "unclassified": 0,
16
+ "single": 0,
17
+ "multi": 0
18
+ },
19
+ "Trainer": {
20
+ "unclassified": 0,
21
+ "single": 0,
22
+ "multi": 0
23
+ },
24
+ "ONNX": {
25
+ "unclassified": 0,
26
+ "single": 0,
27
+ "multi": 0
28
+ },
29
+ "Auto": {
30
+ "unclassified": 0,
31
+ "single": 0,
32
+ "multi": 0
33
+ },
34
+ "Quantization": {
35
+ "unclassified": 0,
36
+ "single": 0,
37
+ "multi": 0
38
+ },
39
+ "Unclassified": {
40
+ "unclassified": 0,
41
+ "single": 0,
42
+ "multi": 0
43
+ }
44
+ },
45
+ "errors": 0,
46
+ "success": 217,
47
+ "skipped": 225,
48
+ "time_spent": [
49
+ 116.29,
50
+ 117.27
51
+ ],
52
+ "error": false,
53
+ "failures": {
54
+ "single": [
55
+ {
56
+ "line": "tests/models/afmoe/test_modeling_afmoe.py::AfmoeModelTest::test_flex_attention_with_grads",
57
+ "trace": "(line 619) torch._inductor.exc.InductorError: RuntimeError: No valid triton configs. OutOfMemoryError: out of resource: triton_tem_fused_0 Required: 147968 Hardware limit:101376 Reducing block sizes or `num_stages` may help."
58
+ },
59
+ {
60
+ "line": "tests/models/afmoe/test_modeling_afmoe.py::AfmoeModelTest::test_sdpa_can_dispatch_on_flash",
61
+ "trace": "(line 92) RuntimeError: No available kernel. Aborting execution."
62
+ }
63
+ ],
64
+ "multi": [
65
+ {
66
+ "line": "tests/models/afmoe/test_modeling_afmoe.py::AfmoeModelTest::test_flex_attention_with_grads",
67
+ "trace": "(line 619) torch._inductor.exc.InductorError: RuntimeError: No valid triton configs. OutOfMemoryError: out of resource: triton_tem_fused_0 Required: 147968 Hardware limit:101376 Reducing block sizes or `num_stages` may help."
68
+ },
69
+ {
70
+ "line": "tests/models/afmoe/test_modeling_afmoe.py::AfmoeModelTest::test_sdpa_can_dispatch_on_flash",
71
+ "trace": "(line 92) RuntimeError: No available kernel. Aborting execution."
72
+ }
73
+ ]
74
+ },
75
+ "job_link": {
76
+ "single": "https://github.com/huggingface/transformers/actions/runs/20918795329/job/60099266307",
77
+ "multi": "https://github.com/huggingface/transformers/actions/runs/20918795329/job/60099266291"
78
+ },
79
+ "captured_info": {}
80
+ },
81
+ "models_apertus": {
82
+ "failed": {
83
+ "PyTorch": {
84
+ "unclassified": 0,
85
+ "single": 0,
86
+ "multi": 0
87
+ },
88
+ "Tokenizers": {
89
+ "unclassified": 0,
90
+ "single": 0,
91
+ "multi": 0
92
+ },
93
+ "Pipelines": {
94
+ "unclassified": 0,
95
+ "single": 0,
96
+ "multi": 0
97
+ },
98
+ "Trainer": {
99
+ "unclassified": 0,
100
+ "single": 0,
101
+ "multi": 0
102
+ },
103
+ "ONNX": {
104
+ "unclassified": 0,
105
+ "single": 0,
106
+ "multi": 0
107
+ },
108
+ "Auto": {
109
+ "unclassified": 0,
110
+ "single": 0,
111
+ "multi": 0
112
+ },
113
+ "Quantization": {
114
+ "unclassified": 0,
115
+ "single": 0,
116
+ "multi": 0
117
+ },
118
+ "Unclassified": {
119
+ "unclassified": 0,
120
+ "single": 0,
121
+ "multi": 0
122
+ }
123
+ },
124
+ "errors": 0,
125
+ "success": 249,
126
+ "skipped": 197,
127
+ "time_spent": [
128
+ 99.6,
129
+ 99.65
130
+ ],
131
+ "error": false,
132
+ "failures": {},
133
+ "job_link": {
134
+ "single": "https://github.com/huggingface/transformers/actions/runs/20918795329/job/60099266333",
135
+ "multi": "https://github.com/huggingface/transformers/actions/runs/20918795329/job/60099266311"
136
+ },
137
+ "captured_info": {
138
+ "single": "https://github.com/huggingface/transformers/actions/runs/20918795329/job/60099266333#step:16:1",
139
+ "multi": "https://github.com/huggingface/transformers/actions/runs/20918795329/job/60099266311#step:16:1"
140
+ }
141
+ },
142
+ "models_arcee": {
143
+ "failed": {
144
+ "PyTorch": {
145
+ "unclassified": 0,
146
+ "single": 0,
147
+ "multi": 0
148
+ },
149
+ "Tokenizers": {
150
+ "unclassified": 0,
151
+ "single": 0,
152
+ "multi": 0
153
+ },
154
+ "Pipelines": {
155
+ "unclassified": 0,
156
+ "single": 0,
157
+ "multi": 0
158
+ },
159
+ "Trainer": {
160
+ "unclassified": 0,
161
+ "single": 0,
162
+ "multi": 0
163
+ },
164
+ "ONNX": {
165
+ "unclassified": 0,
166
+ "single": 0,
167
+ "multi": 0
168
+ },
169
+ "Auto": {
170
+ "unclassified": 0,
171
+ "single": 0,
172
+ "multi": 0
173
+ },
174
+ "Quantization": {
175
+ "unclassified": 0,
176
+ "single": 0,
177
+ "multi": 0
178
+ },
179
+ "Unclassified": {
180
+ "unclassified": 0,
181
+ "single": 0,
182
+ "multi": 0
183
+ }
184
+ },
185
+ "errors": 0,
186
+ "success": 261,
187
+ "skipped": 193,
188
+ "time_spent": [
189
+ 151.07,
190
+ 148.87
191
+ ],
192
+ "error": false,
193
+ "failures": {},
194
+ "job_link": {
195
+ "multi": "https://github.com/huggingface/transformers/actions/runs/20918795329/job/60099266325",
196
+ "single": "https://github.com/huggingface/transformers/actions/runs/20918795329/job/60099266464"
197
+ },
198
+ "captured_info": {
199
+ "multi": "https://github.com/huggingface/transformers/actions/runs/20918795329/job/60099266325#step:16:1",
200
+ "single": "https://github.com/huggingface/transformers/actions/runs/20918795329/job/60099266464#step:16:1"
201
+ }
202
+ },
203
+ "models_aria": {
204
+ "failed": {
205
+ "PyTorch": {
206
+ "unclassified": 0,
207
+ "single": 0,
208
+ "multi": 0
209
+ },
210
+ "Tokenizers": {
211
+ "unclassified": 0,
212
+ "single": 0,
213
+ "multi": 0
214
+ },
215
+ "Pipelines": {
216
+ "unclassified": 0,
217
+ "single": 0,
218
+ "multi": 0
219
+ },
220
+ "Trainer": {
221
+ "unclassified": 0,
222
+ "single": 0,
223
+ "multi": 0
224
+ },
225
+ "ONNX": {
226
+ "unclassified": 0,
227
+ "single": 0,
228
+ "multi": 0
229
+ },
230
+ "Auto": {
231
+ "unclassified": 0,
232
+ "single": 0,
233
+ "multi": 0
234
+ },
235
+ "Quantization": {
236
+ "unclassified": 0,
237
+ "single": 0,
238
+ "multi": 0
239
+ },
240
+ "Unclassified": {
241
+ "unclassified": 0,
242
+ "single": 0,
243
+ "multi": 0
244
+ }
245
+ },
246
+ "errors": 0,
247
+ "success": 305,
248
+ "skipped": 181,
249
+ "time_spent": [
250
+ 222.49,
251
+ 221.05
252
+ ],
253
+ "error": false,
254
+ "failures": {},
255
+ "job_link": {
256
+ "single": "https://github.com/huggingface/transformers/actions/runs/20918795329/job/60099266326",
257
+ "multi": "https://github.com/huggingface/transformers/actions/runs/20918795329/job/60099266320"
258
+ },
259
+ "captured_info": {}
260
+ },
261
+ "models_bamba": {
262
+ "failed": {
263
+ "PyTorch": {
264
+ "unclassified": 0,
265
+ "single": 3,
266
+ "multi": 3
267
+ },
268
+ "Tokenizers": {
269
+ "unclassified": 0,
270
+ "single": 0,
271
+ "multi": 0
272
+ },
273
+ "Pipelines": {
274
+ "unclassified": 0,
275
+ "single": 0,
276
+ "multi": 0
277
+ },
278
+ "Trainer": {
279
+ "unclassified": 0,
280
+ "single": 0,
281
+ "multi": 0
282
+ },
283
+ "ONNX": {
284
+ "unclassified": 0,
285
+ "single": 0,
286
+ "multi": 0
287
+ },
288
+ "Auto": {
289
+ "unclassified": 0,
290
+ "single": 0,
291
+ "multi": 0
292
+ },
293
+ "Quantization": {
294
+ "unclassified": 0,
295
+ "single": 0,
296
+ "multi": 0
297
+ },
298
+ "Unclassified": {
299
+ "unclassified": 0,
300
+ "single": 0,
301
+ "multi": 0
302
+ }
303
+ },
304
+ "errors": 0,
305
+ "success": 203,
306
+ "skipped": 223,
307
+ "time_spent": [
308
+ 73.82,
309
+ 75.08
310
+ ],
311
+ "error": false,
312
+ "failures": {
313
+ "single": [
314
+ {
315
+ "line": "tests/models/bamba/test_modeling_bamba.py::BambaModelTest::test_sdpa_can_compile_dynamic",
316
+ "trace": "(line 1677) torch._dynamo.exc.TorchRuntimeError: Dynamo failed to run FX node with fake tensors: call_function <built-in function scaled_dot_product_attention>(*(FakeTensor(..., device='cuda:0', size=(s47, 4, s87, 8), dtype=torch.float16), FakeTensor(..., device='cuda:0', size=(s47, 4, s87, 8), dtype=torch.float16), FakeTensor(..., device='cuda:0', size=(s47, 4, s87, 8), dtype=torch.float16)), **{'attn_mask': FakeTensor(..., device='cuda:0', size=(s84, 1, s34, s30), dtype=torch.float16), 'dropout_p': 0.0, 'scale': 0.353553390593274, 'is_causal': False}): got RuntimeError('expand: attempting to expand a dimension of length s30 -> s87!')"
317
+ },
318
+ {
319
+ "line": "tests/models/bamba/test_modeling_bamba.py::BambaModelIntegrationTest::test_simple_batched_generate_with_padding",
320
+ "trace": "(line 904) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 8.00 GiB. GPU 0 has a total capacity of 22.30 GiB of which 3.40 GiB is free. Process 129639 has 18.90 GiB memory in use. Of the allocated memory 18.54 GiB is allocated by PyTorch, and 4.28 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)"
321
+ },
322
+ {
323
+ "line": "tests/models/bamba/test_modeling_bamba.py::BambaModelIntegrationTest::test_simple_generate",
324
+ "trace": "(line 904) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 4.00 GiB. GPU 0 has a total capacity of 22.30 GiB of which 3.25 GiB is free. Process 129639 has 19.04 GiB memory in use. Of the allocated memory 18.67 GiB is allocated by PyTorch, and 11.38 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)"
325
+ }
326
+ ],
327
+ "multi": [
328
+ {
329
+ "line": "tests/models/bamba/test_modeling_bamba.py::BambaModelTest::test_sdpa_can_compile_dynamic",
330
+ "trace": "(line 1677) torch._dynamo.exc.TorchRuntimeError: Dynamo failed to run FX node with fake tensors: call_function <built-in function scaled_dot_product_attention>(*(FakeTensor(..., device='cuda:0', size=(s47, 4, s87, 8), dtype=torch.float16), FakeTensor(..., device='cuda:0', size=(s47, 4, s87, 8), dtype=torch.float16), FakeTensor(..., device='cuda:0', size=(s47, 4, s87, 8), dtype=torch.float16)), **{'attn_mask': FakeTensor(..., device='cuda:0', size=(s84, 1, s34, s30), dtype=torch.float16), 'dropout_p': 0.0, 'scale': 0.353553390593274, 'is_causal': False}): got RuntimeError('expand: attempting to expand a dimension of length s30 -> s87!')"
331
+ },
332
+ {
333
+ "line": "tests/models/bamba/test_modeling_bamba.py::BambaModelIntegrationTest::test_simple_batched_generate_with_padding",
334
+ "trace": "(line 904) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 8.00 GiB. GPU 0 has a total capacity of 22.30 GiB of which 3.29 GiB is free. Process 188090 has 19.00 GiB memory in use. Of the allocated memory 18.54 GiB is allocated by PyTorch, and 4.16 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)"
335
+ },
336
+ {
337
+ "line": "tests/models/bamba/test_modeling_bamba.py::BambaModelIntegrationTest::test_simple_generate",
338
+ "trace": "(line 904) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 4.00 GiB. GPU 0 has a total capacity of 22.30 GiB of which 3.15 GiB is free. Process 188090 has 19.15 GiB memory in use. Of the allocated memory 18.67 GiB is allocated by PyTorch, and 11.27 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)"
339
+ }
340
+ ]
341
+ },
342
+ "job_link": {
343
+ "single": "https://github.com/huggingface/transformers/actions/runs/20918795329/job/60099266334",
344
+ "multi": "https://github.com/huggingface/transformers/actions/runs/20918795329/job/60099266338"
345
+ },
346
+ "captured_info": {}
347
+ },
348
+ "models_cohere": {
349
+ "failed": {
350
+ "PyTorch": {
351
+ "unclassified": 0,
352
+ "single": 0,
353
+ "multi": 1
354
+ },
355
+ "Tokenizers": {
356
+ "unclassified": 0,
357
+ "single": 0,
358
+ "multi": 1
359
+ },
360
+ "Pipelines": {
361
+ "unclassified": 0,
362
+ "single": 0,
363
+ "multi": 0
364
+ },
365
+ "Trainer": {
366
+ "unclassified": 0,
367
+ "single": 0,
368
+ "multi": 0
369
+ },
370
+ "ONNX": {
371
+ "unclassified": 0,
372
+ "single": 0,
373
+ "multi": 0
374
+ },
375
+ "Auto": {
376
+ "unclassified": 0,
377
+ "single": 0,
378
+ "multi": 0
379
+ },
380
+ "Quantization": {
381
+ "unclassified": 0,
382
+ "single": 0,
383
+ "multi": 0
384
+ },
385
+ "Unclassified": {
386
+ "unclassified": 0,
387
+ "single": 0,
388
+ "multi": 0
389
+ }
390
+ },
391
+ "errors": 0,
392
+ "success": 349,
393
+ "skipped": 193,
394
+ "time_spent": [
395
+ 87.98,
396
+ 87.01
397
+ ],
398
+ "error": false,
399
+ "failures": {
400
+ "multi": [
401
+ {
402
+ "line": "tests/models/cohere/test_modeling_cohere.py::CohereIntegrationTest::test_batched_4bit",
403
+ "trace": "(line 483) OSError: You are trying to access a gated repo."
404
+ },
405
+ {
406
+ "line": "tests/models/cohere/test_tokenization_cohere.py::CohereTokenizationTest::test_torch_encode_plus_sent_to_model",
407
+ "trace": "(line 47) AttributeError: 'super' object has no attribute 'test_torch_encode_plus_sent_to_model'"
408
+ }
409
+ ]
410
+ },
411
+ "job_link": {
412
+ "multi": "https://github.com/huggingface/transformers/actions/runs/20918795329/job/60099266442",
413
+ "single": "https://github.com/huggingface/transformers/actions/runs/20918795329/job/60099266340"
414
+ },
415
+ "captured_info": {}
416
+ },
417
+ "models_cohere2": {
418
+ "failed": {
419
+ "PyTorch": {
420
+ "unclassified": 0,
421
+ "single": 1,
422
+ "multi": 1
423
+ },
424
+ "Tokenizers": {
425
+ "unclassified": 0,
426
+ "single": 0,
427
+ "multi": 0
428
+ },
429
+ "Pipelines": {
430
+ "unclassified": 0,
431
+ "single": 0,
432
+ "multi": 0
433
+ },
434
+ "Trainer": {
435
+ "unclassified": 0,
436
+ "single": 0,
437
+ "multi": 0
438
+ },
439
+ "ONNX": {
440
+ "unclassified": 0,
441
+ "single": 0,
442
+ "multi": 0
443
+ },
444
+ "Auto": {
445
+ "unclassified": 0,
446
+ "single": 0,
447
+ "multi": 0
448
+ },
449
+ "Quantization": {
450
+ "unclassified": 0,
451
+ "single": 0,
452
+ "multi": 0
453
+ },
454
+ "Unclassified": {
455
+ "unclassified": 0,
456
+ "single": 0,
457
+ "multi": 0
458
+ }
459
+ },
460
+ "errors": 0,
461
+ "success": 241,
462
+ "skipped": 197,
463
+ "time_spent": [
464
+ 191.89,
465
+ 184.85
466
+ ],
467
+ "error": false,
468
+ "failures": {
469
+ "single": [
470
+ {
471
+ "line": "tests/models/cohere2/test_modeling_cohere2.py::Cohere2IntegrationTest::test_model_pipeline_bf16",
472
+ "trace": "(line 193) torch.AcceleratorError: CUDA error: an illegal memory access was encountered"
473
+ }
474
+ ],
475
+ "multi": [
476
+ {
477
+ "line": "tests/models/cohere2/test_modeling_cohere2.py::Cohere2IntegrationTest::test_model_pipeline_bf16",
478
+ "trace": "(line 193) torch.AcceleratorError: CUDA error: an illegal memory access was encountered"
479
+ }
480
+ ]
481
+ },
482
+ "job_link": {
483
+ "single": "https://github.com/huggingface/transformers/actions/runs/20918795329/job/60099266390",
484
+ "multi": "https://github.com/huggingface/transformers/actions/runs/20918795329/job/60099266318"
485
+ },
486
+ "captured_info": {}
487
+ },
488
+ "models_cwm": {
489
+ "failed": {
490
+ "PyTorch": {
491
+ "unclassified": 0,
492
+ "single": 0,
493
+ "multi": 1
494
+ },
495
+ "Tokenizers": {
496
+ "unclassified": 0,
497
+ "single": 0,
498
+ "multi": 0
499
+ },
500
+ "Pipelines": {
501
+ "unclassified": 0,
502
+ "single": 0,
503
+ "multi": 0
504
+ },
505
+ "Trainer": {
506
+ "unclassified": 0,
507
+ "single": 0,
508
+ "multi": 0
509
+ },
510
+ "ONNX": {
511
+ "unclassified": 0,
512
+ "single": 0,
513
+ "multi": 0
514
+ },
515
+ "Auto": {
516
+ "unclassified": 0,
517
+ "single": 0,
518
+ "multi": 0
519
+ },
520
+ "Quantization": {
521
+ "unclassified": 0,
522
+ "single": 0,
523
+ "multi": 0
524
+ },
525
+ "Unclassified": {
526
+ "unclassified": 0,
527
+ "single": 0,
528
+ "multi": 0
529
+ }
530
+ },
531
+ "errors": 0,
532
+ "success": 268,
533
+ "skipped": 201,
534
+ "time_spent": [
535
+ 230.51,
536
+ 415.55
537
+ ],
538
+ "error": false,
539
+ "failures": {
540
+ "multi": [
541
+ {
542
+ "line": "tests/models/cwm/test_modeling_cwm.py::CwmIntegrationTest::test_cwm_sliding_window_long_sequence",
543
+ "trace": "(line 260) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 102.00 MiB. GPU 1 has a total capacity of 22.30 GiB of which 42.69 MiB is free. Process 168039 has 22.25 GiB memory in use. Of the allocated memory 21.82 GiB is allocated by PyTorch, and 24.19 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)"
544
+ }
545
+ ]
546
+ },
547
+ "job_link": {
548
+ "multi": "https://github.com/huggingface/transformers/actions/runs/20918795329/job/60099266330",
549
+ "single": "https://github.com/huggingface/transformers/actions/runs/20918795329/job/60099266418"
550
+ },
551
+ "captured_info": {
552
+ "multi": "https://github.com/huggingface/transformers/actions/runs/20918795329/job/60099266330#step:16:1",
553
+ "single": "https://github.com/huggingface/transformers/actions/runs/20918795329/job/60099266418#step:16:1"
554
+ }
555
+ },
556
+ "models_dbrx": {
557
+ "failed": {
558
+ "PyTorch": {
559
+ "unclassified": 0,
560
+ "single": 2,
561
+ "multi": 2
562
+ },
563
+ "Tokenizers": {
564
+ "unclassified": 0,
565
+ "single": 0,
566
+ "multi": 0
567
+ },
568
+ "Pipelines": {
569
+ "unclassified": 0,
570
+ "single": 0,
571
+ "multi": 0
572
+ },
573
+ "Trainer": {
574
+ "unclassified": 0,
575
+ "single": 0,
576
+ "multi": 0
577
+ },
578
+ "ONNX": {
579
+ "unclassified": 0,
580
+ "single": 0,
581
+ "multi": 0
582
+ },
583
+ "Auto": {
584
+ "unclassified": 0,
585
+ "single": 0,
586
+ "multi": 0
587
+ },
588
+ "Quantization": {
589
+ "unclassified": 0,
590
+ "single": 0,
591
+ "multi": 0
592
+ },
593
+ "Unclassified": {
594
+ "unclassified": 0,
595
+ "single": 0,
596
+ "multi": 0
597
+ }
598
+ },
599
+ "errors": 0,
600
+ "success": 225,
601
+ "skipped": 221,
602
+ "time_spent": [
603
+ 44.11,
604
+ 42.82
605
+ ],
606
+ "error": false,
607
+ "failures": {
608
+ "multi": [
609
+ {
610
+ "line": "tests/models/dbrx/test_modeling_dbrx.py::DbrxModelTest::test_model_from_pretrained",
611
+ "trace": "(line 248) RuntimeError: You set `ignore_mismatched_sizes` to `False`, thus raising an error. For details look at the above report!"
612
+ },
613
+ {
614
+ "line": "tests/models/dbrx/test_modeling_dbrx.py::DbrxModelIntegrationTest::test_tiny_model_logits",
615
+ "trace": "(line 248) RuntimeError: You set `ignore_mismatched_sizes` to `False`, thus raising an error. For details look at the above report!"
616
+ }
617
+ ],
618
+ "single": [
619
+ {
620
+ "line": "tests/models/dbrx/test_modeling_dbrx.py::DbrxModelTest::test_model_from_pretrained",
621
+ "trace": "(line 248) RuntimeError: You set `ignore_mismatched_sizes` to `False`, thus raising an error. For details look at the above report!"
622
+ },
623
+ {
624
+ "line": "tests/models/dbrx/test_modeling_dbrx.py::DbrxModelIntegrationTest::test_tiny_model_logits",
625
+ "trace": "(line 248) RuntimeError: You set `ignore_mismatched_sizes` to `False`, thus raising an error. For details look at the above report!"
626
+ }
627
+ ]
628
+ },
629
+ "job_link": {
630
+ "multi": "https://github.com/huggingface/transformers/actions/runs/20918795329/job/60099266424",
631
+ "single": "https://github.com/huggingface/transformers/actions/runs/20918795329/job/60099266374"
632
+ },
633
+ "captured_info": {
634
+ "multi": "https://github.com/huggingface/transformers/actions/runs/20918795329/job/60099266424#step:16:1",
635
+ "single": "https://github.com/huggingface/transformers/actions/runs/20918795329/job/60099266374#step:16:1"
636
+ }
637
+ },
638
+ "models_deepseek_v2": {
639
+ "failed": {
640
+ "PyTorch": {
641
+ "unclassified": 0,
642
+ "single": 6,
643
+ "multi": 6
644
+ },
645
+ "Tokenizers": {
646
+ "unclassified": 0,
647
+ "single": 0,
648
+ "multi": 0
649
+ },
650
+ "Pipelines": {
651
+ "unclassified": 0,
652
+ "single": 0,
653
+ "multi": 0
654
+ },
655
+ "Trainer": {
656
+ "unclassified": 0,
657
+ "single": 0,
658
+ "multi": 0
659
+ },
660
+ "ONNX": {
661
+ "unclassified": 0,
662
+ "single": 0,
663
+ "multi": 0
664
+ },
665
+ "Auto": {
666
+ "unclassified": 0,
667
+ "single": 0,
668
+ "multi": 0
669
+ },
670
+ "Quantization": {
671
+ "unclassified": 0,
672
+ "single": 0,
673
+ "multi": 0
674
+ },
675
+ "Unclassified": {
676
+ "unclassified": 0,
677
+ "single": 0,
678
+ "multi": 0
679
+ }
680
+ },
681
+ "errors": 0,
682
+ "success": 239,
683
+ "skipped": 201,
684
+ "time_spent": [
685
+ 334.72,
686
+ 550.3
687
+ ],
688
+ "error": false,
689
+ "failures": {
690
+ "single": [
691
+ {
692
+ "line": "tests/models/deepseek_v2/test_modeling_deepseek_v2.py::DeepseekV2ModelTest::test_cpu_offload",
693
+ "trace": "(line 2495) AssertionError: Tensor-likes are not close!"
694
+ },
695
+ {
696
+ "line": "tests/models/deepseek_v2/test_modeling_deepseek_v2.py::DeepseekV2ModelTest::test_disk_offload_bin",
697
+ "trace": "(line 2413) AssertionError: Tensor-likes are not close!"
698
+ },
699
+ {
700
+ "line": "tests/models/deepseek_v2/test_modeling_deepseek_v2.py::DeepseekV2ModelTest::test_disk_offload_safetensors",
701
+ "trace": "(line 2453) AssertionError: Tensor-likes are not close!"
702
+ },
703
+ {
704
+ "line": "tests/models/deepseek_v2/test_modeling_deepseek_v2.py::DeepseekV2IntegrationTest::test_batch_fa2",
705
+ "trace": "(line 243) RuntimeError: We encountered some issues during automatic conversion of the weights. For details look at the `CONVERSION` entries of the above report!"
706
+ },
707
+ {
708
+ "line": "tests/models/deepseek_v2/test_modeling_deepseek_v2.py::DeepseekV2IntegrationTest::test_deepseek_v2_lite",
709
+ "trace": "(line 516) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 400.00 MiB. GPU 0 has a total capacity of 22.30 GiB of which 138.69 MiB is free. Process 17429 has 22.16 GiB memory in use. Of the allocated memory 20.56 GiB is allocated by PyTorch, and 1.05 GiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)"
710
+ },
711
+ {
712
+ "line": "tests/models/deepseek_v2/test_modeling_deepseek_v2.py::DeepseekV2IntegrationTest::test_logits_eager",
713
+ "trace": "(line 516) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 24.00 MiB. GPU 0 has a total capacity of 22.30 GiB of which 10.69 MiB is free. Process 17429 has 22.29 GiB memory in use. Of the allocated memory 21.34 GiB is allocated by PyTorch, and 395.40 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)"
714
+ }
715
+ ],
716
+ "multi": [
717
+ {
718
+ "line": "tests/models/deepseek_v2/test_modeling_deepseek_v2.py::DeepseekV2ModelTest::test_cpu_offload",
719
+ "trace": "(line 2495) AssertionError: Tensor-likes are not close!"
720
+ },
721
+ {
722
+ "line": "tests/models/deepseek_v2/test_modeling_deepseek_v2.py::DeepseekV2ModelTest::test_disk_offload_bin",
723
+ "trace": "(line 2413) AssertionError: Tensor-likes are not close!"
724
+ },
725
+ {
726
+ "line": "tests/models/deepseek_v2/test_modeling_deepseek_v2.py::DeepseekV2ModelTest::test_disk_offload_safetensors",
727
+ "trace": "(line 2453) AssertionError: Tensor-likes are not close!"
728
+ },
729
+ {
730
+ "line": "tests/models/deepseek_v2/test_modeling_deepseek_v2.py::DeepseekV2IntegrationTest::test_batch_fa2",
731
+ "trace": "(line 243) RuntimeError: We encountered some issues during automatic conversion of the weights. For details look at the `CONVERSION` entries of the above report!"
732
+ },
733
+ {
734
+ "line": "tests/models/deepseek_v2/test_modeling_deepseek_v2.py::DeepseekV2IntegrationTest::test_deepseek_v2_lite",
735
+ "trace": "(line 516) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 400.00 MiB. GPU 0 has a total capacity of 22.30 GiB of which 100.69 MiB is free. Process 22619 has 22.20 GiB memory in use. Of the allocated memory 20.22 GiB is allocated by PyTorch, and 1.35 GiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)"
736
+ },
737
+ {
738
+ "line": "tests/models/deepseek_v2/test_modeling_deepseek_v2.py::DeepseekV2IntegrationTest::test_logits_eager",
739
+ "trace": "(line 243) RuntimeError: We encountered some issues during automatic conversion of the weights. For details look at the `CONVERSION` entries of the above report!"
740
+ }
741
+ ]
742
+ },
743
+ "job_link": {
744
+ "single": "https://github.com/huggingface/transformers/actions/runs/20918795329/job/60099266467",
745
+ "multi": "https://github.com/huggingface/transformers/actions/runs/20918795329/job/60099266446"
746
+ },
747
+ "captured_info": {
748
+ "single": "https://github.com/huggingface/transformers/actions/runs/20918795329/job/60099266467#step:16:1",
749
+ "multi": "https://github.com/huggingface/transformers/actions/runs/20918795329/job/60099266446#step:16:1"
750
+ }
751
+ },
752
+ "models_deepseek_v3": {
753
+ "failed": {
754
+ "PyTorch": {
755
+ "unclassified": 0,
756
+ "single": 2,
757
+ "multi": 2
758
+ },
759
+ "Tokenizers": {
760
+ "unclassified": 0,
761
+ "single": 0,
762
+ "multi": 0
763
+ },
764
+ "Pipelines": {
765
+ "unclassified": 0,
766
+ "single": 0,
767
+ "multi": 0
768
+ },
769
+ "Trainer": {
770
+ "unclassified": 0,
771
+ "single": 0,
772
+ "multi": 0
773
+ },
774
+ "ONNX": {
775
+ "unclassified": 0,
776
+ "single": 0,
777
+ "multi": 0
778
+ },
779
+ "Auto": {
780
+ "unclassified": 0,
781
+ "single": 0,
782
+ "multi": 0
783
+ },
784
+ "Quantization": {
785
+ "unclassified": 0,
786
+ "single": 0,
787
+ "multi": 0
788
+ },
789
+ "Unclassified": {
790
+ "unclassified": 0,
791
+ "single": 0,
792
+ "multi": 0
793
+ }
794
+ },
795
+ "errors": 0,
796
+ "success": 215,
797
+ "skipped": 207,
798
+ "time_spent": [
799
+ 126.93,
800
+ 122.9
801
+ ],
802
+ "error": false,
803
+ "failures": {
804
+ "single": [
805
+ {
806
+ "line": "tests/models/deepseek_v3/test_modeling_deepseek_v3.py::DeepseekV3ModelTest::test_eager_matches_sdpa_generate",
807
+ "trace": "(line 337) AssertionError:"
808
+ },
809
+ {
810
+ "line": "tests/models/deepseek_v3/test_modeling_deepseek_v3.py::DeepseekV3IntegrationTest::test_compile_static_cache",
811
+ "trace": "(line 423) AssertionError: Lists differ: ['Simply put, the theory of relativity states tha[370 chars]rew'] != ['Simplyput,thetheoryofrelativitystatesthatugarĠp[432 chars]est']"
812
+ }
813
+ ],
814
+ "multi": [
815
+ {
816
+ "line": "tests/models/deepseek_v3/test_modeling_deepseek_v3.py::DeepseekV3ModelTest::test_eager_matches_sdpa_generate",
817
+ "trace": "(line 337) AssertionError:"
818
+ },
819
+ {
820
+ "line": "tests/models/deepseek_v3/test_modeling_deepseek_v3.py::DeepseekV3IntegrationTest::test_compile_static_cache",
821
+ "trace": "(line 423) AssertionError: Lists differ: ['Simply put, the theory of relativity states tha[370 chars]rew'] != ['Simplyput,thetheoryofrelativitystatesthatugarĠp[432 chars]est']"
822
+ }
823
+ ]
824
+ },
825
+ "job_link": {
826
+ "single": "https://github.com/huggingface/transformers/actions/runs/20918795329/job/60099266372",
827
+ "multi": "https://github.com/huggingface/transformers/actions/runs/20918795329/job/60099266405"
828
+ },
829
+ "captured_info": {
830
+ "single": "https://github.com/huggingface/transformers/actions/runs/20918795329/job/60099266372#step:16:1",
831
+ "multi": "https://github.com/huggingface/transformers/actions/runs/20918795329/job/60099266405#step:16:1"
832
+ }
833
+ },
834
+ "models_diffllama": {
835
+ "failed": {
836
+ "PyTorch": {
837
+ "unclassified": 0,
838
+ "single": 1,
839
+ "multi": 1
840
+ },
841
+ "Tokenizers": {
842
+ "unclassified": 0,
843
+ "single": 0,
844
+ "multi": 0
845
+ },
846
+ "Pipelines": {
847
+ "unclassified": 0,
848
+ "single": 0,
849
+ "multi": 0
850
+ },
851
+ "Trainer": {
852
+ "unclassified": 0,
853
+ "single": 0,
854
+ "multi": 0
855
+ },
856
+ "ONNX": {
857
+ "unclassified": 0,
858
+ "single": 0,
859
+ "multi": 0
860
+ },
861
+ "Auto": {
862
+ "unclassified": 0,
863
+ "single": 0,
864
+ "multi": 0
865
+ },
866
+ "Quantization": {
867
+ "unclassified": 0,
868
+ "single": 0,
869
+ "multi": 0
870
+ },
871
+ "Unclassified": {
872
+ "unclassified": 0,
873
+ "single": 0,
874
+ "multi": 0
875
+ }
876
+ },
877
+ "errors": 0,
878
+ "success": 241,
879
+ "skipped": 203,
880
+ "time_spent": [
881
+ 149.46,
882
+ 149.96
883
+ ],
884
+ "error": false,
885
+ "failures": {
886
+ "single": [
887
+ {
888
+ "line": "tests/models/diffllama/test_modeling_diffllama.py::DiffLlamaIntegrationTest::test_compile_static_cache",
889
+ "trace": "(line 494) AssertionError: Lists differ: ['Sim[41 chars]that 1) the speed of light is constant in all [301 chars]y p'] != ['Sim[41 chars]that 2.5 5 5 5 5 5 5 5 5 5 5 5 5 5 5 5 5 5 5 '[133 chars]a a']"
890
+ }
891
+ ],
892
+ "multi": [
893
+ {
894
+ "line": "tests/models/diffllama/test_modeling_diffllama.py::DiffLlamaIntegrationTest::test_compile_static_cache",
895
+ "trace": "(line 494) AssertionError: Lists differ: ['Sim[41 chars]that 1) the speed of light is constant in all [301 chars]y p'] != ['Sim[41 chars]that 2.5 5 5 5 5 5 5 5 5 5 5 5 5 5 5 5 5 5 5 '[133 chars]a a']"
896
+ }
897
+ ]
898
+ },
899
+ "job_link": {
900
+ "single": "https://github.com/huggingface/transformers/actions/runs/20918795329/job/60099266369",
901
+ "multi": "https://github.com/huggingface/transformers/actions/runs/20918795329/job/60099266410"
902
+ },
903
+ "captured_info": {
904
+ "single": "https://github.com/huggingface/transformers/actions/runs/20918795329/job/60099266369#step:16:1",
905
+ "multi": "https://github.com/huggingface/transformers/actions/runs/20918795329/job/60099266410#step:16:1"
906
+ }
907
+ },
908
+ "models_doge": {
909
+ "failed": {
910
+ "PyTorch": {
911
+ "unclassified": 0,
912
+ "single": 0,
913
+ "multi": 0
914
+ },
915
+ "Tokenizers": {
916
+ "unclassified": 0,
917
+ "single": 0,
918
+ "multi": 0
919
+ },
920
+ "Pipelines": {
921
+ "unclassified": 0,
922
+ "single": 0,
923
+ "multi": 0
924
+ },
925
+ "Trainer": {
926
+ "unclassified": 0,
927
+ "single": 0,
928
+ "multi": 0
929
+ },
930
+ "ONNX": {
931
+ "unclassified": 0,
932
+ "single": 0,
933
+ "multi": 0
934
+ },
935
+ "Auto": {
936
+ "unclassified": 0,
937
+ "single": 0,
938
+ "multi": 0
939
+ },
940
+ "Quantization": {
941
+ "unclassified": 0,
942
+ "single": 0,
943
+ "multi": 0
944
+ },
945
+ "Unclassified": {
946
+ "unclassified": 0,
947
+ "single": 0,
948
+ "multi": 0
949
+ }
950
+ },
951
+ "errors": 0,
952
+ "success": 167,
953
+ "skipped": 265,
954
+ "time_spent": [
955
+ 42.44,
956
+ 43.91
957
+ ],
958
+ "error": false,
959
+ "failures": {},
960
+ "job_link": {
961
+ "single": "https://github.com/huggingface/transformers/actions/runs/20918795329/job/60099266387",
962
+ "multi": "https://github.com/huggingface/transformers/actions/runs/20918795329/job/60099266422"
963
+ },
964
+ "captured_info": {}
965
+ },
966
+ "models_dots1": {
967
+ "failed": {
968
+ "PyTorch": {
969
+ "unclassified": 0,
970
+ "single": 0,
971
+ "multi": 0
972
+ },
973
+ "Tokenizers": {
974
+ "unclassified": 0,
975
+ "single": 0,
976
+ "multi": 0
977
+ },
978
+ "Pipelines": {
979
+ "unclassified": 0,
980
+ "single": 0,
981
+ "multi": 0
982
+ },
983
+ "Trainer": {
984
+ "unclassified": 0,
985
+ "single": 0,
986
+ "multi": 0
987
+ },
988
+ "ONNX": {
989
+ "unclassified": 0,
990
+ "single": 0,
991
+ "multi": 0
992
+ },
993
+ "Auto": {
994
+ "unclassified": 0,
995
+ "single": 0,
996
+ "multi": 0
997
+ },
998
+ "Quantization": {
999
+ "unclassified": 0,
1000
+ "single": 0,
1001
+ "multi": 0
1002
+ },
1003
+ "Unclassified": {
1004
+ "unclassified": 0,
1005
+ "single": 0,
1006
+ "multi": 0
1007
+ }
1008
+ },
1009
+ "errors": 0,
1010
+ "success": 239,
1011
+ "skipped": 209,
1012
+ "time_spent": [
1013
+ 156.7,
1014
+ 124.96
1015
+ ],
1016
+ "error": false,
1017
+ "failures": {},
1018
+ "job_link": {
1019
+ "single": "https://github.com/huggingface/transformers/actions/runs/20918795329/job/60099266392",
1020
+ "multi": "https://github.com/huggingface/transformers/actions/runs/20918795329/job/60099266427"
1021
+ },
1022
+ "captured_info": {
1023
+ "single": "https://github.com/huggingface/transformers/actions/runs/20918795329/job/60099266392#step:16:1",
1024
+ "multi": "https://github.com/huggingface/transformers/actions/runs/20918795329/job/60099266427#step:16:1"
1025
+ }
1026
+ },
1027
+ "models_emu3": {
1028
+ "failed": {
1029
+ "PyTorch": {
1030
+ "unclassified": 0,
1031
+ "single": 4,
1032
+ "multi": 4
1033
+ },
1034
+ "Tokenizers": {
1035
+ "unclassified": 0,
1036
+ "single": 0,
1037
+ "multi": 0
1038
+ },
1039
+ "Pipelines": {
1040
+ "unclassified": 0,
1041
+ "single": 0,
1042
+ "multi": 0
1043
+ },
1044
+ "Trainer": {
1045
+ "unclassified": 0,
1046
+ "single": 0,
1047
+ "multi": 0
1048
+ },
1049
+ "ONNX": {
1050
+ "unclassified": 0,
1051
+ "single": 0,
1052
+ "multi": 0
1053
+ },
1054
+ "Auto": {
1055
+ "unclassified": 0,
1056
+ "single": 0,
1057
+ "multi": 0
1058
+ },
1059
+ "Quantization": {
1060
+ "unclassified": 0,
1061
+ "single": 0,
1062
+ "multi": 0
1063
+ },
1064
+ "Unclassified": {
1065
+ "unclassified": 0,
1066
+ "single": 0,
1067
+ "multi": 0
1068
+ }
1069
+ },
1070
+ "errors": 0,
1071
+ "success": 484,
1072
+ "skipped": 468,
1073
+ "time_spent": [
1074
+ 339.35,
1075
+ 335.49
1076
+ ],
1077
+ "error": false,
1078
+ "failures": {
1079
+ "multi": [
1080
+ {
1081
+ "line": "tests/models/emu3/test_modeling_emu3.py::Emu3IntegrationTest::test_model_generate_images",
1082
+ "trace": "(line 516) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 224.00 MiB. GPU 0 has a total capacity of 22.30 GiB of which 174.69 MiB is free. Process 196200 has 22.12 GiB memory in use. Of the allocated memory 21.42 GiB is allocated by PyTorch, and 212.87 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)"
1083
+ },
1084
+ {
1085
+ "line": "tests/models/emu3/test_modeling_emu3.py::Emu3IntegrationTest::test_model_generation",
1086
+ "trace": "(line 516) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 224.00 MiB. GPU 0 has a total capacity of 22.30 GiB of which 188.69 MiB is free. Process 196200 has 22.11 GiB memory in use. Of the allocated memory 21.55 GiB is allocated by PyTorch, and 70.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)"
1087
+ },
1088
+ {
1089
+ "line": "tests/models/emu3/test_modeling_emu3.py::Emu3IntegrationTest::test_model_generation_batched",
1090
+ "trace": "(line 516) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 2.82 GiB. GPU 0 has a total capacity of 22.30 GiB of which 1.75 GiB is free. Process 196200 has 20.54 GiB memory in use. Of the allocated memory 19.98 GiB is allocated by PyTorch, and 66.64 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)"
1091
+ },
1092
+ {
1093
+ "line": "tests/models/emu3/test_modeling_emu3.py::Emu3IntegrationTest::test_model_generation_multi_image",
1094
+ "trace": "(line 516) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 224.00 MiB. GPU 0 has a total capacity of 22.30 GiB of which 22.69 MiB is free. Process 196200 has 22.27 GiB memory in use. Of the allocated memory 21.75 GiB is allocated by PyTorch, and 31.62 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)"
1095
+ }
1096
+ ],
1097
+ "single": [
1098
+ {
1099
+ "line": "tests/models/emu3/test_modeling_emu3.py::Emu3IntegrationTest::test_model_generate_images",
1100
+ "trace": "(line 516) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 64.00 MiB. GPU 0 has a total capacity of 22.30 GiB of which 54.69 MiB is free. Process 176354 has 22.24 GiB memory in use. Of the allocated memory 21.83 GiB is allocated by PyTorch, and 28.37 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)"
1101
+ },
1102
+ {
1103
+ "line": "tests/models/emu3/test_modeling_emu3.py::Emu3IntegrationTest::test_model_generation",
1104
+ "trace": "(line 516) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 224.00 MiB. GPU 0 has a total capacity of 22.30 GiB of which 136.69 MiB is free. Process 176354 has 22.16 GiB memory in use. Of the allocated memory 21.73 GiB is allocated by PyTorch, and 54.80 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)"
1105
+ },
1106
+ {
1107
+ "line": "tests/models/emu3/test_modeling_emu3.py::Emu3IntegrationTest::test_model_generation_batched",
1108
+ "trace": "(line 516) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 2.82 GiB. GPU 0 has a total capacity of 22.30 GiB of which 2.40 GiB is free. Process 176354 has 19.89 GiB memory in use. Of the allocated memory 19.30 GiB is allocated by PyTorch, and 212.20 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)"
1109
+ },
1110
+ {
1111
+ "line": "tests/models/emu3/test_modeling_emu3.py::Emu3IntegrationTest::test_model_generation_multi_image",
1112
+ "trace": "(line 516) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 224.00 MiB. GPU 0 has a total capacity of 22.30 GiB of which 30.69 MiB is free. Process 176354 has 22.27 GiB memory in use. Of the allocated memory 21.83 GiB is allocated by PyTorch, and 53.63 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)"
1113
+ }
1114
+ ]
1115
+ },
1116
+ "job_link": {
1117
+ "multi": "https://github.com/huggingface/transformers/actions/runs/20918795329/job/60099266384",
1118
+ "single": "https://github.com/huggingface/transformers/actions/runs/20918795329/job/60099266400"
1119
+ },
1120
+ "captured_info": {}
1121
+ },
1122
+ "models_ernie4_5": {
1123
+ "failed": {
1124
+ "PyTorch": {
1125
+ "unclassified": 0,
1126
+ "single": 0,
1127
+ "multi": 0
1128
+ },
1129
+ "Tokenizers": {
1130
+ "unclassified": 0,
1131
+ "single": 0,
1132
+ "multi": 0
1133
+ },
1134
+ "Pipelines": {
1135
+ "unclassified": 0,
1136
+ "single": 0,
1137
+ "multi": 0
1138
+ },
1139
+ "Trainer": {
1140
+ "unclassified": 0,
1141
+ "single": 0,
1142
+ "multi": 0
1143
+ },
1144
+ "ONNX": {
1145
+ "unclassified": 0,
1146
+ "single": 0,
1147
+ "multi": 0
1148
+ },
1149
+ "Auto": {
1150
+ "unclassified": 0,
1151
+ "single": 0,
1152
+ "multi": 0
1153
+ },
1154
+ "Quantization": {
1155
+ "unclassified": 0,
1156
+ "single": 0,
1157
+ "multi": 0
1158
+ },
1159
+ "Unclassified": {
1160
+ "unclassified": 0,
1161
+ "single": 0,
1162
+ "multi": 0
1163
+ }
1164
+ },
1165
+ "errors": 0,
1166
+ "success": 247,
1167
+ "skipped": 201,
1168
+ "time_spent": [
1169
+ 88.29,
1170
+ 90.23
1171
+ ],
1172
+ "error": false,
1173
+ "failures": {},
1174
+ "job_link": {
1175
+ "single": "https://github.com/huggingface/transformers/actions/runs/20918795329/job/60099266425",
1176
+ "multi": "https://github.com/huggingface/transformers/actions/runs/20918795329/job/60099266379"
1177
+ },
1178
+ "captured_info": {
1179
+ "single": "https://github.com/huggingface/transformers/actions/runs/20918795329/job/60099266425#step:16:1",
1180
+ "multi": "https://github.com/huggingface/transformers/actions/runs/20918795329/job/60099266379#step:16:1"
1181
+ }
1182
+ }
1183
+ }