hf-transformers-bot commited on
Commit
3f1b4c3
·
verified ·
1 Parent(s): 4446f35

Upload 2025-11-03/runs/5702-19037454680/ci_results_run_models_gpu/model_results.json with huggingface_hub

Browse files
2025-11-03/runs/5702-19037454680/ci_results_run_models_gpu/model_results.json ADDED
@@ -0,0 +1,522 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "models_csm": {
3
+ "failed": {
4
+ "PyTorch": {
5
+ "unclassified": 0,
6
+ "single": 0,
7
+ "multi": 2
8
+ },
9
+ "Tokenizers": {
10
+ "unclassified": 0,
11
+ "single": 0,
12
+ "multi": 0
13
+ },
14
+ "Pipelines": {
15
+ "unclassified": 0,
16
+ "single": 0,
17
+ "multi": 0
18
+ },
19
+ "Trainer": {
20
+ "unclassified": 0,
21
+ "single": 0,
22
+ "multi": 0
23
+ },
24
+ "ONNX": {
25
+ "unclassified": 0,
26
+ "single": 0,
27
+ "multi": 0
28
+ },
29
+ "Auto": {
30
+ "unclassified": 0,
31
+ "single": 0,
32
+ "multi": 0
33
+ },
34
+ "Quantization": {
35
+ "unclassified": 0,
36
+ "single": 0,
37
+ "multi": 0
38
+ },
39
+ "Unclassified": {
40
+ "unclassified": 0,
41
+ "single": 0,
42
+ "multi": 0
43
+ }
44
+ },
45
+ "errors": 0,
46
+ "success": 238,
47
+ "skipped": 158,
48
+ "time_spent": [
49
+ 137.19,
50
+ 138.46
51
+ ],
52
+ "error": false,
53
+ "failures": {
54
+ "multi": [
55
+ {
56
+ "line": "tests/models/csm/test_modeling_csm.py::CsmForConditionalGenerationTest::test_generate_with_static_cache",
57
+ "trace": "(line 4971) IndexError: tuple index out of range"
58
+ },
59
+ {
60
+ "line": "tests/models/csm/test_modeling_csm.py::CsmForConditionalGenerationTest::test_multi_gpu_data_parallel_forward",
61
+ "trace": "(line 66) RuntimeError: NCCL Error 2: unhandled system error (run with NCCL_DEBUG=INFO for details)"
62
+ }
63
+ ]
64
+ },
65
+ "job_link": {
66
+ "multi": "https://github.com/huggingface/transformers/actions/runs/19037454680/job/54365964359",
67
+ "single": "https://github.com/huggingface/transformers/actions/runs/19037454680/job/54365964255"
68
+ },
69
+ "captured_info": {}
70
+ },
71
+ "models_gemma3n": {
72
+ "failed": {
73
+ "PyTorch": {
74
+ "unclassified": 0,
75
+ "single": 4,
76
+ "multi": 6
77
+ },
78
+ "Tokenizers": {
79
+ "unclassified": 0,
80
+ "single": 0,
81
+ "multi": 0
82
+ },
83
+ "Pipelines": {
84
+ "unclassified": 0,
85
+ "single": 0,
86
+ "multi": 0
87
+ },
88
+ "Trainer": {
89
+ "unclassified": 0,
90
+ "single": 0,
91
+ "multi": 0
92
+ },
93
+ "ONNX": {
94
+ "unclassified": 0,
95
+ "single": 0,
96
+ "multi": 0
97
+ },
98
+ "Auto": {
99
+ "unclassified": 0,
100
+ "single": 0,
101
+ "multi": 0
102
+ },
103
+ "Quantization": {
104
+ "unclassified": 0,
105
+ "single": 0,
106
+ "multi": 0
107
+ },
108
+ "Unclassified": {
109
+ "unclassified": 0,
110
+ "single": 0,
111
+ "multi": 0
112
+ }
113
+ },
114
+ "errors": 0,
115
+ "success": 317,
116
+ "skipped": 683,
117
+ "time_spent": [
118
+ 361.5,
119
+ 369.68
120
+ ],
121
+ "error": false,
122
+ "failures": {
123
+ "single": [
124
+ {
125
+ "line": "tests/models/gemma3n/test_modeling_gemma3n.py::Gemma3nTextModelTest::test_flash_attn_2_equivalence",
126
+ "trace": "(line 581) AssertionError: Tensor-likes are not close!"
127
+ },
128
+ {
129
+ "line": "tests/models/gemma3n/test_modeling_gemma3n.py::Gemma3nTextModelTest::test_flash_attn_2_inference_equivalence",
130
+ "trace": "(line 2740) AssertionError: Tensor-likes are not close!"
131
+ },
132
+ {
133
+ "line": "tests/models/gemma3n/test_modeling_gemma3n.py::Gemma3nTextModelTest::test_flash_attn_2_inference_equivalence_right_padding",
134
+ "trace": "(line 2742) AssertionError: Tensor-likes are not close!"
135
+ },
136
+ {
137
+ "line": "tests/models/gemma3n/test_modeling_gemma3n.py::Gemma3nIntegrationTest::test_generation_beyond_sliding_window_with_generation_config",
138
+ "trace": "(line 644) AssertionError: Lists differ: [\" and I think it's a nice place to visit.[99 chars]re'\"] != [' and I am glad to be here. This is a nic[102 chars]are']"
139
+ }
140
+ ],
141
+ "multi": [
142
+ {
143
+ "line": "tests/models/gemma3n/test_modeling_gemma3n.py::Gemma3nTextModelTest::test_flash_attention_2_padding_matches_padding_free_with_position_ids_and_fa_kwargs",
144
+ "trace": "(line 2059) AssertionError: Tensor-likes are not close!"
145
+ },
146
+ {
147
+ "line": "tests/models/gemma3n/test_modeling_gemma3n.py::Gemma3nTextModelTest::test_flash_attn_2_equivalence",
148
+ "trace": "(line 581) AssertionError: Tensor-likes are not close!"
149
+ },
150
+ {
151
+ "line": "tests/models/gemma3n/test_modeling_gemma3n.py::Gemma3nTextModelTest::test_flash_attn_2_inference_equivalence",
152
+ "trace": "(line 2738) AssertionError: Tensor-likes are not close!"
153
+ },
154
+ {
155
+ "line": "tests/models/gemma3n/test_modeling_gemma3n.py::Gemma3nTextModelTest::test_flash_attn_2_inference_equivalence_right_padding",
156
+ "trace": "(line 2738) AssertionError: Tensor-likes are not close!"
157
+ },
158
+ {
159
+ "line": "tests/models/gemma3n/test_modeling_gemma3n.py::Gemma3nTextModelTest::test_multi_gpu_data_parallel_forward",
160
+ "trace": "(line 66) RuntimeError: NCCL Error 2: unhandled system error (run with NCCL_DEBUG=INFO for details)"
161
+ },
162
+ {
163
+ "line": "tests/models/gemma3n/test_modeling_gemma3n.py::Gemma3nIntegrationTest::test_generation_beyond_sliding_window_with_generation_config",
164
+ "trace": "(line 644) AssertionError: Lists differ: [\" and I think it's a nice place to visit.[99 chars]re'\"] != [' and I am glad to be here. This is a nic[102 chars]are']"
165
+ }
166
+ ]
167
+ },
168
+ "job_link": {
169
+ "single": "https://github.com/huggingface/transformers/actions/runs/19037454680/job/54365964298",
170
+ "multi": "https://github.com/huggingface/transformers/actions/runs/19037454680/job/54365964366"
171
+ },
172
+ "captured_info": {
173
+ "single": "https://github.com/huggingface/transformers/actions/runs/19037454680/job/54365964298#step:16:1",
174
+ "multi": "https://github.com/huggingface/transformers/actions/runs/19037454680/job/54365964366#step:16:1"
175
+ }
176
+ },
177
+ "models_llama": {
178
+ "failed": {
179
+ "PyTorch": {
180
+ "unclassified": 0,
181
+ "single": 0,
182
+ "multi": 1
183
+ },
184
+ "Tokenizers": {
185
+ "unclassified": 0,
186
+ "single": 0,
187
+ "multi": 0
188
+ },
189
+ "Pipelines": {
190
+ "unclassified": 0,
191
+ "single": 0,
192
+ "multi": 0
193
+ },
194
+ "Trainer": {
195
+ "unclassified": 0,
196
+ "single": 0,
197
+ "multi": 0
198
+ },
199
+ "ONNX": {
200
+ "unclassified": 0,
201
+ "single": 0,
202
+ "multi": 0
203
+ },
204
+ "Auto": {
205
+ "unclassified": 0,
206
+ "single": 0,
207
+ "multi": 0
208
+ },
209
+ "Quantization": {
210
+ "unclassified": 0,
211
+ "single": 0,
212
+ "multi": 0
213
+ },
214
+ "Unclassified": {
215
+ "unclassified": 0,
216
+ "single": 0,
217
+ "multi": 0
218
+ }
219
+ },
220
+ "errors": 0,
221
+ "success": 492,
222
+ "skipped": 199,
223
+ "time_spent": [
224
+ 261.92,
225
+ 258.84
226
+ ],
227
+ "error": false,
228
+ "failures": {
229
+ "multi": [
230
+ {
231
+ "line": "tests/models/llama/test_modeling_llama.py::LlamaModelTest::test_multi_gpu_data_parallel_forward",
232
+ "trace": "(line 66) RuntimeError: NCCL Error 2: unhandled system error (run with NCCL_DEBUG=INFO for details)"
233
+ }
234
+ ]
235
+ },
236
+ "job_link": {
237
+ "single": "https://github.com/huggingface/transformers/actions/runs/19037454680/job/54365964382",
238
+ "multi": "https://github.com/huggingface/transformers/actions/runs/19037454680/job/54365964362"
239
+ },
240
+ "captured_info": {}
241
+ },
242
+ "models_modernbert": {
243
+ "failed": {
244
+ "PyTorch": {
245
+ "unclassified": 0,
246
+ "single": 0,
247
+ "multi": 1
248
+ },
249
+ "Tokenizers": {
250
+ "unclassified": 0,
251
+ "single": 0,
252
+ "multi": 0
253
+ },
254
+ "Pipelines": {
255
+ "unclassified": 0,
256
+ "single": 0,
257
+ "multi": 0
258
+ },
259
+ "Trainer": {
260
+ "unclassified": 0,
261
+ "single": 0,
262
+ "multi": 0
263
+ },
264
+ "ONNX": {
265
+ "unclassified": 0,
266
+ "single": 0,
267
+ "multi": 0
268
+ },
269
+ "Auto": {
270
+ "unclassified": 0,
271
+ "single": 0,
272
+ "multi": 0
273
+ },
274
+ "Quantization": {
275
+ "unclassified": 0,
276
+ "single": 0,
277
+ "multi": 0
278
+ },
279
+ "Unclassified": {
280
+ "unclassified": 0,
281
+ "single": 0,
282
+ "multi": 0
283
+ }
284
+ },
285
+ "errors": 0,
286
+ "success": 179,
287
+ "skipped": 170,
288
+ "time_spent": [
289
+ 124.34,
290
+ 122.35
291
+ ],
292
+ "error": false,
293
+ "failures": {
294
+ "multi": [
295
+ {
296
+ "line": "tests/models/modernbert/test_modeling_modernbert.py::ModernBertModelTest::test_multi_gpu_data_parallel_forward",
297
+ "trace": "(line 66) RuntimeError: NCCL Error 2: unhandled system error (run with NCCL_DEBUG=INFO for details)"
298
+ }
299
+ ]
300
+ },
301
+ "job_link": {
302
+ "multi": "https://github.com/huggingface/transformers/actions/runs/19037454680/job/54365964412",
303
+ "single": "https://github.com/huggingface/transformers/actions/runs/19037454680/job/54365964301"
304
+ },
305
+ "captured_info": {}
306
+ },
307
+ "models_qwen2": {
308
+ "failed": {
309
+ "PyTorch": {
310
+ "unclassified": 0,
311
+ "single": 0,
312
+ "multi": 1
313
+ },
314
+ "Tokenizers": {
315
+ "unclassified": 0,
316
+ "single": 0,
317
+ "multi": 0
318
+ },
319
+ "Pipelines": {
320
+ "unclassified": 0,
321
+ "single": 0,
322
+ "multi": 0
323
+ },
324
+ "Trainer": {
325
+ "unclassified": 0,
326
+ "single": 0,
327
+ "multi": 0
328
+ },
329
+ "ONNX": {
330
+ "unclassified": 0,
331
+ "single": 0,
332
+ "multi": 0
333
+ },
334
+ "Auto": {
335
+ "unclassified": 0,
336
+ "single": 0,
337
+ "multi": 0
338
+ },
339
+ "Quantization": {
340
+ "unclassified": 0,
341
+ "single": 0,
342
+ "multi": 0
343
+ },
344
+ "Unclassified": {
345
+ "unclassified": 0,
346
+ "single": 0,
347
+ "multi": 0
348
+ }
349
+ },
350
+ "errors": 0,
351
+ "success": 468,
352
+ "skipped": 187,
353
+ "time_spent": [
354
+ 182.84,
355
+ 180.7
356
+ ],
357
+ "error": false,
358
+ "failures": {
359
+ "multi": [
360
+ {
361
+ "line": "tests/models/qwen2/test_modeling_qwen2.py::Qwen2ModelTest::test_multi_gpu_data_parallel_forward",
362
+ "trace": "(line 66) RuntimeError: NCCL Error 2: unhandled system error (run with NCCL_DEBUG=INFO for details)"
363
+ }
364
+ ]
365
+ },
366
+ "job_link": {
367
+ "single": "https://github.com/huggingface/transformers/actions/runs/19037454680/job/54365964419",
368
+ "multi": "https://github.com/huggingface/transformers/actions/runs/19037454680/job/54365964379"
369
+ },
370
+ "captured_info": {
371
+ "single": "https://github.com/huggingface/transformers/actions/runs/19037454680/job/54365964419#step:16:1",
372
+ "multi": "https://github.com/huggingface/transformers/actions/runs/19037454680/job/54365964379#step:16:1"
373
+ }
374
+ },
375
+ "models_qwen2_5_omni": {
376
+ "failed": {
377
+ "PyTorch": {
378
+ "unclassified": 0,
379
+ "single": 0,
380
+ "multi": 1
381
+ },
382
+ "Tokenizers": {
383
+ "unclassified": 0,
384
+ "single": 0,
385
+ "multi": 0
386
+ },
387
+ "Pipelines": {
388
+ "unclassified": 0,
389
+ "single": 0,
390
+ "multi": 0
391
+ },
392
+ "Trainer": {
393
+ "unclassified": 0,
394
+ "single": 0,
395
+ "multi": 0
396
+ },
397
+ "ONNX": {
398
+ "unclassified": 0,
399
+ "single": 0,
400
+ "multi": 0
401
+ },
402
+ "Auto": {
403
+ "unclassified": 0,
404
+ "single": 0,
405
+ "multi": 0
406
+ },
407
+ "Quantization": {
408
+ "unclassified": 0,
409
+ "single": 0,
410
+ "multi": 0
411
+ },
412
+ "Unclassified": {
413
+ "unclassified": 0,
414
+ "single": 0,
415
+ "multi": 0
416
+ }
417
+ },
418
+ "errors": 0,
419
+ "success": 312,
420
+ "skipped": 91,
421
+ "time_spent": [
422
+ 166.85,
423
+ 175.08
424
+ ],
425
+ "error": false,
426
+ "failures": {
427
+ "multi": [
428
+ {
429
+ "line": "tests/models/qwen2_5_omni/test_modeling_qwen2_5_omni.py::Qwen2_5OmniThinkerForConditionalGenerationModelTest::test_multi_gpu_data_parallel_forward",
430
+ "trace": "(line 66) RuntimeError: NCCL Error 2: unhandled system error (run with NCCL_DEBUG=INFO for details)"
431
+ }
432
+ ]
433
+ },
434
+ "job_link": {
435
+ "multi": "https://github.com/huggingface/transformers/actions/runs/19037454680/job/54365964450",
436
+ "single": "https://github.com/huggingface/transformers/actions/runs/19037454680/job/54365964259"
437
+ },
438
+ "captured_info": {}
439
+ },
440
+ "models_qwen2_5_vl": {
441
+ "failed": {
442
+ "PyTorch": {
443
+ "unclassified": 0,
444
+ "single": 2,
445
+ "multi": 2
446
+ },
447
+ "Tokenizers": {
448
+ "unclassified": 0,
449
+ "single": 0,
450
+ "multi": 0
451
+ },
452
+ "Pipelines": {
453
+ "unclassified": 0,
454
+ "single": 0,
455
+ "multi": 0
456
+ },
457
+ "Trainer": {
458
+ "unclassified": 0,
459
+ "single": 0,
460
+ "multi": 0
461
+ },
462
+ "ONNX": {
463
+ "unclassified": 0,
464
+ "single": 0,
465
+ "multi": 0
466
+ },
467
+ "Auto": {
468
+ "unclassified": 0,
469
+ "single": 0,
470
+ "multi": 0
471
+ },
472
+ "Quantization": {
473
+ "unclassified": 0,
474
+ "single": 0,
475
+ "multi": 0
476
+ },
477
+ "Unclassified": {
478
+ "unclassified": 0,
479
+ "single": 0,
480
+ "multi": 0
481
+ }
482
+ },
483
+ "errors": 0,
484
+ "success": 329,
485
+ "skipped": 87,
486
+ "time_spent": [
487
+ 204.87,
488
+ 207.82
489
+ ],
490
+ "error": false,
491
+ "failures": {
492
+ "multi": [
493
+ {
494
+ "line": "tests/models/qwen2_5_vl/test_modeling_qwen2_5_vl.py::Qwen2_5_VLIntegrationTest::test_small_model_integration_test_batch_different_resolutions",
495
+ "trace": "(line 625) AssertionError: 'syst[73 chars]ant\\n addCriterion\\nThe dog in the picture app[95 chars]h is' != 'syst[73 chars]ant\\nThe dog in the picture appears to be a La[94 chars]t in'"
496
+ },
497
+ {
498
+ "line": "tests/models/qwen2_5_vl/test_modeling_qwen2_5_vl.py::Qwen2_5_VLIntegrationTest::test_small_model_integration_test_batch_wo_image_flashatt2",
499
+ "trace": "(line 698) AssertionError: Lists differ: ['sys[216 chars]in', 'system\\nYou are a helpful assistant.\\nus[29 chars]aks'] != ['sys[216 chars]in', \"system\\nYou are a helpful assistant.\\nus[162 chars]ing\"]"
500
+ }
501
+ ],
502
+ "single": [
503
+ {
504
+ "line": "tests/models/qwen2_5_vl/test_modeling_qwen2_5_vl.py::Qwen2_5_VLIntegrationTest::test_small_model_integration_test_batch_different_resolutions",
505
+ "trace": "(line 625) AssertionError: 'syst[73 chars]ant\\n addCriterion\\nThe dog in the picture app[95 chars]h is' != 'syst[73 chars]ant\\nThe dog in the picture appears to be a La[94 chars]t in'"
506
+ },
507
+ {
508
+ "line": "tests/models/qwen2_5_vl/test_modeling_qwen2_5_vl.py::Qwen2_5_VLIntegrationTest::test_small_model_integration_test_batch_wo_image_flashatt2",
509
+ "trace": "(line 698) AssertionError: Lists differ: ['sys[216 chars]in', 'system\\nYou are a helpful assistant.\\nus[29 chars]aks'] != ['sys[216 chars]in', \"system\\nYou are a helpful assistant.\\nus[162 chars]ing\"]"
510
+ }
511
+ ]
512
+ },
513
+ "job_link": {
514
+ "multi": "https://github.com/huggingface/transformers/actions/runs/19037454680/job/54365964385",
515
+ "single": "https://github.com/huggingface/transformers/actions/runs/19037454680/job/54365964300"
516
+ },
517
+ "captured_info": {
518
+ "multi": "https://github.com/huggingface/transformers/actions/runs/19037454680/job/54365964385#step:16:1",
519
+ "single": "https://github.com/huggingface/transformers/actions/runs/19037454680/job/54365964300#step:16:1"
520
+ }
521
+ }
522
+ }