hf-transformers-bot commited on
Commit
296dfd2
·
verified ·
1 Parent(s): c8d672c

Upload 2025-12-05/runs/24768-19962235018/ci_results_run_models_gpu/model_results.json with huggingface_hub

Browse files
2025-12-05/runs/24768-19962235018/ci_results_run_models_gpu/model_results.json ADDED
@@ -0,0 +1,678 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "models_deepseek_v3": {
3
+ "failed": {
4
+ "PyTorch": {
5
+ "unclassified": 0,
6
+ "single": 2,
7
+ "multi": 2
8
+ },
9
+ "Tokenizers": {
10
+ "unclassified": 0,
11
+ "single": 0,
12
+ "multi": 0
13
+ },
14
+ "Pipelines": {
15
+ "unclassified": 0,
16
+ "single": 0,
17
+ "multi": 0
18
+ },
19
+ "Trainer": {
20
+ "unclassified": 0,
21
+ "single": 0,
22
+ "multi": 0
23
+ },
24
+ "ONNX": {
25
+ "unclassified": 0,
26
+ "single": 0,
27
+ "multi": 0
28
+ },
29
+ "Auto": {
30
+ "unclassified": 0,
31
+ "single": 0,
32
+ "multi": 0
33
+ },
34
+ "Quantization": {
35
+ "unclassified": 0,
36
+ "single": 0,
37
+ "multi": 0
38
+ },
39
+ "Unclassified": {
40
+ "unclassified": 0,
41
+ "single": 0,
42
+ "multi": 0
43
+ }
44
+ },
45
+ "errors": 0,
46
+ "success": 215,
47
+ "skipped": 207,
48
+ "time_spent": [
49
+ 741.96,
50
+ 746.02
51
+ ],
52
+ "error": false,
53
+ "failures": {
54
+ "single": [
55
+ {
56
+ "line": "tests/models/deepseek_v3/test_modeling_deepseek_v3.py::DeepseekV3ModelTest::test_eager_matches_sdpa_generate",
57
+ "trace": "(line 240) RuntimeError: You set `ignore_mismatched_sizes` to `False`, thus raising an error. For details look at the above report!"
58
+ },
59
+ {
60
+ "line": "tests/models/deepseek_v3/test_modeling_deepseek_v3.py::DeepseekV3IntegrationTest::test_compile_static_cache",
61
+ "trace": "(line 240) RuntimeError: You set `ignore_mismatched_sizes` to `False`, thus raising an error. For details look at the above report!"
62
+ }
63
+ ],
64
+ "multi": [
65
+ {
66
+ "line": "tests/models/deepseek_v3/test_modeling_deepseek_v3.py::DeepseekV3ModelTest::test_eager_matches_sdpa_generate",
67
+ "trace": "(line 240) RuntimeError: You set `ignore_mismatched_sizes` to `False`, thus raising an error. For details look at the above report!"
68
+ },
69
+ {
70
+ "line": "tests/models/deepseek_v3/test_modeling_deepseek_v3.py::DeepseekV3IntegrationTest::test_compile_static_cache",
71
+ "trace": "(line 240) RuntimeError: You set `ignore_mismatched_sizes` to `False`, thus raising an error. For details look at the above report!"
72
+ }
73
+ ]
74
+ },
75
+ "job_link": {
76
+ "single": "https://github.com/huggingface/transformers/actions/runs/19962235018/job/57245316238",
77
+ "multi": "https://github.com/huggingface/transformers/actions/runs/19962235018/job/57245316188"
78
+ },
79
+ "captured_info": {}
80
+ },
81
+ "models_gemma2": {
82
+ "failed": {
83
+ "PyTorch": {
84
+ "unclassified": 0,
85
+ "single": 0,
86
+ "multi": 0
87
+ },
88
+ "Tokenizers": {
89
+ "unclassified": 0,
90
+ "single": 0,
91
+ "multi": 0
92
+ },
93
+ "Pipelines": {
94
+ "unclassified": 0,
95
+ "single": 0,
96
+ "multi": 0
97
+ },
98
+ "Trainer": {
99
+ "unclassified": 0,
100
+ "single": 0,
101
+ "multi": 0
102
+ },
103
+ "ONNX": {
104
+ "unclassified": 0,
105
+ "single": 0,
106
+ "multi": 0
107
+ },
108
+ "Auto": {
109
+ "unclassified": 0,
110
+ "single": 0,
111
+ "multi": 0
112
+ },
113
+ "Quantization": {
114
+ "unclassified": 0,
115
+ "single": 0,
116
+ "multi": 0
117
+ },
118
+ "Unclassified": {
119
+ "unclassified": 0,
120
+ "single": 0,
121
+ "multi": 0
122
+ }
123
+ },
124
+ "errors": 0,
125
+ "success": 270,
126
+ "skipped": 206,
127
+ "time_spent": [
128
+ 584.59,
129
+ 769.18
130
+ ],
131
+ "error": false,
132
+ "failures": {},
133
+ "job_link": {
134
+ "multi": "https://github.com/huggingface/transformers/actions/runs/19962235018/job/57245316168",
135
+ "single": "https://github.com/huggingface/transformers/actions/runs/19962235018/job/57245316214"
136
+ },
137
+ "captured_info": {}
138
+ },
139
+ "models_gemma3": {
140
+ "failed": {
141
+ "PyTorch": {
142
+ "unclassified": 0,
143
+ "single": 0,
144
+ "multi": 0
145
+ },
146
+ "Tokenizers": {
147
+ "unclassified": 0,
148
+ "single": 0,
149
+ "multi": 0
150
+ },
151
+ "Pipelines": {
152
+ "unclassified": 0,
153
+ "single": 0,
154
+ "multi": 0
155
+ },
156
+ "Trainer": {
157
+ "unclassified": 0,
158
+ "single": 0,
159
+ "multi": 0
160
+ },
161
+ "ONNX": {
162
+ "unclassified": 0,
163
+ "single": 0,
164
+ "multi": 0
165
+ },
166
+ "Auto": {
167
+ "unclassified": 0,
168
+ "single": 0,
169
+ "multi": 0
170
+ },
171
+ "Quantization": {
172
+ "unclassified": 0,
173
+ "single": 0,
174
+ "multi": 0
175
+ },
176
+ "Unclassified": {
177
+ "unclassified": 0,
178
+ "single": 0,
179
+ "multi": 0
180
+ }
181
+ },
182
+ "errors": 0,
183
+ "success": 572,
184
+ "skipped": 362,
185
+ "time_spent": [
186
+ 696.12,
187
+ 694.04
188
+ ],
189
+ "error": false,
190
+ "failures": {},
191
+ "job_link": {
192
+ "multi": "https://github.com/huggingface/transformers/actions/runs/19962235018/job/57245316172",
193
+ "single": "https://github.com/huggingface/transformers/actions/runs/19962235018/job/57245316242"
194
+ },
195
+ "captured_info": {
196
+ "multi": "https://github.com/huggingface/transformers/actions/runs/19962235018/job/57245316172#step:16:1",
197
+ "single": "https://github.com/huggingface/transformers/actions/runs/19962235018/job/57245316242#step:16:1"
198
+ }
199
+ },
200
+ "models_llama": {
201
+ "failed": {
202
+ "PyTorch": {
203
+ "unclassified": 0,
204
+ "single": 0,
205
+ "multi": 0
206
+ },
207
+ "Tokenizers": {
208
+ "unclassified": 0,
209
+ "single": 5,
210
+ "multi": 5
211
+ },
212
+ "Pipelines": {
213
+ "unclassified": 0,
214
+ "single": 0,
215
+ "multi": 0
216
+ },
217
+ "Trainer": {
218
+ "unclassified": 0,
219
+ "single": 0,
220
+ "multi": 0
221
+ },
222
+ "ONNX": {
223
+ "unclassified": 0,
224
+ "single": 0,
225
+ "multi": 0
226
+ },
227
+ "Auto": {
228
+ "unclassified": 0,
229
+ "single": 0,
230
+ "multi": 0
231
+ },
232
+ "Quantization": {
233
+ "unclassified": 0,
234
+ "single": 0,
235
+ "multi": 0
236
+ },
237
+ "Unclassified": {
238
+ "unclassified": 0,
239
+ "single": 0,
240
+ "multi": 0
241
+ }
242
+ },
243
+ "errors": 0,
244
+ "success": 365,
245
+ "skipped": 195,
246
+ "time_spent": [
247
+ 313.76,
248
+ 305.13
249
+ ],
250
+ "error": false,
251
+ "failures": {
252
+ "single": [
253
+ {
254
+ "line": "tests/models/llama/test_tokenization_llama.py::LlamaTokenizationTest::test_added_tokens_serialization",
255
+ "trace": "(line 489) OSError: You are trying to access a gated repo."
256
+ },
257
+ {
258
+ "line": "tests/models/llama/test_tokenization_llama.py::LlamaTokenizationTest::test_chat_template_return_assistant_tokens_mask",
259
+ "trace": "(line 489) OSError: You are trying to access a gated repo."
260
+ },
261
+ {
262
+ "line": "tests/models/llama/test_tokenization_llama.py::LlamaTokenizationTest::test_chat_template_return_assistant_tokens_mask_truncated",
263
+ "trace": "(line 489) OSError: You are trying to access a gated repo."
264
+ },
265
+ {
266
+ "line": "tests/models/llama/test_tokenization_llama.py::LlamaTokenizationTest::test_padding_side_in_kwargs",
267
+ "trace": "(line 489) OSError: You are trying to access a gated repo."
268
+ },
269
+ {
270
+ "line": "tests/models/llama/test_tokenization_llama.py::LlamaTokenizationTest::test_truncation_side_in_kwargs",
271
+ "trace": "(line 489) OSError: You are trying to access a gated repo."
272
+ }
273
+ ],
274
+ "multi": [
275
+ {
276
+ "line": "tests/models/llama/test_tokenization_llama.py::LlamaTokenizationTest::test_added_tokens_serialization",
277
+ "trace": "(line 489) OSError: You are trying to access a gated repo."
278
+ },
279
+ {
280
+ "line": "tests/models/llama/test_tokenization_llama.py::LlamaTokenizationTest::test_chat_template_return_assistant_tokens_mask",
281
+ "trace": "(line 489) OSError: You are trying to access a gated repo."
282
+ },
283
+ {
284
+ "line": "tests/models/llama/test_tokenization_llama.py::LlamaTokenizationTest::test_chat_template_return_assistant_tokens_mask_truncated",
285
+ "trace": "(line 489) OSError: You are trying to access a gated repo."
286
+ },
287
+ {
288
+ "line": "tests/models/llama/test_tokenization_llama.py::LlamaTokenizationTest::test_padding_side_in_kwargs",
289
+ "trace": "(line 489) OSError: You are trying to access a gated repo."
290
+ },
291
+ {
292
+ "line": "tests/models/llama/test_tokenization_llama.py::LlamaTokenizationTest::test_truncation_side_in_kwargs",
293
+ "trace": "(line 489) OSError: You are trying to access a gated repo."
294
+ }
295
+ ]
296
+ },
297
+ "job_link": {
298
+ "single": "https://github.com/huggingface/transformers/actions/runs/19962235018/job/57245316218",
299
+ "multi": "https://github.com/huggingface/transformers/actions/runs/19962235018/job/57245316189"
300
+ },
301
+ "captured_info": {}
302
+ },
303
+ "models_mistral": {
304
+ "failed": {
305
+ "PyTorch": {
306
+ "unclassified": 0,
307
+ "single": 0,
308
+ "multi": 0
309
+ },
310
+ "Tokenizers": {
311
+ "unclassified": 0,
312
+ "single": 0,
313
+ "multi": 0
314
+ },
315
+ "Pipelines": {
316
+ "unclassified": 0,
317
+ "single": 0,
318
+ "multi": 0
319
+ },
320
+ "Trainer": {
321
+ "unclassified": 0,
322
+ "single": 0,
323
+ "multi": 0
324
+ },
325
+ "ONNX": {
326
+ "unclassified": 0,
327
+ "single": 0,
328
+ "multi": 0
329
+ },
330
+ "Auto": {
331
+ "unclassified": 0,
332
+ "single": 0,
333
+ "multi": 0
334
+ },
335
+ "Quantization": {
336
+ "unclassified": 0,
337
+ "single": 0,
338
+ "multi": 0
339
+ },
340
+ "Unclassified": {
341
+ "unclassified": 0,
342
+ "single": 0,
343
+ "multi": 0
344
+ }
345
+ },
346
+ "errors": 0,
347
+ "success": 285,
348
+ "skipped": 187,
349
+ "time_spent": [
350
+ 229.96,
351
+ 230.5
352
+ ],
353
+ "error": false,
354
+ "failures": {},
355
+ "job_link": {
356
+ "single": "https://github.com/huggingface/transformers/actions/runs/19962235018/job/57245316226",
357
+ "multi": "https://github.com/huggingface/transformers/actions/runs/19962235018/job/57245316184"
358
+ },
359
+ "captured_info": {
360
+ "single": "https://github.com/huggingface/transformers/actions/runs/19962235018/job/57245316226#step:16:1",
361
+ "multi": "https://github.com/huggingface/transformers/actions/runs/19962235018/job/57245316184#step:16:1"
362
+ }
363
+ },
364
+ "models_phi": {
365
+ "failed": {
366
+ "PyTorch": {
367
+ "unclassified": 0,
368
+ "single": 0,
369
+ "multi": 0
370
+ },
371
+ "Tokenizers": {
372
+ "unclassified": 0,
373
+ "single": 0,
374
+ "multi": 0
375
+ },
376
+ "Pipelines": {
377
+ "unclassified": 0,
378
+ "single": 0,
379
+ "multi": 0
380
+ },
381
+ "Trainer": {
382
+ "unclassified": 0,
383
+ "single": 0,
384
+ "multi": 0
385
+ },
386
+ "ONNX": {
387
+ "unclassified": 0,
388
+ "single": 0,
389
+ "multi": 0
390
+ },
391
+ "Auto": {
392
+ "unclassified": 0,
393
+ "single": 0,
394
+ "multi": 0
395
+ },
396
+ "Quantization": {
397
+ "unclassified": 0,
398
+ "single": 0,
399
+ "multi": 0
400
+ },
401
+ "Unclassified": {
402
+ "unclassified": 0,
403
+ "single": 0,
404
+ "multi": 0
405
+ }
406
+ },
407
+ "errors": 0,
408
+ "success": 259,
409
+ "skipped": 193,
410
+ "time_spent": [
411
+ 140.03,
412
+ 134.94
413
+ ],
414
+ "error": false,
415
+ "failures": {},
416
+ "job_link": {
417
+ "multi": "https://github.com/huggingface/transformers/actions/runs/19962235018/job/57245316177",
418
+ "single": "https://github.com/huggingface/transformers/actions/runs/19962235018/job/57245316216"
419
+ },
420
+ "captured_info": {
421
+ "multi": "https://github.com/huggingface/transformers/actions/runs/19962235018/job/57245316177#step:16:1",
422
+ "single": "https://github.com/huggingface/transformers/actions/runs/19962235018/job/57245316216#step:16:1"
423
+ }
424
+ },
425
+ "models_phi3": {
426
+ "failed": {
427
+ "PyTorch": {
428
+ "unclassified": 0,
429
+ "single": 6,
430
+ "multi": 6
431
+ },
432
+ "Tokenizers": {
433
+ "unclassified": 0,
434
+ "single": 0,
435
+ "multi": 0
436
+ },
437
+ "Pipelines": {
438
+ "unclassified": 0,
439
+ "single": 0,
440
+ "multi": 0
441
+ },
442
+ "Trainer": {
443
+ "unclassified": 0,
444
+ "single": 0,
445
+ "multi": 0
446
+ },
447
+ "ONNX": {
448
+ "unclassified": 0,
449
+ "single": 0,
450
+ "multi": 0
451
+ },
452
+ "Auto": {
453
+ "unclassified": 0,
454
+ "single": 0,
455
+ "multi": 0
456
+ },
457
+ "Quantization": {
458
+ "unclassified": 0,
459
+ "single": 0,
460
+ "multi": 0
461
+ },
462
+ "Unclassified": {
463
+ "unclassified": 0,
464
+ "single": 0,
465
+ "multi": 0
466
+ }
467
+ },
468
+ "errors": 0,
469
+ "success": 255,
470
+ "skipped": 193,
471
+ "time_spent": [
472
+ 149.24,
473
+ 166.68
474
+ ],
475
+ "error": false,
476
+ "failures": {
477
+ "multi": [
478
+ {
479
+ "line": "tests/models/phi3/test_modeling_phi3.py::Phi3IntegrationTest::test_export_static_cache",
480
+ "trace": "(line 841) torch._dynamo.exc.Unsupported: Data-dependent branching"
481
+ },
482
+ {
483
+ "line": "tests/models/phi3/test_modeling_phi3.py::Phi3IntegrationTest::test_phi3_mini_128k_instruct_generation",
484
+ "trace": "(line 279) AttributeError"
485
+ },
486
+ {
487
+ "line": "tests/models/phi3/test_modeling_phi3.py::Phi3IntegrationTest::test_phi3_mini_128k_instruct_with_static_cache",
488
+ "trace": "(line 279) AttributeError"
489
+ },
490
+ {
491
+ "line": "tests/models/phi3/test_modeling_phi3.py::Phi3IntegrationTest::test_phi3_mini_4k_instruct_generation",
492
+ "trace": "(line 279) AttributeError"
493
+ },
494
+ {
495
+ "line": "tests/models/phi3/test_modeling_phi3.py::Phi3IntegrationTest::test_phi3_mini_4k_instruct_with_static_cache",
496
+ "trace": "(line 279) AttributeError"
497
+ },
498
+ {
499
+ "line": "tests/models/phi3/test_modeling_phi3.py::Phi3IntegrationTest::test_phi3_mini_4k_sliding_window",
500
+ "trace": "(line 319) AssertionError: Lists differ: ['1. [279 chars]: Boeuf Bourguignon is a traditional French beef stew that is'] != ['1. [279 chars]: Boeuf Bourguignon is a traditional French beef stew that']"
501
+ }
502
+ ],
503
+ "single": [
504
+ {
505
+ "line": "tests/models/phi3/test_modeling_phi3.py::Phi3IntegrationTest::test_export_static_cache",
506
+ "trace": "(line 841) torch._dynamo.exc.Unsupported: Data-dependent branching"
507
+ },
508
+ {
509
+ "line": "tests/models/phi3/test_modeling_phi3.py::Phi3IntegrationTest::test_phi3_mini_128k_instruct_generation",
510
+ "trace": "(line 279) AttributeError"
511
+ },
512
+ {
513
+ "line": "tests/models/phi3/test_modeling_phi3.py::Phi3IntegrationTest::test_phi3_mini_128k_instruct_with_static_cache",
514
+ "trace": "(line 279) AttributeError"
515
+ },
516
+ {
517
+ "line": "tests/models/phi3/test_modeling_phi3.py::Phi3IntegrationTest::test_phi3_mini_4k_instruct_generation",
518
+ "trace": "(line 279) AttributeError"
519
+ },
520
+ {
521
+ "line": "tests/models/phi3/test_modeling_phi3.py::Phi3IntegrationTest::test_phi3_mini_4k_instruct_with_static_cache",
522
+ "trace": "(line 279) AttributeError"
523
+ },
524
+ {
525
+ "line": "tests/models/phi3/test_modeling_phi3.py::Phi3IntegrationTest::test_phi3_mini_4k_sliding_window",
526
+ "trace": "(line 319) AssertionError: Lists differ: ['1. [279 chars]: Boeuf Bourguignon is a traditional French beef stew that is'] != ['1. [279 chars]: Boeuf Bourguignon is a traditional French beef stew that']"
527
+ }
528
+ ]
529
+ },
530
+ "job_link": {
531
+ "multi": "https://github.com/huggingface/transformers/actions/runs/19962235018/job/57245316292",
532
+ "single": "https://github.com/huggingface/transformers/actions/runs/19962235018/job/57245316202"
533
+ },
534
+ "captured_info": {
535
+ "multi": "https://github.com/huggingface/transformers/actions/runs/19962235018/job/57245316292#step:16:1",
536
+ "single": "https://github.com/huggingface/transformers/actions/runs/19962235018/job/57245316202#step:16:1"
537
+ }
538
+ },
539
+ "models_qwen2": {
540
+ "failed": {
541
+ "PyTorch": {
542
+ "unclassified": 0,
543
+ "single": 1,
544
+ "multi": 1
545
+ },
546
+ "Tokenizers": {
547
+ "unclassified": 0,
548
+ "single": 0,
549
+ "multi": 0
550
+ },
551
+ "Pipelines": {
552
+ "unclassified": 0,
553
+ "single": 0,
554
+ "multi": 0
555
+ },
556
+ "Trainer": {
557
+ "unclassified": 0,
558
+ "single": 0,
559
+ "multi": 0
560
+ },
561
+ "ONNX": {
562
+ "unclassified": 0,
563
+ "single": 0,
564
+ "multi": 0
565
+ },
566
+ "Auto": {
567
+ "unclassified": 0,
568
+ "single": 0,
569
+ "multi": 0
570
+ },
571
+ "Quantization": {
572
+ "unclassified": 0,
573
+ "single": 0,
574
+ "multi": 0
575
+ },
576
+ "Unclassified": {
577
+ "unclassified": 0,
578
+ "single": 0,
579
+ "multi": 0
580
+ }
581
+ },
582
+ "errors": 0,
583
+ "success": 365,
584
+ "skipped": 197,
585
+ "time_spent": [
586
+ 223.6,
587
+ 222.0
588
+ ],
589
+ "error": false,
590
+ "failures": {
591
+ "multi": [
592
+ {
593
+ "line": "tests/models/qwen2/test_modeling_qwen2.py::Qwen2IntegrationTest::test_export_static_cache",
594
+ "trace": "(line 275) AssertionError: Lists differ: ['My [35 chars], organic, gluten free, vegan, and vegetarian. I love to use'] != ['My [35 chars], organic, gluten free, vegan, and free from preservatives. I']"
595
+ }
596
+ ],
597
+ "single": [
598
+ {
599
+ "line": "tests/models/qwen2/test_modeling_qwen2.py::Qwen2IntegrationTest::test_export_static_cache",
600
+ "trace": "(line 275) AssertionError: Lists differ: ['My [35 chars], organic, gluten free, vegan, and vegetarian. I love to use'] != ['My [35 chars], organic, gluten free, vegan, and free from preservatives. I']"
601
+ }
602
+ ]
603
+ },
604
+ "job_link": {
605
+ "multi": "https://github.com/huggingface/transformers/actions/runs/19962235018/job/57245316180",
606
+ "single": "https://github.com/huggingface/transformers/actions/runs/19962235018/job/57245316200"
607
+ },
608
+ "captured_info": {
609
+ "multi": "https://github.com/huggingface/transformers/actions/runs/19962235018/job/57245316180#step:16:1",
610
+ "single": "https://github.com/huggingface/transformers/actions/runs/19962235018/job/57245316200#step:16:1"
611
+ }
612
+ },
613
+ "models_qwen2_vl": {
614
+ "failed": {
615
+ "PyTorch": {
616
+ "unclassified": 0,
617
+ "single": 0,
618
+ "multi": 1
619
+ },
620
+ "Tokenizers": {
621
+ "unclassified": 0,
622
+ "single": 0,
623
+ "multi": 0
624
+ },
625
+ "Pipelines": {
626
+ "unclassified": 0,
627
+ "single": 0,
628
+ "multi": 0
629
+ },
630
+ "Trainer": {
631
+ "unclassified": 0,
632
+ "single": 0,
633
+ "multi": 0
634
+ },
635
+ "ONNX": {
636
+ "unclassified": 0,
637
+ "single": 0,
638
+ "multi": 0
639
+ },
640
+ "Auto": {
641
+ "unclassified": 0,
642
+ "single": 0,
643
+ "multi": 0
644
+ },
645
+ "Quantization": {
646
+ "unclassified": 0,
647
+ "single": 0,
648
+ "multi": 0
649
+ },
650
+ "Unclassified": {
651
+ "unclassified": 0,
652
+ "single": 0,
653
+ "multi": 0
654
+ }
655
+ },
656
+ "errors": 0,
657
+ "success": 402,
658
+ "skipped": 239,
659
+ "time_spent": [
660
+ 279.56,
661
+ 275.36
662
+ ],
663
+ "error": false,
664
+ "failures": {
665
+ "multi": [
666
+ {
667
+ "line": "tests/models/qwen2_vl/test_modeling_qwen2_vl.py::Qwen2VLIntegrationTest::test_small_model_integration_test_batch_wo_image",
668
+ "trace": "(line 529) OSError: Qwen/Qwen2-VL-7B-Instruct does not appear to have a file named model-00001-of-00005.safetensors. Checkout 'https://huggingface.co/Qwen/Qwen2-VL-7B-Instruct/tree/main' for available files."
669
+ }
670
+ ]
671
+ },
672
+ "job_link": {
673
+ "single": "https://github.com/huggingface/transformers/actions/runs/19962235018/job/57245316204",
674
+ "multi": "https://github.com/huggingface/transformers/actions/runs/19962235018/job/57245316234"
675
+ },
676
+ "captured_info": {}
677
+ }
678
+ }