hf-transformers-bot commited on
Commit
30e817e
·
verified ·
1 Parent(s): 7ec0aee

Upload 2025-12-05/runs/24810-19968726358/ci_results_run_models_gpu/model_results.json with huggingface_hub

Browse files
2025-12-05/runs/24810-19968726358/ci_results_run_models_gpu/model_results.json ADDED
@@ -0,0 +1,1138 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "models_align": {
3
+ "failed": {
4
+ "PyTorch": {
5
+ "unclassified": 0,
6
+ "single": 0,
7
+ "multi": 2
8
+ },
9
+ "Tokenizers": {
10
+ "unclassified": 0,
11
+ "single": 0,
12
+ "multi": 0
13
+ },
14
+ "Pipelines": {
15
+ "unclassified": 0,
16
+ "single": 0,
17
+ "multi": 0
18
+ },
19
+ "Trainer": {
20
+ "unclassified": 0,
21
+ "single": 0,
22
+ "multi": 0
23
+ },
24
+ "ONNX": {
25
+ "unclassified": 0,
26
+ "single": 0,
27
+ "multi": 0
28
+ },
29
+ "Auto": {
30
+ "unclassified": 0,
31
+ "single": 0,
32
+ "multi": 0
33
+ },
34
+ "Quantization": {
35
+ "unclassified": 0,
36
+ "single": 0,
37
+ "multi": 0
38
+ },
39
+ "Unclassified": {
40
+ "unclassified": 0,
41
+ "single": 0,
42
+ "multi": 0
43
+ }
44
+ },
45
+ "errors": 0,
46
+ "success": 305,
47
+ "skipped": 585,
48
+ "time_spent": [
49
+ 72.62,
50
+ 73.85
51
+ ],
52
+ "error": false,
53
+ "failures": {
54
+ "multi": [
55
+ {
56
+ "line": "tests/models/align/test_modeling_align.py::AlignTextModelTest::test_model_parallelism",
57
+ "trace": "(line 590) RuntimeError: Expected all tensors to be on the same device, but got mat2 is on cuda:1, different from other tensors on cuda:0 (when checking argument in method wrapper_CUDA_bmm)"
58
+ },
59
+ {
60
+ "line": "tests/models/align/test_modeling_align.py::AlignTextModelTest::test_multi_gpu_data_parallel_forward",
61
+ "trace": "(line 769) StopIteration: Caught StopIteration in replica 1 on device 1."
62
+ }
63
+ ]
64
+ },
65
+ "job_link": {
66
+ "multi": "https://github.com/huggingface/transformers/actions/runs/19968726358/job/57268271035",
67
+ "single": "https://github.com/huggingface/transformers/actions/runs/19968726358/job/57268270969"
68
+ },
69
+ "captured_info": {}
70
+ },
71
+ "models_altclip": {
72
+ "failed": {
73
+ "PyTorch": {
74
+ "unclassified": 0,
75
+ "single": 2,
76
+ "multi": 4
77
+ },
78
+ "Tokenizers": {
79
+ "unclassified": 0,
80
+ "single": 0,
81
+ "multi": 0
82
+ },
83
+ "Pipelines": {
84
+ "unclassified": 0,
85
+ "single": 0,
86
+ "multi": 0
87
+ },
88
+ "Trainer": {
89
+ "unclassified": 0,
90
+ "single": 0,
91
+ "multi": 0
92
+ },
93
+ "ONNX": {
94
+ "unclassified": 0,
95
+ "single": 0,
96
+ "multi": 0
97
+ },
98
+ "Auto": {
99
+ "unclassified": 0,
100
+ "single": 0,
101
+ "multi": 0
102
+ },
103
+ "Quantization": {
104
+ "unclassified": 0,
105
+ "single": 0,
106
+ "multi": 0
107
+ },
108
+ "Unclassified": {
109
+ "unclassified": 0,
110
+ "single": 0,
111
+ "multi": 0
112
+ }
113
+ },
114
+ "errors": 0,
115
+ "success": 304,
116
+ "skipped": 582,
117
+ "time_spent": [
118
+ 171.67,
119
+ 180.32
120
+ ],
121
+ "error": false,
122
+ "failures": {
123
+ "multi": [
124
+ {
125
+ "line": "tests/models/altclip/test_modeling_altclip.py::AltCLIPVisionModelTest::test_batching_equivalence",
126
+ "trace": "(line 1214) AssertionError: Batched and Single row outputs are not equal in AltCLIPVisionModel for key=last_hidden_state."
127
+ },
128
+ {
129
+ "line": "tests/models/altclip/test_modeling_altclip.py::AltCLIPTextModelTest::test_multi_gpu_data_parallel_forward",
130
+ "trace": "(line 769) StopIteration: Caught StopIteration in replica 1 on device 1."
131
+ },
132
+ {
133
+ "line": "tests/models/altclip/test_modeling_altclip.py::AltCLIPModelTest::test_batching_equivalence",
134
+ "trace": "(line 1214) AssertionError: Batched and Single row outputs are not equal in AltCLIPModel for key=logits_per_image."
135
+ },
136
+ {
137
+ "line": "tests/models/altclip/test_modeling_altclip.py::AltCLIPModelTest::test_multi_gpu_data_parallel_forward",
138
+ "trace": "(line 769) StopIteration: Caught StopIteration in replica 1 on device 1."
139
+ }
140
+ ],
141
+ "single": [
142
+ {
143
+ "line": "tests/models/altclip/test_modeling_altclip.py::AltCLIPVisionModelTest::test_batching_equivalence",
144
+ "trace": "(line 1214) AssertionError: Batched and Single row outputs are not equal in AltCLIPVisionModel for key=last_hidden_state."
145
+ },
146
+ {
147
+ "line": "tests/models/altclip/test_modeling_altclip.py::AltCLIPModelTest::test_batching_equivalence",
148
+ "trace": "(line 1214) AssertionError: Batched and Single row outputs are not equal in AltCLIPModel for key=logits_per_image."
149
+ }
150
+ ]
151
+ },
152
+ "job_link": {
153
+ "multi": "https://github.com/huggingface/transformers/actions/runs/19968726358/job/57268271041",
154
+ "single": "https://github.com/huggingface/transformers/actions/runs/19968726358/job/57268270961"
155
+ },
156
+ "captured_info": {
157
+ "multi": "https://github.com/huggingface/transformers/actions/runs/19968726358/job/57268271041#step:16:1",
158
+ "single": "https://github.com/huggingface/transformers/actions/runs/19968726358/job/57268270961#step:16:1"
159
+ }
160
+ },
161
+ "models_chinese_clip": {
162
+ "failed": {
163
+ "PyTorch": {
164
+ "unclassified": 0,
165
+ "single": 2,
166
+ "multi": 5
167
+ },
168
+ "Tokenizers": {
169
+ "unclassified": 0,
170
+ "single": 0,
171
+ "multi": 0
172
+ },
173
+ "Pipelines": {
174
+ "unclassified": 0,
175
+ "single": 0,
176
+ "multi": 0
177
+ },
178
+ "Trainer": {
179
+ "unclassified": 0,
180
+ "single": 0,
181
+ "multi": 0
182
+ },
183
+ "ONNX": {
184
+ "unclassified": 0,
185
+ "single": 0,
186
+ "multi": 0
187
+ },
188
+ "Auto": {
189
+ "unclassified": 0,
190
+ "single": 0,
191
+ "multi": 0
192
+ },
193
+ "Quantization": {
194
+ "unclassified": 0,
195
+ "single": 0,
196
+ "multi": 0
197
+ },
198
+ "Unclassified": {
199
+ "unclassified": 0,
200
+ "single": 0,
201
+ "multi": 0
202
+ }
203
+ },
204
+ "errors": 0,
205
+ "success": 387,
206
+ "skipped": 586,
207
+ "time_spent": [
208
+ 71.44,
209
+ 72.7
210
+ ],
211
+ "error": false,
212
+ "failures": {
213
+ "multi": [
214
+ {
215
+ "line": "tests/models/chinese_clip/test_modeling_chinese_clip.py::ChineseCLIPTextModelTest::test_model_parallelism",
216
+ "trace": "(line 257) RuntimeError: Expected all tensors to be on the same device, but got mat2 is on cuda:1, different from other tensors on cuda:0 (when checking argument in method wrapper_CUDA_bmm)"
217
+ },
218
+ {
219
+ "line": "tests/models/chinese_clip/test_modeling_chinese_clip.py::ChineseCLIPTextModelTest::test_multi_gpu_data_parallel_forward",
220
+ "trace": "(line 769) StopIteration: Caught StopIteration in replica 1 on device 1."
221
+ },
222
+ {
223
+ "line": "tests/models/chinese_clip/test_modeling_chinese_clip.py::ChineseCLIPVisionModelTest::test_batching_equivalence",
224
+ "trace": "(line 1214) AssertionError: Batched and Single row outputs are not equal in ChineseCLIPVisionModel for key=last_hidden_state."
225
+ },
226
+ {
227
+ "line": "tests/models/chinese_clip/test_modeling_chinese_clip.py::ChineseCLIPModelTest::test_batching_equivalence",
228
+ "trace": "(line 1214) AssertionError: Batched and Single row outputs are not equal in ChineseCLIPModel for key=vision_model_output."
229
+ },
230
+ {
231
+ "line": "tests/models/chinese_clip/test_modeling_chinese_clip.py::ChineseCLIPModelTest::test_multi_gpu_data_parallel_forward",
232
+ "trace": "(line 769) StopIteration: Caught StopIteration in replica 1 on device 1."
233
+ }
234
+ ],
235
+ "single": [
236
+ {
237
+ "line": "tests/models/chinese_clip/test_modeling_chinese_clip.py::ChineseCLIPVisionModelTest::test_batching_equivalence",
238
+ "trace": "(line 1214) AssertionError: Batched and Single row outputs are not equal in ChineseCLIPVisionModel for key=last_hidden_state."
239
+ },
240
+ {
241
+ "line": "tests/models/chinese_clip/test_modeling_chinese_clip.py::ChineseCLIPModelTest::test_batching_equivalence",
242
+ "trace": "(line 1214) AssertionError: Batched and Single row outputs are not equal in ChineseCLIPModel for key=logits_per_image."
243
+ }
244
+ ]
245
+ },
246
+ "job_link": {
247
+ "multi": "https://github.com/huggingface/transformers/actions/runs/19968726358/job/57268271054",
248
+ "single": "https://github.com/huggingface/transformers/actions/runs/19968726358/job/57268270956"
249
+ },
250
+ "captured_info": {
251
+ "multi": "https://github.com/huggingface/transformers/actions/runs/19968726358/job/57268271054#step:16:1",
252
+ "single": "https://github.com/huggingface/transformers/actions/runs/19968726358/job/57268270956#step:16:1"
253
+ }
254
+ },
255
+ "models_clap": {
256
+ "failed": {
257
+ "PyTorch": {
258
+ "unclassified": 0,
259
+ "single": 0,
260
+ "multi": 2
261
+ },
262
+ "Tokenizers": {
263
+ "unclassified": 0,
264
+ "single": 0,
265
+ "multi": 0
266
+ },
267
+ "Pipelines": {
268
+ "unclassified": 0,
269
+ "single": 0,
270
+ "multi": 0
271
+ },
272
+ "Trainer": {
273
+ "unclassified": 0,
274
+ "single": 0,
275
+ "multi": 0
276
+ },
277
+ "ONNX": {
278
+ "unclassified": 0,
279
+ "single": 0,
280
+ "multi": 0
281
+ },
282
+ "Auto": {
283
+ "unclassified": 0,
284
+ "single": 0,
285
+ "multi": 0
286
+ },
287
+ "Quantization": {
288
+ "unclassified": 0,
289
+ "single": 0,
290
+ "multi": 0
291
+ },
292
+ "Unclassified": {
293
+ "unclassified": 0,
294
+ "single": 0,
295
+ "multi": 0
296
+ }
297
+ },
298
+ "errors": 0,
299
+ "success": 330,
300
+ "skipped": 518,
301
+ "time_spent": [
302
+ 83.48,
303
+ 81.03
304
+ ],
305
+ "error": false,
306
+ "failures": {
307
+ "multi": [
308
+ {
309
+ "line": "tests/models/clap/test_modeling_clap.py::ClapTextModelTest::test_multi_gpu_data_parallel_forward",
310
+ "trace": "(line 769) StopIteration: Caught StopIteration in replica 1 on device 1."
311
+ },
312
+ {
313
+ "line": "tests/models/clap/test_modeling_clap.py::ClapModelTest::test_multi_gpu_data_parallel_forward",
314
+ "trace": "(line 769) StopIteration: Caught StopIteration in replica 1 on device 1."
315
+ }
316
+ ]
317
+ },
318
+ "job_link": {
319
+ "single": "https://github.com/huggingface/transformers/actions/runs/19968726358/job/57268270982",
320
+ "multi": "https://github.com/huggingface/transformers/actions/runs/19968726358/job/57268271049"
321
+ },
322
+ "captured_info": {}
323
+ },
324
+ "models_clvp": {
325
+ "failed": {
326
+ "PyTorch": {
327
+ "unclassified": 0,
328
+ "single": 2,
329
+ "multi": 2
330
+ },
331
+ "Tokenizers": {
332
+ "unclassified": 0,
333
+ "single": 0,
334
+ "multi": 0
335
+ },
336
+ "Pipelines": {
337
+ "unclassified": 0,
338
+ "single": 0,
339
+ "multi": 0
340
+ },
341
+ "Trainer": {
342
+ "unclassified": 0,
343
+ "single": 0,
344
+ "multi": 0
345
+ },
346
+ "ONNX": {
347
+ "unclassified": 0,
348
+ "single": 0,
349
+ "multi": 0
350
+ },
351
+ "Auto": {
352
+ "unclassified": 0,
353
+ "single": 0,
354
+ "multi": 0
355
+ },
356
+ "Quantization": {
357
+ "unclassified": 0,
358
+ "single": 0,
359
+ "multi": 0
360
+ },
361
+ "Unclassified": {
362
+ "unclassified": 0,
363
+ "single": 0,
364
+ "multi": 0
365
+ }
366
+ },
367
+ "errors": 0,
368
+ "success": 465,
369
+ "skipped": 571,
370
+ "time_spent": [
371
+ 99.78,
372
+ 107.6
373
+ ],
374
+ "error": false,
375
+ "failures": {
376
+ "single": [
377
+ {
378
+ "line": "tests/models/clvp/test_modeling_clvp.py::ClvpIntegrationTest::test_conditional_encoder",
379
+ "trace": "(line 554) AssertionError: Tensor-likes are not close!"
380
+ },
381
+ {
382
+ "line": "tests/models/clvp/test_modeling_clvp.py::ClvpIntegrationTest::test_full_model_integration",
383
+ "trace": "(line 3853) TypeError: ClvpForCausalLM("
384
+ }
385
+ ],
386
+ "multi": [
387
+ {
388
+ "line": "tests/models/clvp/test_modeling_clvp.py::ClvpIntegrationTest::test_conditional_encoder",
389
+ "trace": "(line 554) AssertionError: Tensor-likes are not close!"
390
+ },
391
+ {
392
+ "line": "tests/models/clvp/test_modeling_clvp.py::ClvpIntegrationTest::test_full_model_integration",
393
+ "trace": "(line 3853) TypeError: ClvpForCausalLM("
394
+ }
395
+ ]
396
+ },
397
+ "job_link": {
398
+ "single": "https://github.com/huggingface/transformers/actions/runs/19968726358/job/57268270955",
399
+ "multi": "https://github.com/huggingface/transformers/actions/runs/19968726358/job/57268271074"
400
+ },
401
+ "captured_info": {
402
+ "single": "https://github.com/huggingface/transformers/actions/runs/19968726358/job/57268270955#step:16:1",
403
+ "multi": "https://github.com/huggingface/transformers/actions/runs/19968726358/job/57268271074#step:16:1"
404
+ }
405
+ },
406
+ "models_falcon_mamba": {
407
+ "failed": {
408
+ "PyTorch": {
409
+ "unclassified": 0,
410
+ "single": 0,
411
+ "multi": 0
412
+ },
413
+ "Tokenizers": {
414
+ "unclassified": 0,
415
+ "single": 0,
416
+ "multi": 0
417
+ },
418
+ "Pipelines": {
419
+ "unclassified": 0,
420
+ "single": 0,
421
+ "multi": 0
422
+ },
423
+ "Trainer": {
424
+ "unclassified": 0,
425
+ "single": 0,
426
+ "multi": 0
427
+ },
428
+ "ONNX": {
429
+ "unclassified": 0,
430
+ "single": 0,
431
+ "multi": 0
432
+ },
433
+ "Auto": {
434
+ "unclassified": 0,
435
+ "single": 0,
436
+ "multi": 0
437
+ },
438
+ "Quantization": {
439
+ "unclassified": 0,
440
+ "single": 0,
441
+ "multi": 0
442
+ },
443
+ "Unclassified": {
444
+ "unclassified": 0,
445
+ "single": 0,
446
+ "multi": 0
447
+ }
448
+ },
449
+ "errors": 0,
450
+ "success": 155,
451
+ "skipped": 291,
452
+ "time_spent": [
453
+ 76.82,
454
+ 69.6
455
+ ],
456
+ "error": false,
457
+ "failures": {},
458
+ "job_link": {
459
+ "multi": "https://github.com/huggingface/transformers/actions/runs/19968726358/job/57268271051",
460
+ "single": "https://github.com/huggingface/transformers/actions/runs/19968726358/job/57268270991"
461
+ },
462
+ "captured_info": {}
463
+ },
464
+ "models_fast_vlm": {
465
+ "failed": {
466
+ "PyTorch": {
467
+ "unclassified": 0,
468
+ "single": 0,
469
+ "multi": 1
470
+ },
471
+ "Tokenizers": {
472
+ "unclassified": 0,
473
+ "single": 0,
474
+ "multi": 0
475
+ },
476
+ "Pipelines": {
477
+ "unclassified": 0,
478
+ "single": 0,
479
+ "multi": 0
480
+ },
481
+ "Trainer": {
482
+ "unclassified": 0,
483
+ "single": 0,
484
+ "multi": 0
485
+ },
486
+ "ONNX": {
487
+ "unclassified": 0,
488
+ "single": 0,
489
+ "multi": 0
490
+ },
491
+ "Auto": {
492
+ "unclassified": 0,
493
+ "single": 0,
494
+ "multi": 0
495
+ },
496
+ "Quantization": {
497
+ "unclassified": 0,
498
+ "single": 0,
499
+ "multi": 0
500
+ },
501
+ "Unclassified": {
502
+ "unclassified": 0,
503
+ "single": 0,
504
+ "multi": 0
505
+ }
506
+ },
507
+ "errors": 0,
508
+ "success": 224,
509
+ "skipped": 91,
510
+ "time_spent": [
511
+ 74.05,
512
+ 75.17
513
+ ],
514
+ "error": false,
515
+ "failures": {
516
+ "multi": [
517
+ {
518
+ "line": "tests/models/fast_vlm/test_modeling_fast_vlm.py::FastVlmForConditionalGenerationModelTest::test_multi_gpu_data_parallel_forward",
519
+ "trace": "(line 769) StopIteration: Caught StopIteration in replica 1 on device 1."
520
+ }
521
+ ]
522
+ },
523
+ "job_link": {
524
+ "multi": "https://github.com/huggingface/transformers/actions/runs/19968726358/job/57268271085",
525
+ "single": "https://github.com/huggingface/transformers/actions/runs/19968726358/job/57268270988"
526
+ },
527
+ "captured_info": {}
528
+ },
529
+ "models_internvl": {
530
+ "failed": {
531
+ "PyTorch": {
532
+ "unclassified": 0,
533
+ "single": 0,
534
+ "multi": 1
535
+ },
536
+ "Tokenizers": {
537
+ "unclassified": 0,
538
+ "single": 0,
539
+ "multi": 0
540
+ },
541
+ "Pipelines": {
542
+ "unclassified": 0,
543
+ "single": 0,
544
+ "multi": 0
545
+ },
546
+ "Trainer": {
547
+ "unclassified": 0,
548
+ "single": 0,
549
+ "multi": 0
550
+ },
551
+ "ONNX": {
552
+ "unclassified": 0,
553
+ "single": 0,
554
+ "multi": 0
555
+ },
556
+ "Auto": {
557
+ "unclassified": 0,
558
+ "single": 0,
559
+ "multi": 0
560
+ },
561
+ "Quantization": {
562
+ "unclassified": 0,
563
+ "single": 0,
564
+ "multi": 0
565
+ },
566
+ "Unclassified": {
567
+ "unclassified": 0,
568
+ "single": 0,
569
+ "multi": 0
570
+ }
571
+ },
572
+ "errors": 0,
573
+ "success": 376,
574
+ "skipped": 225,
575
+ "time_spent": [
576
+ 284.39,
577
+ 281.84
578
+ ],
579
+ "error": false,
580
+ "failures": {
581
+ "multi": [
582
+ {
583
+ "line": "tests/models/internvl/test_modeling_internvl.py::InternVLModelTest::test_multi_gpu_data_parallel_forward",
584
+ "trace": "(line 769) StopIteration: Caught StopIteration in replica 1 on device 1."
585
+ }
586
+ ]
587
+ },
588
+ "job_link": {
589
+ "single": "https://github.com/huggingface/transformers/actions/runs/19968726358/job/57268270981",
590
+ "multi": "https://github.com/huggingface/transformers/actions/runs/19968726358/job/57268271089"
591
+ },
592
+ "captured_info": {}
593
+ },
594
+ "models_layoutlm": {
595
+ "failed": {
596
+ "PyTorch": {
597
+ "unclassified": 0,
598
+ "single": 0,
599
+ "multi": 1
600
+ },
601
+ "Tokenizers": {
602
+ "unclassified": 0,
603
+ "single": 0,
604
+ "multi": 0
605
+ },
606
+ "Pipelines": {
607
+ "unclassified": 0,
608
+ "single": 0,
609
+ "multi": 0
610
+ },
611
+ "Trainer": {
612
+ "unclassified": 0,
613
+ "single": 0,
614
+ "multi": 0
615
+ },
616
+ "ONNX": {
617
+ "unclassified": 0,
618
+ "single": 0,
619
+ "multi": 0
620
+ },
621
+ "Auto": {
622
+ "unclassified": 0,
623
+ "single": 0,
624
+ "multi": 0
625
+ },
626
+ "Quantization": {
627
+ "unclassified": 0,
628
+ "single": 0,
629
+ "multi": 0
630
+ },
631
+ "Unclassified": {
632
+ "unclassified": 0,
633
+ "single": 0,
634
+ "multi": 0
635
+ }
636
+ },
637
+ "errors": 0,
638
+ "success": 109,
639
+ "skipped": 242,
640
+ "time_spent": [
641
+ 32.2,
642
+ 30.73
643
+ ],
644
+ "error": false,
645
+ "failures": {
646
+ "multi": [
647
+ {
648
+ "line": "tests/models/layoutlm/test_modeling_layoutlm.py::LayoutLMModelTest::test_multi_gpu_data_parallel_forward",
649
+ "trace": "(line 769) StopIteration: Caught StopIteration in replica 1 on device 1."
650
+ }
651
+ ]
652
+ },
653
+ "job_link": {
654
+ "multi": "https://github.com/huggingface/transformers/actions/runs/19968726358/job/57268271124",
655
+ "single": "https://github.com/huggingface/transformers/actions/runs/19968726358/job/57268271025"
656
+ },
657
+ "captured_info": {}
658
+ },
659
+ "models_layoutlmv3": {
660
+ "failed": {
661
+ "PyTorch": {
662
+ "unclassified": 0,
663
+ "single": 0,
664
+ "multi": 0
665
+ },
666
+ "Tokenizers": {
667
+ "unclassified": 0,
668
+ "single": 0,
669
+ "multi": 0
670
+ },
671
+ "Pipelines": {
672
+ "unclassified": 0,
673
+ "single": 0,
674
+ "multi": 0
675
+ },
676
+ "Trainer": {
677
+ "unclassified": 0,
678
+ "single": 0,
679
+ "multi": 0
680
+ },
681
+ "ONNX": {
682
+ "unclassified": 0,
683
+ "single": 0,
684
+ "multi": 0
685
+ },
686
+ "Auto": {
687
+ "unclassified": 0,
688
+ "single": 0,
689
+ "multi": 0
690
+ },
691
+ "Quantization": {
692
+ "unclassified": 0,
693
+ "single": 0,
694
+ "multi": 0
695
+ },
696
+ "Unclassified": {
697
+ "unclassified": 0,
698
+ "single": 0,
699
+ "multi": 0
700
+ }
701
+ },
702
+ "errors": 0,
703
+ "success": 298,
704
+ "skipped": 362,
705
+ "time_spent": [
706
+ 101.81,
707
+ 103.67
708
+ ],
709
+ "error": false,
710
+ "failures": {},
711
+ "job_link": {
712
+ "single": "https://github.com/huggingface/transformers/actions/runs/19968726358/job/57268270995",
713
+ "multi": "https://github.com/huggingface/transformers/actions/runs/19968726358/job/57268271148"
714
+ },
715
+ "captured_info": {}
716
+ },
717
+ "models_lilt": {
718
+ "failed": {
719
+ "PyTorch": {
720
+ "unclassified": 0,
721
+ "single": 0,
722
+ "multi": 2
723
+ },
724
+ "Tokenizers": {
725
+ "unclassified": 0,
726
+ "single": 0,
727
+ "multi": 0
728
+ },
729
+ "Pipelines": {
730
+ "unclassified": 0,
731
+ "single": 0,
732
+ "multi": 0
733
+ },
734
+ "Trainer": {
735
+ "unclassified": 0,
736
+ "single": 0,
737
+ "multi": 0
738
+ },
739
+ "ONNX": {
740
+ "unclassified": 0,
741
+ "single": 0,
742
+ "multi": 0
743
+ },
744
+ "Auto": {
745
+ "unclassified": 0,
746
+ "single": 0,
747
+ "multi": 0
748
+ },
749
+ "Quantization": {
750
+ "unclassified": 0,
751
+ "single": 0,
752
+ "multi": 0
753
+ },
754
+ "Unclassified": {
755
+ "unclassified": 0,
756
+ "single": 0,
757
+ "multi": 0
758
+ }
759
+ },
760
+ "errors": 0,
761
+ "success": 120,
762
+ "skipped": 222,
763
+ "time_spent": [
764
+ 32.39,
765
+ 33.77
766
+ ],
767
+ "error": false,
768
+ "failures": {
769
+ "multi": [
770
+ {
771
+ "line": "tests/models/lilt/test_modeling_lilt.py::LiltModelTest::test_model_parallelism",
772
+ "trace": "(line 162) RuntimeError: Expected all tensors to be on the same device, but got tensors is on cuda:1, different from other tensors on cuda:0 (when checking argument in method wrapper_CUDA_cat)"
773
+ },
774
+ {
775
+ "line": "tests/models/lilt/test_modeling_lilt.py::LiltModelTest::test_multi_gpu_data_parallel_forward",
776
+ "trace": "(line 769) StopIteration: Caught StopIteration in replica 1 on device 1."
777
+ }
778
+ ]
779
+ },
780
+ "job_link": {
781
+ "single": "https://github.com/huggingface/transformers/actions/runs/19968726358/job/57268271007",
782
+ "multi": "https://github.com/huggingface/transformers/actions/runs/19968726358/job/57268271127"
783
+ },
784
+ "captured_info": {}
785
+ },
786
+ "models_mamba": {
787
+ "failed": {
788
+ "PyTorch": {
789
+ "unclassified": 0,
790
+ "single": 1,
791
+ "multi": 1
792
+ },
793
+ "Tokenizers": {
794
+ "unclassified": 0,
795
+ "single": 0,
796
+ "multi": 0
797
+ },
798
+ "Pipelines": {
799
+ "unclassified": 0,
800
+ "single": 0,
801
+ "multi": 0
802
+ },
803
+ "Trainer": {
804
+ "unclassified": 0,
805
+ "single": 0,
806
+ "multi": 0
807
+ },
808
+ "ONNX": {
809
+ "unclassified": 0,
810
+ "single": 0,
811
+ "multi": 0
812
+ },
813
+ "Auto": {
814
+ "unclassified": 0,
815
+ "single": 0,
816
+ "multi": 0
817
+ },
818
+ "Quantization": {
819
+ "unclassified": 0,
820
+ "single": 0,
821
+ "multi": 0
822
+ },
823
+ "Unclassified": {
824
+ "unclassified": 0,
825
+ "single": 0,
826
+ "multi": 0
827
+ }
828
+ },
829
+ "errors": 0,
830
+ "success": 166,
831
+ "skipped": 292,
832
+ "time_spent": [
833
+ 94.22,
834
+ 101.44
835
+ ],
836
+ "error": false,
837
+ "failures": {
838
+ "multi": [
839
+ {
840
+ "line": "tests/models/mamba/test_modeling_mamba.py::MambaIntegrationTests::test_compile_mamba_cache",
841
+ "trace": "(line 841) torch._dynamo.exc.Unsupported: Attempted to call function marked as skipped"
842
+ }
843
+ ],
844
+ "single": [
845
+ {
846
+ "line": "tests/models/mamba/test_modeling_mamba.py::MambaIntegrationTests::test_compile_mamba_cache",
847
+ "trace": "(line 841) torch._dynamo.exc.Unsupported: Attempted to call function marked as skipped"
848
+ }
849
+ ]
850
+ },
851
+ "job_link": {
852
+ "multi": "https://github.com/huggingface/transformers/actions/runs/19968726358/job/57268271108",
853
+ "single": "https://github.com/huggingface/transformers/actions/runs/19968726358/job/57268271026"
854
+ },
855
+ "captured_info": {}
856
+ },
857
+ "models_markuplm": {
858
+ "failed": {
859
+ "PyTorch": {
860
+ "unclassified": 0,
861
+ "single": 1,
862
+ "multi": 2
863
+ },
864
+ "Tokenizers": {
865
+ "unclassified": 0,
866
+ "single": 0,
867
+ "multi": 0
868
+ },
869
+ "Pipelines": {
870
+ "unclassified": 0,
871
+ "single": 0,
872
+ "multi": 0
873
+ },
874
+ "Trainer": {
875
+ "unclassified": 0,
876
+ "single": 0,
877
+ "multi": 0
878
+ },
879
+ "ONNX": {
880
+ "unclassified": 0,
881
+ "single": 0,
882
+ "multi": 0
883
+ },
884
+ "Auto": {
885
+ "unclassified": 0,
886
+ "single": 0,
887
+ "multi": 0
888
+ },
889
+ "Quantization": {
890
+ "unclassified": 0,
891
+ "single": 0,
892
+ "multi": 0
893
+ },
894
+ "Unclassified": {
895
+ "unclassified": 0,
896
+ "single": 0,
897
+ "multi": 0
898
+ }
899
+ },
900
+ "errors": 0,
901
+ "success": 263,
902
+ "skipped": 248,
903
+ "time_spent": [
904
+ 61.43,
905
+ 64.74
906
+ ],
907
+ "error": false,
908
+ "failures": {
909
+ "single": [
910
+ {
911
+ "line": "tests/models/markuplm/test_modeling_markuplm.py::MarkupLMModelIntegrationTest::test_forward_pass_no_head",
912
+ "trace": "(line 76) OSError: Could not create safetensors conversion PR. The repo does not appear to have a file named pytorch_model.bin or model.safetensors.If you are loading with variant, use `use_safetensors=False` to load the original model."
913
+ }
914
+ ],
915
+ "multi": [
916
+ {
917
+ "line": "tests/models/markuplm/test_modeling_markuplm.py::MarkupLMModelTest::test_multi_gpu_data_parallel_forward",
918
+ "trace": "(line 769) StopIteration: Caught StopIteration in replica 1 on device 1."
919
+ },
920
+ {
921
+ "line": "tests/models/markuplm/test_modeling_markuplm.py::MarkupLMModelIntegrationTest::test_forward_pass_no_head",
922
+ "trace": "(line 76) OSError: Could not create safetensors conversion PR. The repo does not appear to have a file named pytorch_model.bin or model.safetensors.If you are loading with variant, use `use_safetensors=False` to load the original model."
923
+ }
924
+ ]
925
+ },
926
+ "job_link": {
927
+ "single": "https://github.com/huggingface/transformers/actions/runs/19968726358/job/57268270987",
928
+ "multi": "https://github.com/huggingface/transformers/actions/runs/19968726358/job/57268271159"
929
+ },
930
+ "captured_info": {}
931
+ },
932
+ "models_mlcd": {
933
+ "failed": {
934
+ "PyTorch": {
935
+ "unclassified": 0,
936
+ "single": 1,
937
+ "multi": 1
938
+ },
939
+ "Tokenizers": {
940
+ "unclassified": 0,
941
+ "single": 0,
942
+ "multi": 0
943
+ },
944
+ "Pipelines": {
945
+ "unclassified": 0,
946
+ "single": 0,
947
+ "multi": 0
948
+ },
949
+ "Trainer": {
950
+ "unclassified": 0,
951
+ "single": 0,
952
+ "multi": 0
953
+ },
954
+ "ONNX": {
955
+ "unclassified": 0,
956
+ "single": 0,
957
+ "multi": 0
958
+ },
959
+ "Auto": {
960
+ "unclassified": 0,
961
+ "single": 0,
962
+ "multi": 0
963
+ },
964
+ "Quantization": {
965
+ "unclassified": 0,
966
+ "single": 0,
967
+ "multi": 0
968
+ },
969
+ "Unclassified": {
970
+ "unclassified": 0,
971
+ "single": 0,
972
+ "multi": 0
973
+ }
974
+ },
975
+ "errors": 0,
976
+ "success": 150,
977
+ "skipped": 66,
978
+ "time_spent": [
979
+ 47.6,
980
+ 46.72
981
+ ],
982
+ "error": false,
983
+ "failures": {
984
+ "multi": [
985
+ {
986
+ "line": "tests/models/mlcd/test_modeling_mlcd.py::MLCDVisionModelTest::test_batching_equivalence",
987
+ "trace": "(line 1214) AssertionError: Batched and Single row outputs are not equal in MLCDVisionModel for key=last_hidden_state."
988
+ }
989
+ ],
990
+ "single": [
991
+ {
992
+ "line": "tests/models/mlcd/test_modeling_mlcd.py::MLCDVisionModelTest::test_batching_equivalence",
993
+ "trace": "(line 1214) AssertionError: Batched and Single row outputs are not equal in MLCDVisionModel for key=last_hidden_state."
994
+ }
995
+ ]
996
+ },
997
+ "job_link": {
998
+ "multi": "https://github.com/huggingface/transformers/actions/runs/19968726358/job/57268271151",
999
+ "single": "https://github.com/huggingface/transformers/actions/runs/19968726358/job/57268271093"
1000
+ },
1001
+ "captured_info": {
1002
+ "multi": "https://github.com/huggingface/transformers/actions/runs/19968726358/job/57268271151#step:16:1",
1003
+ "single": "https://github.com/huggingface/transformers/actions/runs/19968726358/job/57268271093#step:16:1"
1004
+ }
1005
+ },
1006
+ "models_poolformer": {
1007
+ "failed": {
1008
+ "PyTorch": {
1009
+ "unclassified": 0,
1010
+ "single": 1,
1011
+ "multi": 1
1012
+ },
1013
+ "Tokenizers": {
1014
+ "unclassified": 0,
1015
+ "single": 0,
1016
+ "multi": 0
1017
+ },
1018
+ "Pipelines": {
1019
+ "unclassified": 0,
1020
+ "single": 0,
1021
+ "multi": 0
1022
+ },
1023
+ "Trainer": {
1024
+ "unclassified": 0,
1025
+ "single": 0,
1026
+ "multi": 0
1027
+ },
1028
+ "ONNX": {
1029
+ "unclassified": 0,
1030
+ "single": 0,
1031
+ "multi": 0
1032
+ },
1033
+ "Auto": {
1034
+ "unclassified": 0,
1035
+ "single": 0,
1036
+ "multi": 0
1037
+ },
1038
+ "Quantization": {
1039
+ "unclassified": 0,
1040
+ "single": 0,
1041
+ "multi": 0
1042
+ },
1043
+ "Unclassified": {
1044
+ "unclassified": 0,
1045
+ "single": 0,
1046
+ "multi": 0
1047
+ }
1048
+ },
1049
+ "errors": 0,
1050
+ "success": 166,
1051
+ "skipped": 260,
1052
+ "time_spent": [
1053
+ 41.11,
1054
+ 41.86
1055
+ ],
1056
+ "error": false,
1057
+ "failures": {
1058
+ "single": [
1059
+ {
1060
+ "line": "tests/models/poolformer/test_modeling_poolformer.py::PoolFormerModelIntegrationTest::test_inference_image_classification_head",
1061
+ "trace": "(line 245) AssertionError: Tensor-likes are not close!"
1062
+ }
1063
+ ],
1064
+ "multi": [
1065
+ {
1066
+ "line": "tests/models/poolformer/test_modeling_poolformer.py::PoolFormerModelIntegrationTest::test_inference_image_classification_head",
1067
+ "trace": "(line 245) AssertionError: Tensor-likes are not close!"
1068
+ }
1069
+ ]
1070
+ },
1071
+ "job_link": {
1072
+ "single": "https://github.com/huggingface/transformers/actions/runs/19968726358/job/57268271045",
1073
+ "multi": "https://github.com/huggingface/transformers/actions/runs/19968726358/job/57268271134"
1074
+ },
1075
+ "captured_info": {
1076
+ "single": "https://github.com/huggingface/transformers/actions/runs/19968726358/job/57268271045#step:16:1",
1077
+ "multi": "https://github.com/huggingface/transformers/actions/runs/19968726358/job/57268271134#step:16:1"
1078
+ }
1079
+ },
1080
+ "models_siglip": {
1081
+ "failed": {
1082
+ "PyTorch": {
1083
+ "unclassified": 0,
1084
+ "single": 0,
1085
+ "multi": 0
1086
+ },
1087
+ "Tokenizers": {
1088
+ "unclassified": 0,
1089
+ "single": 0,
1090
+ "multi": 0
1091
+ },
1092
+ "Pipelines": {
1093
+ "unclassified": 0,
1094
+ "single": 0,
1095
+ "multi": 0
1096
+ },
1097
+ "Trainer": {
1098
+ "unclassified": 0,
1099
+ "single": 0,
1100
+ "multi": 0
1101
+ },
1102
+ "ONNX": {
1103
+ "unclassified": 0,
1104
+ "single": 0,
1105
+ "multi": 0
1106
+ },
1107
+ "Auto": {
1108
+ "unclassified": 0,
1109
+ "single": 0,
1110
+ "multi": 0
1111
+ },
1112
+ "Quantization": {
1113
+ "unclassified": 0,
1114
+ "single": 0,
1115
+ "multi": 0
1116
+ },
1117
+ "Unclassified": {
1118
+ "unclassified": 0,
1119
+ "single": 0,
1120
+ "multi": 0
1121
+ }
1122
+ },
1123
+ "errors": 0,
1124
+ "success": 758,
1125
+ "skipped": 522,
1126
+ "time_spent": [
1127
+ 147.13,
1128
+ 148.42
1129
+ ],
1130
+ "error": false,
1131
+ "failures": {},
1132
+ "job_link": {
1133
+ "single": "https://github.com/huggingface/transformers/actions/runs/19968726358/job/57268271061",
1134
+ "multi": "https://github.com/huggingface/transformers/actions/runs/19968726358/job/57268271095"
1135
+ },
1136
+ "captured_info": {}
1137
+ }
1138
+ }