hf-transformers-bot commited on
Commit
5541a6b
·
verified ·
1 Parent(s): 34e268c

Upload 2025-12-01/runs/24334-19822280975/ci_results_run_models_gpu/model_results.json with huggingface_hub

Browse files
2025-12-01/runs/24334-19822280975/ci_results_run_models_gpu/model_results.json ADDED
@@ -0,0 +1,555 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "models_qwen2_5_omni": {
3
+ "failed": {
4
+ "PyTorch": {
5
+ "unclassified": 0,
6
+ "single": 4,
7
+ "multi": 5
8
+ },
9
+ "Tokenizers": {
10
+ "unclassified": 0,
11
+ "single": 0,
12
+ "multi": 0
13
+ },
14
+ "Pipelines": {
15
+ "unclassified": 0,
16
+ "single": 0,
17
+ "multi": 0
18
+ },
19
+ "Trainer": {
20
+ "unclassified": 0,
21
+ "single": 0,
22
+ "multi": 0
23
+ },
24
+ "ONNX": {
25
+ "unclassified": 0,
26
+ "single": 0,
27
+ "multi": 0
28
+ },
29
+ "Auto": {
30
+ "unclassified": 0,
31
+ "single": 0,
32
+ "multi": 0
33
+ },
34
+ "Quantization": {
35
+ "unclassified": 0,
36
+ "single": 0,
37
+ "multi": 0
38
+ },
39
+ "Unclassified": {
40
+ "unclassified": 0,
41
+ "single": 0,
42
+ "multi": 0
43
+ }
44
+ },
45
+ "errors": 0,
46
+ "success": 290,
47
+ "skipped": 237,
48
+ "time_spent": [
49
+ 208.56,
50
+ 267.15
51
+ ],
52
+ "error": false,
53
+ "failures": {
54
+ "multi": [
55
+ {
56
+ "line": "tests/models/qwen2_5_omni/test_modeling_qwen2_5_omni.py::Qwen2_5OmniThinkerForConditionalGenerationModelTest::test_multi_gpu_data_parallel_forward",
57
+ "trace": "(line 769) StopIteration: Caught StopIteration in replica 1 on device 1."
58
+ },
59
+ {
60
+ "line": "tests/models/qwen2_5_omni/test_modeling_qwen2_5_omni.py::Qwen2_5OmniModelIntegrationTest::test_small_model_integration_test",
61
+ "trace": "(line 692) AssertionError: \"syst[95 chars]ant\\n\" != \"syst[95 chars]ant\\nThe sound is glass shattering, and the do[22 chars]ver.\""
62
+ },
63
+ {
64
+ "line": "tests/models/qwen2_5_omni/test_modeling_qwen2_5_omni.py::Qwen2_5OmniModelIntegrationTest::test_small_model_integration_test_batch",
65
+ "trace": "(line 734) AssertionError: Lists differ: [\"sys[96 chars]ant\\n\", \"system\\nYou are a helpful assistant.\\[64 chars]t\\n\"] != [\"sys[96 chars]ant\\nThe sound is glass shattering, and the do[198 chars]er.\"]"
66
+ },
67
+ {
68
+ "line": "tests/models/qwen2_5_omni/test_modeling_qwen2_5_omni.py::Qwen2_5OmniModelIntegrationTest::test_small_model_integration_test_multiturn",
69
+ "trace": "(line 777) AssertionError: \"syst[177 chars].\\nuser\\nHow about this one?\\nassistant\\n\" != \"syst[177 chars].\\nuser\\nHow about this one?\\nassistant\\nThe sound is a cough.\""
70
+ },
71
+ {
72
+ "line": "tests/models/qwen2_5_omni/test_modeling_qwen2_5_omni.py::Qwen2_5OmniModelIntegrationTest::test_small_model_integration_test_w_audio",
73
+ "trace": "(line 829) AssertionError: 'system\\nYou are Qwen, a virtual human de[146 chars]nt\\n' != \"system\\nYou are Qwen, a virtual human de[232 chars]many\""
74
+ }
75
+ ],
76
+ "single": [
77
+ {
78
+ "line": "tests/models/qwen2_5_omni/test_modeling_qwen2_5_omni.py::Qwen2_5OmniModelIntegrationTest::test_small_model_integration_test",
79
+ "trace": "(line 692) AssertionError: \"syst[95 chars]ant\\n\" != \"syst[95 chars]ant\\nThe sound is glass shattering, and the do[22 chars]ver.\""
80
+ },
81
+ {
82
+ "line": "tests/models/qwen2_5_omni/test_modeling_qwen2_5_omni.py::Qwen2_5OmniModelIntegrationTest::test_small_model_integration_test_batch",
83
+ "trace": "(line 734) AssertionError: Lists differ: [\"sys[96 chars]ant\\n\", \"system\\nYou are a helpful assistant.\\[64 chars]t\\n\"] != [\"sys[96 chars]ant\\nThe sound is glass shattering, and the do[198 chars]er.\"]"
84
+ },
85
+ {
86
+ "line": "tests/models/qwen2_5_omni/test_modeling_qwen2_5_omni.py::Qwen2_5OmniModelIntegrationTest::test_small_model_integration_test_multiturn",
87
+ "trace": "(line 777) AssertionError: \"syst[177 chars].\\nuser\\nHow about this one?\\nassistant\\n\" != \"syst[177 chars].\\nuser\\nHow about this one?\\nassistant\\nThe sound is a cough.\""
88
+ },
89
+ {
90
+ "line": "tests/models/qwen2_5_omni/test_modeling_qwen2_5_omni.py::Qwen2_5OmniModelIntegrationTest::test_small_model_integration_test_w_audio",
91
+ "trace": "(line 829) AssertionError: 'system\\nYou are Qwen, a virtual human de[146 chars]nt\\n' != \"system\\nYou are Qwen, a virtual human de[232 chars]many\""
92
+ }
93
+ ]
94
+ },
95
+ "job_link": {
96
+ "multi": "https://github.com/huggingface/transformers/actions/runs/19822280975/job/56787800730",
97
+ "single": "https://github.com/huggingface/transformers/actions/runs/19822280975/job/56787800696"
98
+ },
99
+ "captured_info": {
100
+ "multi": "https://github.com/huggingface/transformers/actions/runs/19822280975/job/56787800730#step:16:1",
101
+ "single": "https://github.com/huggingface/transformers/actions/runs/19822280975/job/56787800696#step:16:1"
102
+ }
103
+ },
104
+ "models_qwen2_5_vl": {
105
+ "failed": {
106
+ "PyTorch": {
107
+ "unclassified": 0,
108
+ "single": 2,
109
+ "multi": 2
110
+ },
111
+ "Tokenizers": {
112
+ "unclassified": 0,
113
+ "single": 0,
114
+ "multi": 0
115
+ },
116
+ "Pipelines": {
117
+ "unclassified": 0,
118
+ "single": 0,
119
+ "multi": 0
120
+ },
121
+ "Trainer": {
122
+ "unclassified": 0,
123
+ "single": 0,
124
+ "multi": 0
125
+ },
126
+ "ONNX": {
127
+ "unclassified": 0,
128
+ "single": 0,
129
+ "multi": 0
130
+ },
131
+ "Auto": {
132
+ "unclassified": 0,
133
+ "single": 0,
134
+ "multi": 0
135
+ },
136
+ "Quantization": {
137
+ "unclassified": 0,
138
+ "single": 0,
139
+ "multi": 0
140
+ },
141
+ "Unclassified": {
142
+ "unclassified": 0,
143
+ "single": 0,
144
+ "multi": 0
145
+ }
146
+ },
147
+ "errors": 0,
148
+ "success": 313,
149
+ "skipped": 117,
150
+ "time_spent": [
151
+ 252.9,
152
+ 256.77
153
+ ],
154
+ "error": false,
155
+ "failures": {
156
+ "single": [
157
+ {
158
+ "line": "tests/models/qwen2_5_vl/test_modeling_qwen2_5_vl.py::Qwen2_5_VLIntegrationTest::test_small_model_integration_test_batch_different_resolutions",
159
+ "trace": "(line 587) AssertionError: 'syst[73 chars]ant\\n addCriterion\\nThe dog in the picture app[95 chars]h is' != 'syst[73 chars]ant\\nThe dog in the picture appears to be a La[94 chars]t in'"
160
+ },
161
+ {
162
+ "line": "tests/models/qwen2_5_vl/test_modeling_qwen2_5_vl.py::Qwen2_5_VLIntegrationTest::test_small_model_integration_test_with_video",
163
+ "trace": "(line 706) AssertionError: Lists differ: ['sys[124 chars]h a player standing on the baseline, preparing[61 chars]ts,'] != ['sys[124 chars]h a person standing on one side, preparing to [68 chars]ing']"
164
+ }
165
+ ],
166
+ "multi": [
167
+ {
168
+ "line": "tests/models/qwen2_5_vl/test_modeling_qwen2_5_vl.py::Qwen2_5_VLIntegrationTest::test_small_model_integration_test_batch_different_resolutions",
169
+ "trace": "(line 587) AssertionError: 'syst[73 chars]ant\\n addCriterion\\nThe dog in the picture app[95 chars]h is' != 'syst[73 chars]ant\\nThe dog in the picture appears to be a La[94 chars]t in'"
170
+ },
171
+ {
172
+ "line": "tests/models/qwen2_5_vl/test_modeling_qwen2_5_vl.py::Qwen2_5_VLIntegrationTest::test_small_model_integration_test_with_video",
173
+ "trace": "(line 706) AssertionError: Lists differ: ['sys[124 chars]h a player standing on the baseline, preparing[61 chars]ts,'] != ['sys[124 chars]h a person standing on one side, preparing to [68 chars]ing']"
174
+ }
175
+ ]
176
+ },
177
+ "job_link": {
178
+ "single": "https://github.com/huggingface/transformers/actions/runs/19822280975/job/56787800674",
179
+ "multi": "https://github.com/huggingface/transformers/actions/runs/19822280975/job/56787800703"
180
+ },
181
+ "captured_info": {
182
+ "single": "https://github.com/huggingface/transformers/actions/runs/19822280975/job/56787800674#step:16:1",
183
+ "multi": "https://github.com/huggingface/transformers/actions/runs/19822280975/job/56787800703#step:16:1"
184
+ }
185
+ },
186
+ "models_qwen2_vl": {
187
+ "failed": {
188
+ "PyTorch": {
189
+ "unclassified": 0,
190
+ "single": 0,
191
+ "multi": 0
192
+ },
193
+ "Tokenizers": {
194
+ "unclassified": 0,
195
+ "single": 0,
196
+ "multi": 0
197
+ },
198
+ "Pipelines": {
199
+ "unclassified": 0,
200
+ "single": 0,
201
+ "multi": 0
202
+ },
203
+ "Trainer": {
204
+ "unclassified": 0,
205
+ "single": 0,
206
+ "multi": 0
207
+ },
208
+ "ONNX": {
209
+ "unclassified": 0,
210
+ "single": 0,
211
+ "multi": 0
212
+ },
213
+ "Auto": {
214
+ "unclassified": 0,
215
+ "single": 0,
216
+ "multi": 0
217
+ },
218
+ "Quantization": {
219
+ "unclassified": 0,
220
+ "single": 0,
221
+ "multi": 0
222
+ },
223
+ "Unclassified": {
224
+ "unclassified": 0,
225
+ "single": 0,
226
+ "multi": 0
227
+ }
228
+ },
229
+ "errors": 0,
230
+ "success": 401,
231
+ "skipped": 239,
232
+ "time_spent": [
233
+ 278.63,
234
+ 281.02
235
+ ],
236
+ "error": false,
237
+ "failures": {},
238
+ "job_link": {
239
+ "single": "https://github.com/huggingface/transformers/actions/runs/19822280975/job/56787800683",
240
+ "multi": "https://github.com/huggingface/transformers/actions/runs/19822280975/job/56787800729"
241
+ },
242
+ "captured_info": {}
243
+ },
244
+ "models_qwen3_omni_moe": {
245
+ "failed": {
246
+ "PyTorch": {
247
+ "unclassified": 0,
248
+ "single": 4,
249
+ "multi": 6
250
+ },
251
+ "Tokenizers": {
252
+ "unclassified": 0,
253
+ "single": 0,
254
+ "multi": 0
255
+ },
256
+ "Pipelines": {
257
+ "unclassified": 0,
258
+ "single": 0,
259
+ "multi": 0
260
+ },
261
+ "Trainer": {
262
+ "unclassified": 0,
263
+ "single": 0,
264
+ "multi": 0
265
+ },
266
+ "ONNX": {
267
+ "unclassified": 0,
268
+ "single": 0,
269
+ "multi": 0
270
+ },
271
+ "Auto": {
272
+ "unclassified": 0,
273
+ "single": 0,
274
+ "multi": 0
275
+ },
276
+ "Quantization": {
277
+ "unclassified": 0,
278
+ "single": 0,
279
+ "multi": 0
280
+ },
281
+ "Unclassified": {
282
+ "unclassified": 0,
283
+ "single": 0,
284
+ "multi": 0
285
+ }
286
+ },
287
+ "errors": 0,
288
+ "success": 227,
289
+ "skipped": 191,
290
+ "time_spent": [
291
+ 127.61,
292
+ 132.09
293
+ ],
294
+ "error": false,
295
+ "failures": {
296
+ "single": [
297
+ {
298
+ "line": "tests/models/qwen3_omni_moe/test_modeling_qwen3_omni_moe.py::Qwen3OmniModelIntegrationTest::test_small_model_integration_test",
299
+ "trace": "(line 69) TypeError: argument 'merges': failed to extract enum PyMerges ('Merges | Filename')"
300
+ },
301
+ {
302
+ "line": "tests/models/qwen3_omni_moe/test_modeling_qwen3_omni_moe.py::Qwen3OmniModelIntegrationTest::test_small_model_integration_test_batch",
303
+ "trace": "(line 69) TypeError: argument 'merges': failed to extract enum PyMerges ('Merges | Filename')"
304
+ },
305
+ {
306
+ "line": "tests/models/qwen3_omni_moe/test_modeling_qwen3_omni_moe.py::Qwen3OmniModelIntegrationTest::test_small_model_integration_test_multiturn",
307
+ "trace": "(line 69) TypeError: argument 'merges': failed to extract enum PyMerges ('Merges | Filename')"
308
+ },
309
+ {
310
+ "line": "tests/models/qwen3_omni_moe/test_modeling_qwen3_omni_moe.py::Qwen3OmniModelIntegrationTest::test_small_model_integration_test_w_audio",
311
+ "trace": "(line 69) TypeError: argument 'merges': failed to extract enum PyMerges ('Merges | Filename')"
312
+ }
313
+ ],
314
+ "multi": [
315
+ {
316
+ "line": "tests/models/qwen3_omni_moe/test_modeling_qwen3_omni_moe.py::Qwen3OmniMoeThinkerForConditionalGenerationModelTest::test_model_parallelism",
317
+ "trace": "(line 762) RuntimeError: Expected all tensors to be on the same device, but found at least two devices, cuda:1 and cuda:0!"
318
+ },
319
+ {
320
+ "line": "tests/models/qwen3_omni_moe/test_modeling_qwen3_omni_moe.py::Qwen3OmniMoeThinkerForConditionalGenerationModelTest::test_multi_gpu_data_parallel_forward",
321
+ "trace": "(line 769) StopIteration: Caught StopIteration in replica 1 on device 1."
322
+ },
323
+ {
324
+ "line": "tests/models/qwen3_omni_moe/test_modeling_qwen3_omni_moe.py::Qwen3OmniModelIntegrationTest::test_small_model_integration_test",
325
+ "trace": "(line 69) TypeError: argument 'merges': failed to extract enum PyMerges ('Merges | Filename')"
326
+ },
327
+ {
328
+ "line": "tests/models/qwen3_omni_moe/test_modeling_qwen3_omni_moe.py::Qwen3OmniModelIntegrationTest::test_small_model_integration_test_batch",
329
+ "trace": "(line 69) TypeError: argument 'merges': failed to extract enum PyMerges ('Merges | Filename')"
330
+ },
331
+ {
332
+ "line": "tests/models/qwen3_omni_moe/test_modeling_qwen3_omni_moe.py::Qwen3OmniModelIntegrationTest::test_small_model_integration_test_multiturn",
333
+ "trace": "(line 69) TypeError: argument 'merges': failed to extract enum PyMerges ('Merges | Filename')"
334
+ },
335
+ {
336
+ "line": "tests/models/qwen3_omni_moe/test_modeling_qwen3_omni_moe.py::Qwen3OmniModelIntegrationTest::test_small_model_integration_test_w_audio",
337
+ "trace": "(line 69) TypeError: argument 'merges': failed to extract enum PyMerges ('Merges | Filename')"
338
+ }
339
+ ]
340
+ },
341
+ "job_link": {
342
+ "single": "https://github.com/huggingface/transformers/actions/runs/19822280975/job/56787800684",
343
+ "multi": "https://github.com/huggingface/transformers/actions/runs/19822280975/job/56787800725"
344
+ },
345
+ "captured_info": {}
346
+ },
347
+ "models_qwen3_vl": {
348
+ "failed": {
349
+ "PyTorch": {
350
+ "unclassified": 0,
351
+ "single": 1,
352
+ "multi": 2
353
+ },
354
+ "Tokenizers": {
355
+ "unclassified": 0,
356
+ "single": 0,
357
+ "multi": 0
358
+ },
359
+ "Pipelines": {
360
+ "unclassified": 0,
361
+ "single": 0,
362
+ "multi": 0
363
+ },
364
+ "Trainer": {
365
+ "unclassified": 0,
366
+ "single": 0,
367
+ "multi": 0
368
+ },
369
+ "ONNX": {
370
+ "unclassified": 0,
371
+ "single": 0,
372
+ "multi": 0
373
+ },
374
+ "Auto": {
375
+ "unclassified": 0,
376
+ "single": 0,
377
+ "multi": 0
378
+ },
379
+ "Quantization": {
380
+ "unclassified": 0,
381
+ "single": 0,
382
+ "multi": 0
383
+ },
384
+ "Unclassified": {
385
+ "unclassified": 0,
386
+ "single": 0,
387
+ "multi": 0
388
+ }
389
+ },
390
+ "errors": 0,
391
+ "success": 332,
392
+ "skipped": 119,
393
+ "time_spent": [
394
+ 224.54,
395
+ 225.88
396
+ ],
397
+ "error": false,
398
+ "failures": {
399
+ "multi": [
400
+ {
401
+ "line": "tests/models/qwen3_vl/test_modeling_qwen3_vl.py::Qwen3VLModelTest::test_multi_gpu_data_parallel_forward",
402
+ "trace": "(line 769) StopIteration: Caught StopIteration in replica 1 on device 1."
403
+ },
404
+ {
405
+ "line": "tests/models/qwen3_vl/test_modeling_qwen3_vl.py::Qwen3VLModelTest::test_video_forward",
406
+ "trace": "(line 2542) RuntimeError: Expected all tensors to be on the same device, but got index is on cpu, different from other tensors on cuda:0 (when checking argument in method wrapper_CUDA__index_select)"
407
+ }
408
+ ],
409
+ "single": [
410
+ {
411
+ "line": "tests/models/qwen3_vl/test_modeling_qwen3_vl.py::Qwen3VLModelTest::test_video_forward",
412
+ "trace": "(line 2542) RuntimeError: Expected all tensors to be on the same device, but got index is on cpu, different from other tensors on cuda:0 (when checking argument in method wrapper_CUDA__index_select)"
413
+ }
414
+ ]
415
+ },
416
+ "job_link": {
417
+ "multi": "https://github.com/huggingface/transformers/actions/runs/19822280975/job/56787800784",
418
+ "single": "https://github.com/huggingface/transformers/actions/runs/19822280975/job/56787800689"
419
+ },
420
+ "captured_info": {}
421
+ },
422
+ "models_qwen3_vl_moe": {
423
+ "failed": {
424
+ "PyTorch": {
425
+ "unclassified": 0,
426
+ "single": 8,
427
+ "multi": 9
428
+ },
429
+ "Tokenizers": {
430
+ "unclassified": 0,
431
+ "single": 0,
432
+ "multi": 0
433
+ },
434
+ "Pipelines": {
435
+ "unclassified": 0,
436
+ "single": 0,
437
+ "multi": 0
438
+ },
439
+ "Trainer": {
440
+ "unclassified": 0,
441
+ "single": 0,
442
+ "multi": 0
443
+ },
444
+ "ONNX": {
445
+ "unclassified": 0,
446
+ "single": 0,
447
+ "multi": 0
448
+ },
449
+ "Auto": {
450
+ "unclassified": 0,
451
+ "single": 0,
452
+ "multi": 0
453
+ },
454
+ "Quantization": {
455
+ "unclassified": 0,
456
+ "single": 0,
457
+ "multi": 0
458
+ },
459
+ "Unclassified": {
460
+ "unclassified": 0,
461
+ "single": 0,
462
+ "multi": 0
463
+ }
464
+ },
465
+ "errors": 0,
466
+ "success": 220,
467
+ "skipped": 89,
468
+ "time_spent": [
469
+ 766.14,
470
+ 923.2
471
+ ],
472
+ "error": false,
473
+ "failures": {
474
+ "multi": [
475
+ {
476
+ "line": "tests/models/qwen3_vl_moe/test_modeling_qwen3_vl_moe.py::Qwen3VLMoeModelTest::test_multi_gpu_data_parallel_forward",
477
+ "trace": "(line 769) StopIteration: Caught StopIteration in replica 1 on device 1."
478
+ },
479
+ {
480
+ "line": "tests/models/qwen3_vl_moe/test_modeling_qwen3_vl_moe.py::Qwen3VLMoeModelTest::test_video_forward",
481
+ "trace": "(line 2542) RuntimeError: Expected all tensors to be on the same device, but got index is on cpu, different from other tensors on cuda:0 (when checking argument in method wrapper_CUDA__index_select)"
482
+ },
483
+ {
484
+ "line": "tests/models/qwen3_vl_moe/test_modeling_qwen3_vl_moe.py::Qwen3VLMoeIntegrationTest::test_small_model_integration_test",
485
+ "trace": "(line 121) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 324.00 MiB. GPU 1 has a total capacity of 22.30 GiB of which 88.69 MiB is free. Process 24049 has 22.21 GiB memory in use. Of the allocated memory 21.58 GiB is allocated by PyTorch, and 226.05 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)"
486
+ },
487
+ {
488
+ "line": "tests/models/qwen3_vl_moe/test_modeling_qwen3_vl_moe.py::Qwen3VLMoeIntegrationTest::test_small_model_integration_test_batch",
489
+ "trace": "(line 124) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 648.00 MiB. GPU 0 has a total capacity of 22.30 GiB of which 482.69 MiB is free. Process 24049 has 21.82 GiB memory in use. Of the allocated memory 20.82 GiB is allocated by PyTorch, and 30.05 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)"
490
+ },
491
+ {
492
+ "line": "tests/models/qwen3_vl_moe/test_modeling_qwen3_vl_moe.py::Qwen3VLMoeIntegrationTest::test_small_model_integration_test_batch_different_resolutions",
493
+ "trace": "(line 124) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 648.00 MiB. GPU 0 has a total capacity of 22.30 GiB of which 480.69 MiB is free. Process 24049 has 21.83 GiB memory in use. Of the allocated memory 20.81 GiB is allocated by PyTorch, and 45.84 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)"
494
+ },
495
+ {
496
+ "line": "tests/models/qwen3_vl_moe/test_modeling_qwen3_vl_moe.py::Qwen3VLMoeIntegrationTest::test_small_model_integration_test_batch_wo_image",
497
+ "trace": "(line 124) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 648.00 MiB. GPU 0 has a total capacity of 22.30 GiB of which 498.69 MiB is free. Process 24049 has 21.81 GiB memory in use. Of the allocated memory 20.80 GiB is allocated by PyTorch, and 39.18 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)"
498
+ },
499
+ {
500
+ "line": "tests/models/qwen3_vl_moe/test_modeling_qwen3_vl_moe.py::Qwen3VLMoeIntegrationTest::test_small_model_integration_test_expand",
501
+ "trace": "(line 124) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 648.00 MiB. GPU 0 has a total capacity of 22.30 GiB of which 454.69 MiB is free. Process 24049 has 21.85 GiB memory in use. Of the allocated memory 20.83 GiB is allocated by PyTorch, and 43.59 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)"
502
+ },
503
+ {
504
+ "line": "tests/models/qwen3_vl_moe/test_modeling_qwen3_vl_moe.py::Qwen3VLMoeIntegrationTest::test_small_model_integration_test_expand_with_video",
505
+ "trace": "(line 117) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 2.24 GiB. GPU 0 has a total capacity of 22.30 GiB of which 1.87 GiB is free. Process 24049 has 20.42 GiB memory in use. Of the allocated memory 19.30 GiB is allocated by PyTorch, and 155.31 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)"
506
+ },
507
+ {
508
+ "line": "tests/models/qwen3_vl_moe/test_modeling_qwen3_vl_moe.py::Qwen3VLMoeIntegrationTest::test_small_model_integration_test_with_video",
509
+ "trace": "(line 117) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 2.61 GiB. GPU 0 has a total capacity of 22.30 GiB of which 1.98 GiB is free. Process 24049 has 20.32 GiB memory in use. Of the allocated memory 19.29 GiB is allocated by PyTorch, and 54.17 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)"
510
+ }
511
+ ],
512
+ "single": [
513
+ {
514
+ "line": "tests/models/qwen3_vl_moe/test_modeling_qwen3_vl_moe.py::Qwen3VLMoeModelTest::test_video_forward",
515
+ "trace": "(line 2542) RuntimeError: Expected all tensors to be on the same device, but got index is on cpu, different from other tensors on cuda:0 (when checking argument in method wrapper_CUDA__index_select)"
516
+ },
517
+ {
518
+ "line": "tests/models/qwen3_vl_moe/test_modeling_qwen3_vl_moe.py::Qwen3VLMoeIntegrationTest::test_small_model_integration_test",
519
+ "trace": "(line 385) AssertionError: \"user[91 chars]'s a wild cat species native to the grasslands[16 chars]ions\" != \"user[91 chars]'s a small wild cat native to the grasslands and steppes\""
520
+ },
521
+ {
522
+ "line": "tests/models/qwen3_vl_moe/test_modeling_qwen3_vl_moe.py::Qwen3VLMoeIntegrationTest::test_small_model_integration_test_batch",
523
+ "trace": "(line 124) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 648.00 MiB. GPU 0 has a total capacity of 22.30 GiB of which 530.69 MiB is free. Process 17313 has 21.78 GiB memory in use. Of the allocated memory 20.83 GiB is allocated by PyTorch, and 52.73 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)"
524
+ },
525
+ {
526
+ "line": "tests/models/qwen3_vl_moe/test_modeling_qwen3_vl_moe.py::Qwen3VLMoeIntegrationTest::test_small_model_integration_test_batch_different_resolutions",
527
+ "trace": "(line 124) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 648.00 MiB. GPU 0 has a total capacity of 22.30 GiB of which 544.69 MiB is free. Process 17313 has 21.76 GiB memory in use. Of the allocated memory 20.81 GiB is allocated by PyTorch, and 50.03 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)"
528
+ },
529
+ {
530
+ "line": "tests/models/qwen3_vl_moe/test_modeling_qwen3_vl_moe.py::Qwen3VLMoeIntegrationTest::test_small_model_integration_test_batch_wo_image",
531
+ "trace": "(line 124) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 648.00 MiB. GPU 0 has a total capacity of 22.30 GiB of which 554.69 MiB is free. Process 17313 has 21.75 GiB memory in use. Of the allocated memory 20.80 GiB is allocated by PyTorch, and 52.01 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)"
532
+ },
533
+ {
534
+ "line": "tests/models/qwen3_vl_moe/test_modeling_qwen3_vl_moe.py::Qwen3VLMoeIntegrationTest::test_small_model_integration_test_expand",
535
+ "trace": "(line 124) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 648.00 MiB. GPU 0 has a total capacity of 22.30 GiB of which 518.69 MiB is free. Process 17313 has 21.79 GiB memory in use. Of the allocated memory 20.84 GiB is allocated by PyTorch, and 50.02 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)"
536
+ },
537
+ {
538
+ "line": "tests/models/qwen3_vl_moe/test_modeling_qwen3_vl_moe.py::Qwen3VLMoeIntegrationTest::test_small_model_integration_test_expand_with_video",
539
+ "trace": "(line 119) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 1.68 GiB. GPU 0 has a total capacity of 22.30 GiB of which 886.69 MiB is free. Process 17313 has 21.43 GiB memory in use. Of the allocated memory 20.38 GiB is allocated by PyTorch, and 155.82 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)"
540
+ },
541
+ {
542
+ "line": "tests/models/qwen3_vl_moe/test_modeling_qwen3_vl_moe.py::Qwen3VLMoeIntegrationTest::test_small_model_integration_test_with_video",
543
+ "trace": "(line 119) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 1.96 GiB. GPU 0 has a total capacity of 22.30 GiB of which 612.69 MiB is free. Process 17313 has 21.70 GiB memory in use. Of the allocated memory 20.74 GiB is allocated by PyTorch, and 54.68 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)"
544
+ }
545
+ ]
546
+ },
547
+ "job_link": {
548
+ "multi": "https://github.com/huggingface/transformers/actions/runs/19822280975/job/56787800761",
549
+ "single": "https://github.com/huggingface/transformers/actions/runs/19822280975/job/56787800685"
550
+ },
551
+ "captured_info": {
552
+ "single": "https://github.com/huggingface/transformers/actions/runs/19822280975/job/56787800685#step:16:1"
553
+ }
554
+ }
555
+ }