hf-transformers-bot commited on
Commit
9972d27
·
verified ·
1 Parent(s): e9aeb2d

Upload 2025-12-22/runs/25847-20422991545/ci_results_run_models_gpu/model_results.json with huggingface_hub

Browse files
2025-12-22/runs/25847-20422991545/ci_results_run_models_gpu/model_results.json ADDED
@@ -0,0 +1,332 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "models_pe_audio": {
3
+ "failed": {
4
+ "PyTorch": {
5
+ "unclassified": 0,
6
+ "single": 3,
7
+ "multi": 5
8
+ },
9
+ "Tokenizers": {
10
+ "unclassified": 0,
11
+ "single": 0,
12
+ "multi": 0
13
+ },
14
+ "Pipelines": {
15
+ "unclassified": 0,
16
+ "single": 0,
17
+ "multi": 0
18
+ },
19
+ "Trainer": {
20
+ "unclassified": 0,
21
+ "single": 0,
22
+ "multi": 0
23
+ },
24
+ "ONNX": {
25
+ "unclassified": 0,
26
+ "single": 0,
27
+ "multi": 0
28
+ },
29
+ "Auto": {
30
+ "unclassified": 0,
31
+ "single": 0,
32
+ "multi": 0
33
+ },
34
+ "Quantization": {
35
+ "unclassified": 0,
36
+ "single": 0,
37
+ "multi": 0
38
+ },
39
+ "Unclassified": {
40
+ "unclassified": 0,
41
+ "single": 0,
42
+ "multi": 0
43
+ }
44
+ },
45
+ "errors": 0,
46
+ "success": 228,
47
+ "skipped": 216,
48
+ "time_spent": [
49
+ 100.71,
50
+ 102.33
51
+ ],
52
+ "error": false,
53
+ "failures": {
54
+ "single": [
55
+ {
56
+ "line": "tests/models/pe_audio/test_modeling_pe_audio.py::PeAudioEncoderTest::test_all_tensors_are_parameter_or_buffer",
57
+ "trace": "(line 366) RuntimeError: Expected all tensors to be on the same device, but got weight is on cpu, different from other tensors on cuda:0 (when checking argument in method wrapper_CUDA___slow_conv2d_forward)"
58
+ },
59
+ {
60
+ "line": "tests/models/pe_audio/test_modeling_pe_audio.py::PeAudioEncoderTest::test_sdpa_can_dispatch_on_flash",
61
+ "trace": "(line 92) RuntimeError: No available kernel. Aborting execution."
62
+ },
63
+ {
64
+ "line": "tests/models/pe_audio/test_modeling_pe_audio.py::PeAudioModelTest::test_all_tensors_are_parameter_or_buffer",
65
+ "trace": "(line 366) RuntimeError: Expected all tensors to be on the same device, but got weight is on cpu, different from other tensors on cuda:0 (when checking argument in method wrapper_CUDA___slow_conv2d_forward)"
66
+ }
67
+ ],
68
+ "multi": [
69
+ {
70
+ "line": "tests/models/pe_audio/test_modeling_pe_audio.py::PeAudioEncoderTest::test_all_tensors_are_parameter_or_buffer",
71
+ "trace": "(line 366) RuntimeError: Expected all tensors to be on the same device, but got weight is on cpu, different from other tensors on cuda:0 (when checking argument in method wrapper_CUDA___slow_conv2d_forward)"
72
+ },
73
+ {
74
+ "line": "tests/models/pe_audio/test_modeling_pe_audio.py::PeAudioEncoderTest::test_sdpa_can_dispatch_on_flash",
75
+ "trace": "(line 92) RuntimeError: No available kernel. Aborting execution."
76
+ },
77
+ {
78
+ "line": "tests/models/pe_audio/test_modeling_pe_audio.py::PeAudioModelTest::test_all_tensors_are_parameter_or_buffer",
79
+ "trace": "(line 366) RuntimeError: Expected all tensors to be on the same device, but got weight is on cpu, different from other tensors on cuda:0 (when checking argument in method wrapper_CUDA___slow_conv2d_forward)"
80
+ },
81
+ {
82
+ "line": "tests/models/pe_audio/test_modeling_pe_audio.py::PeAudioModelTest::test_model_parallelism",
83
+ "trace": "(line 749) RuntimeError: Expected all tensors to be on the same device, but found at least two devices, cuda:1 and cuda:0!"
84
+ },
85
+ {
86
+ "line": "tests/models/pe_audio/test_modeling_pe_audio.py::PeAudioModelTest::test_multi_gpu_data_parallel_forward",
87
+ "trace": "(line 769) StopIteration: Caught StopIteration in replica 1 on device 1."
88
+ }
89
+ ]
90
+ },
91
+ "job_link": {
92
+ "single": "https://github.com/huggingface/transformers/actions/runs/20422991545/job/58678187772",
93
+ "multi": "https://github.com/huggingface/transformers/actions/runs/20422991545/job/58678187768"
94
+ },
95
+ "captured_info": {}
96
+ },
97
+ "models_pe_audio_video": {
98
+ "failed": {
99
+ "PyTorch": {
100
+ "unclassified": 0,
101
+ "single": 1,
102
+ "multi": 2
103
+ },
104
+ "Tokenizers": {
105
+ "unclassified": 0,
106
+ "single": 0,
107
+ "multi": 0
108
+ },
109
+ "Pipelines": {
110
+ "unclassified": 0,
111
+ "single": 0,
112
+ "multi": 0
113
+ },
114
+ "Trainer": {
115
+ "unclassified": 0,
116
+ "single": 0,
117
+ "multi": 0
118
+ },
119
+ "ONNX": {
120
+ "unclassified": 0,
121
+ "single": 0,
122
+ "multi": 0
123
+ },
124
+ "Auto": {
125
+ "unclassified": 0,
126
+ "single": 0,
127
+ "multi": 0
128
+ },
129
+ "Quantization": {
130
+ "unclassified": 0,
131
+ "single": 0,
132
+ "multi": 0
133
+ },
134
+ "Unclassified": {
135
+ "unclassified": 0,
136
+ "single": 0,
137
+ "multi": 0
138
+ }
139
+ },
140
+ "errors": 0,
141
+ "success": 132,
142
+ "skipped": 95,
143
+ "time_spent": [
144
+ 36.32,
145
+ 35.2
146
+ ],
147
+ "error": false,
148
+ "failures": {
149
+ "multi": [
150
+ {
151
+ "line": "tests/models/pe_audio_video/test_modeling_pe_audio_video.py::PeAudioVideoEncoderTest::test_all_tensors_are_parameter_or_buffer",
152
+ "trace": "(line 366) RuntimeError: Expected all tensors to be on the same device, but got weight is on cpu, different from other tensors on cuda:0 (when checking argument in method wrapper_CUDA___slow_conv2d_forward)"
153
+ },
154
+ {
155
+ "line": "tests/models/pe_audio_video/test_modeling_pe_audio_video.py::PeAudioVideoEncoderTest::test_multi_gpu_data_parallel_forward",
156
+ "trace": "(line 769) StopIteration: Caught StopIteration in replica 1 on device 1."
157
+ }
158
+ ],
159
+ "single": [
160
+ {
161
+ "line": "tests/models/pe_audio_video/test_modeling_pe_audio_video.py::PeAudioVideoEncoderTest::test_all_tensors_are_parameter_or_buffer",
162
+ "trace": "(line 366) RuntimeError: Expected all tensors to be on the same device, but got weight is on cpu, different from other tensors on cuda:0 (when checking argument in method wrapper_CUDA___slow_conv2d_forward)"
163
+ }
164
+ ]
165
+ },
166
+ "job_link": {
167
+ "multi": "https://github.com/huggingface/transformers/actions/runs/20422991545/job/58678187775",
168
+ "single": "https://github.com/huggingface/transformers/actions/runs/20422991545/job/58678187776"
169
+ },
170
+ "captured_info": {}
171
+ },
172
+ "models_pe_video": {
173
+ "failed": {
174
+ "PyTorch": {
175
+ "unclassified": 0,
176
+ "single": 2,
177
+ "multi": 5
178
+ },
179
+ "Tokenizers": {
180
+ "unclassified": 0,
181
+ "single": 0,
182
+ "multi": 0
183
+ },
184
+ "Pipelines": {
185
+ "unclassified": 0,
186
+ "single": 0,
187
+ "multi": 0
188
+ },
189
+ "Trainer": {
190
+ "unclassified": 0,
191
+ "single": 0,
192
+ "multi": 0
193
+ },
194
+ "ONNX": {
195
+ "unclassified": 0,
196
+ "single": 0,
197
+ "multi": 0
198
+ },
199
+ "Auto": {
200
+ "unclassified": 0,
201
+ "single": 0,
202
+ "multi": 0
203
+ },
204
+ "Quantization": {
205
+ "unclassified": 0,
206
+ "single": 0,
207
+ "multi": 0
208
+ },
209
+ "Unclassified": {
210
+ "unclassified": 0,
211
+ "single": 0,
212
+ "multi": 0
213
+ }
214
+ },
215
+ "errors": 0,
216
+ "success": 210,
217
+ "skipped": 237,
218
+ "time_spent": [
219
+ 71.04,
220
+ 73.0
221
+ ],
222
+ "error": false,
223
+ "failures": {
224
+ "single": [
225
+ {
226
+ "line": "tests/models/pe_video/test_modeling_pe_video.py::PeVideoEncoderTest::test_all_tensors_are_parameter_or_buffer",
227
+ "trace": "(line 858) RuntimeError: Expected all tensors to be on the same device, but found at least two devices, cuda:0 and cpu!"
228
+ },
229
+ {
230
+ "line": "tests/models/pe_video/test_modeling_pe_video.py::PeVideoModelTest::test_all_tensors_are_parameter_or_buffer",
231
+ "trace": "(line 858) RuntimeError: Expected all tensors to be on the same device, but found at least two devices, cuda:0 and cpu!"
232
+ }
233
+ ],
234
+ "multi": [
235
+ {
236
+ "line": "tests/models/pe_video/test_modeling_pe_video.py::PeVideoEncoderTest::test_all_tensors_are_parameter_or_buffer",
237
+ "trace": "(line 858) RuntimeError: Expected all tensors to be on the same device, but found at least two devices, cuda:0 and cpu!"
238
+ },
239
+ {
240
+ "line": "tests/models/pe_video/test_modeling_pe_video.py::PeVideoEncoderTest::test_multi_gpu_data_parallel_forward",
241
+ "trace": "(line 769) StopIteration: Caught StopIteration in replica 1 on device 1."
242
+ },
243
+ {
244
+ "line": "tests/models/pe_video/test_modeling_pe_video.py::PeVideoModelTest::test_all_tensors_are_parameter_or_buffer",
245
+ "trace": "(line 858) RuntimeError: Expected all tensors to be on the same device, but found at least two devices, cuda:0 and cpu!"
246
+ },
247
+ {
248
+ "line": "tests/models/pe_video/test_modeling_pe_video.py::PeVideoModelTest::test_model_parallelism",
249
+ "trace": "(line 360) RuntimeError: Expected all tensors to be on the same device, but found at least two devices, cuda:0 and cuda:1!"
250
+ },
251
+ {
252
+ "line": "tests/models/pe_video/test_modeling_pe_video.py::PeVideoModelTest::test_multi_gpu_data_parallel_forward",
253
+ "trace": "(line 769) StopIteration: Caught StopIteration in replica 1 on device 1."
254
+ }
255
+ ]
256
+ },
257
+ "job_link": {
258
+ "single": "https://github.com/huggingface/transformers/actions/runs/20422991545/job/58678187771",
259
+ "multi": "https://github.com/huggingface/transformers/actions/runs/20422991545/job/58678187774"
260
+ },
261
+ "captured_info": {}
262
+ },
263
+ "models_timm_wrapper": {
264
+ "failed": {
265
+ "PyTorch": {
266
+ "unclassified": 0,
267
+ "single": 0,
268
+ "multi": 2
269
+ },
270
+ "Tokenizers": {
271
+ "unclassified": 0,
272
+ "single": 0,
273
+ "multi": 0
274
+ },
275
+ "Pipelines": {
276
+ "unclassified": 0,
277
+ "single": 0,
278
+ "multi": 0
279
+ },
280
+ "Trainer": {
281
+ "unclassified": 0,
282
+ "single": 0,
283
+ "multi": 0
284
+ },
285
+ "ONNX": {
286
+ "unclassified": 0,
287
+ "single": 0,
288
+ "multi": 0
289
+ },
290
+ "Auto": {
291
+ "unclassified": 0,
292
+ "single": 0,
293
+ "multi": 0
294
+ },
295
+ "Quantization": {
296
+ "unclassified": 0,
297
+ "single": 0,
298
+ "multi": 0
299
+ },
300
+ "Unclassified": {
301
+ "unclassified": 0,
302
+ "single": 0,
303
+ "multi": 0
304
+ }
305
+ },
306
+ "errors": 0,
307
+ "success": 124,
308
+ "skipped": 248,
309
+ "time_spent": [
310
+ 47.23,
311
+ 49.87
312
+ ],
313
+ "error": false,
314
+ "failures": {
315
+ "multi": [
316
+ {
317
+ "line": "tests/models/timm_wrapper/test_modeling_timm_wrapper.py::TimmWrapperModelTest::test_model_parallelism",
318
+ "trace": "(line 268) RuntimeError: Expected all tensors to be on the same device, but found at least two devices, cuda:0 and cuda:1!"
319
+ },
320
+ {
321
+ "line": "tests/models/timm_wrapper/test_modeling_timm_wrapper.py::TimmWrapperModelTest::test_multi_gpu_data_parallel_forward",
322
+ "trace": "(line 769) StopIteration: Caught StopIteration in replica 1 on device 1."
323
+ }
324
+ ]
325
+ },
326
+ "job_link": {
327
+ "single": "https://github.com/huggingface/transformers/actions/runs/20422991545/job/58678187767",
328
+ "multi": "https://github.com/huggingface/transformers/actions/runs/20422991545/job/58678187765"
329
+ },
330
+ "captured_info": {}
331
+ }
332
+ }