hf-transformers-bot commited on
Commit
c770e22
·
verified ·
1 Parent(s): e55db8f

Upload 2026-01-12/runs/26599-20926567527/ci_results_run_models_gpu/model_results.json with huggingface_hub

Browse files
2026-01-12/runs/26599-20926567527/ci_results_run_models_gpu/model_results.json ADDED
@@ -0,0 +1,2031 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "models_afmoe": {
3
+ "failed": {
4
+ "PyTorch": {
5
+ "unclassified": 0,
6
+ "single": 2,
7
+ "multi": 2
8
+ },
9
+ "Tokenizers": {
10
+ "unclassified": 0,
11
+ "single": 0,
12
+ "multi": 0
13
+ },
14
+ "Pipelines": {
15
+ "unclassified": 0,
16
+ "single": 0,
17
+ "multi": 0
18
+ },
19
+ "Trainer": {
20
+ "unclassified": 0,
21
+ "single": 0,
22
+ "multi": 0
23
+ },
24
+ "ONNX": {
25
+ "unclassified": 0,
26
+ "single": 0,
27
+ "multi": 0
28
+ },
29
+ "Auto": {
30
+ "unclassified": 0,
31
+ "single": 0,
32
+ "multi": 0
33
+ },
34
+ "Quantization": {
35
+ "unclassified": 0,
36
+ "single": 0,
37
+ "multi": 0
38
+ },
39
+ "Unclassified": {
40
+ "unclassified": 0,
41
+ "single": 0,
42
+ "multi": 0
43
+ }
44
+ },
45
+ "errors": 0,
46
+ "success": 229,
47
+ "skipped": 229,
48
+ "time_spent": [
49
+ 116.57,
50
+ 119.07
51
+ ],
52
+ "error": false,
53
+ "failures": {
54
+ "single": [
55
+ {
56
+ "line": "tests/models/afmoe/test_modeling_afmoe.py::AfmoeModelTest::test_flex_attention_with_grads",
57
+ "trace": "(line 619) torch._inductor.exc.InductorError: RuntimeError: No valid triton configs. OutOfMemoryError: out of resource: triton_tem_fused_0 Required: 147968 Hardware limit:101376 Reducing block sizes or `num_stages` may help."
58
+ },
59
+ {
60
+ "line": "tests/models/afmoe/test_modeling_afmoe.py::AfmoeModelTest::test_sdpa_can_dispatch_on_flash",
61
+ "trace": "(line 92) RuntimeError: No available kernel. Aborting execution."
62
+ }
63
+ ],
64
+ "multi": [
65
+ {
66
+ "line": "tests/models/afmoe/test_modeling_afmoe.py::AfmoeModelTest::test_flex_attention_with_grads",
67
+ "trace": "(line 619) torch._inductor.exc.InductorError: RuntimeError: No valid triton configs. OutOfMemoryError: out of resource: triton_tem_fused_0 Required: 147968 Hardware limit:101376 Reducing block sizes or `num_stages` may help."
68
+ },
69
+ {
70
+ "line": "tests/models/afmoe/test_modeling_afmoe.py::AfmoeModelTest::test_sdpa_can_dispatch_on_flash",
71
+ "trace": "(line 92) RuntimeError: No available kernel. Aborting execution."
72
+ }
73
+ ]
74
+ },
75
+ "job_link": {
76
+ "single": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219037",
77
+ "multi": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219145"
78
+ },
79
+ "captured_info": {}
80
+ },
81
+ "models_aimv2": {
82
+ "failed": {
83
+ "PyTorch": {
84
+ "unclassified": 0,
85
+ "single": 2,
86
+ "multi": 2
87
+ },
88
+ "Tokenizers": {
89
+ "unclassified": 0,
90
+ "single": 0,
91
+ "multi": 0
92
+ },
93
+ "Pipelines": {
94
+ "unclassified": 0,
95
+ "single": 0,
96
+ "multi": 0
97
+ },
98
+ "Trainer": {
99
+ "unclassified": 0,
100
+ "single": 0,
101
+ "multi": 0
102
+ },
103
+ "ONNX": {
104
+ "unclassified": 0,
105
+ "single": 0,
106
+ "multi": 0
107
+ },
108
+ "Auto": {
109
+ "unclassified": 0,
110
+ "single": 0,
111
+ "multi": 0
112
+ },
113
+ "Quantization": {
114
+ "unclassified": 0,
115
+ "single": 0,
116
+ "multi": 0
117
+ },
118
+ "Unclassified": {
119
+ "unclassified": 0,
120
+ "single": 0,
121
+ "multi": 0
122
+ }
123
+ },
124
+ "errors": 0,
125
+ "success": 457,
126
+ "skipped": 359,
127
+ "time_spent": [
128
+ 108.21,
129
+ 107.29
130
+ ],
131
+ "error": false,
132
+ "failures": {
133
+ "multi": [
134
+ {
135
+ "line": "tests/models/aimv2/test_modeling_aimv2.py::Aimv2VisionModelTest::test_batching_equivalence",
136
+ "trace": "(line 1465) AssertionError: Batched and Single row outputs are not equal in Aimv2VisionModel for key=last_hidden_state."
137
+ },
138
+ {
139
+ "line": "tests/models/aimv2/test_modeling_aimv2.py::Aimv2ModelTest::test_batching_equivalence",
140
+ "trace": "(line 1465) AssertionError: Batched and Single row outputs are not equal in Aimv2Model for key=logits_per_image."
141
+ }
142
+ ],
143
+ "single": [
144
+ {
145
+ "line": "tests/models/aimv2/test_modeling_aimv2.py::Aimv2VisionModelTest::test_batching_equivalence",
146
+ "trace": "(line 1465) AssertionError: Batched and Single row outputs are not equal in Aimv2VisionModel for key=last_hidden_state."
147
+ },
148
+ {
149
+ "line": "tests/models/aimv2/test_modeling_aimv2.py::Aimv2ModelTest::test_batching_equivalence",
150
+ "trace": "(line 1465) AssertionError: Batched and Single row outputs are not equal in Aimv2Model for key=logits_per_image."
151
+ }
152
+ ]
153
+ },
154
+ "job_link": {
155
+ "multi": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219126",
156
+ "single": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219077"
157
+ },
158
+ "captured_info": {
159
+ "multi": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219126#step:16:1",
160
+ "single": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219077#step:16:1"
161
+ }
162
+ },
163
+ "models_albert": {
164
+ "failed": {
165
+ "PyTorch": {
166
+ "unclassified": 0,
167
+ "single": 0,
168
+ "multi": 0
169
+ },
170
+ "Tokenizers": {
171
+ "unclassified": 0,
172
+ "single": 0,
173
+ "multi": 0
174
+ },
175
+ "Pipelines": {
176
+ "unclassified": 0,
177
+ "single": 0,
178
+ "multi": 0
179
+ },
180
+ "Trainer": {
181
+ "unclassified": 0,
182
+ "single": 0,
183
+ "multi": 0
184
+ },
185
+ "ONNX": {
186
+ "unclassified": 0,
187
+ "single": 0,
188
+ "multi": 0
189
+ },
190
+ "Auto": {
191
+ "unclassified": 0,
192
+ "single": 0,
193
+ "multi": 0
194
+ },
195
+ "Quantization": {
196
+ "unclassified": 0,
197
+ "single": 0,
198
+ "multi": 0
199
+ },
200
+ "Unclassified": {
201
+ "unclassified": 0,
202
+ "single": 0,
203
+ "multi": 0
204
+ }
205
+ },
206
+ "errors": 0,
207
+ "success": 284,
208
+ "skipped": 186,
209
+ "time_spent": [
210
+ 146.16,
211
+ 145.67
212
+ ],
213
+ "error": false,
214
+ "failures": {},
215
+ "job_link": {
216
+ "multi": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219153",
217
+ "single": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219110"
218
+ },
219
+ "captured_info": {}
220
+ },
221
+ "models_align": {
222
+ "failed": {
223
+ "PyTorch": {
224
+ "unclassified": 0,
225
+ "single": 0,
226
+ "multi": 2
227
+ },
228
+ "Tokenizers": {
229
+ "unclassified": 0,
230
+ "single": 0,
231
+ "multi": 0
232
+ },
233
+ "Pipelines": {
234
+ "unclassified": 0,
235
+ "single": 0,
236
+ "multi": 0
237
+ },
238
+ "Trainer": {
239
+ "unclassified": 0,
240
+ "single": 0,
241
+ "multi": 0
242
+ },
243
+ "ONNX": {
244
+ "unclassified": 0,
245
+ "single": 0,
246
+ "multi": 0
247
+ },
248
+ "Auto": {
249
+ "unclassified": 0,
250
+ "single": 0,
251
+ "multi": 0
252
+ },
253
+ "Quantization": {
254
+ "unclassified": 0,
255
+ "single": 0,
256
+ "multi": 0
257
+ },
258
+ "Unclassified": {
259
+ "unclassified": 0,
260
+ "single": 0,
261
+ "multi": 0
262
+ }
263
+ },
264
+ "errors": 0,
265
+ "success": 323,
266
+ "skipped": 603,
267
+ "time_spent": [
268
+ 84.18,
269
+ 84.45
270
+ ],
271
+ "error": false,
272
+ "failures": {
273
+ "multi": [
274
+ {
275
+ "line": "tests/models/align/test_modeling_align.py::AlignTextModelTest::test_model_parallelism",
276
+ "trace": "(line 589) RuntimeError: Expected all tensors to be on the same device, but got mat2 is on cuda:1, different from other tensors on cuda:0 (when checking argument in method wrapper_CUDA_bmm)"
277
+ },
278
+ {
279
+ "line": "tests/models/align/test_modeling_align.py::AlignTextModelTest::test_multi_gpu_data_parallel_forward",
280
+ "trace": "(line 769) StopIteration: Caught StopIteration in replica 1 on device 1."
281
+ }
282
+ ]
283
+ },
284
+ "job_link": {
285
+ "single": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219044",
286
+ "multi": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219129"
287
+ },
288
+ "captured_info": {}
289
+ },
290
+ "models_altclip": {
291
+ "failed": {
292
+ "PyTorch": {
293
+ "unclassified": 0,
294
+ "single": 2,
295
+ "multi": 4
296
+ },
297
+ "Tokenizers": {
298
+ "unclassified": 0,
299
+ "single": 0,
300
+ "multi": 0
301
+ },
302
+ "Pipelines": {
303
+ "unclassified": 0,
304
+ "single": 0,
305
+ "multi": 0
306
+ },
307
+ "Trainer": {
308
+ "unclassified": 0,
309
+ "single": 0,
310
+ "multi": 0
311
+ },
312
+ "ONNX": {
313
+ "unclassified": 0,
314
+ "single": 0,
315
+ "multi": 0
316
+ },
317
+ "Auto": {
318
+ "unclassified": 0,
319
+ "single": 0,
320
+ "multi": 0
321
+ },
322
+ "Quantization": {
323
+ "unclassified": 0,
324
+ "single": 0,
325
+ "multi": 0
326
+ },
327
+ "Unclassified": {
328
+ "unclassified": 0,
329
+ "single": 0,
330
+ "multi": 0
331
+ }
332
+ },
333
+ "errors": 0,
334
+ "success": 322,
335
+ "skipped": 600,
336
+ "time_spent": [
337
+ 184.82,
338
+ 191.22
339
+ ],
340
+ "error": false,
341
+ "failures": {
342
+ "multi": [
343
+ {
344
+ "line": "tests/models/altclip/test_modeling_altclip.py::AltCLIPVisionModelTest::test_batching_equivalence",
345
+ "trace": "(line 1465) AssertionError: Batched and Single row outputs are not equal in AltCLIPVisionModel for key=last_hidden_state."
346
+ },
347
+ {
348
+ "line": "tests/models/altclip/test_modeling_altclip.py::AltCLIPTextModelTest::test_multi_gpu_data_parallel_forward",
349
+ "trace": "(line 769) StopIteration: Caught StopIteration in replica 1 on device 1."
350
+ },
351
+ {
352
+ "line": "tests/models/altclip/test_modeling_altclip.py::AltCLIPModelTest::test_batching_equivalence",
353
+ "trace": "(line 1465) AssertionError: Batched and Single row outputs are not equal in AltCLIPModel for key=logits_per_image."
354
+ },
355
+ {
356
+ "line": "tests/models/altclip/test_modeling_altclip.py::AltCLIPModelTest::test_multi_gpu_data_parallel_forward",
357
+ "trace": "(line 769) StopIteration: Caught StopIteration in replica 1 on device 1."
358
+ }
359
+ ],
360
+ "single": [
361
+ {
362
+ "line": "tests/models/altclip/test_modeling_altclip.py::AltCLIPVisionModelTest::test_batching_equivalence",
363
+ "trace": "(line 1465) AssertionError: Batched and Single row outputs are not equal in AltCLIPVisionModel for key=last_hidden_state."
364
+ },
365
+ {
366
+ "line": "tests/models/altclip/test_modeling_altclip.py::AltCLIPModelTest::test_batching_equivalence",
367
+ "trace": "(line 1465) AssertionError: Batched and Single row outputs are not equal in AltCLIPModel for key=vision_model_output."
368
+ }
369
+ ]
370
+ },
371
+ "job_link": {
372
+ "multi": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219149",
373
+ "single": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219036"
374
+ },
375
+ "captured_info": {
376
+ "multi": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219149#step:16:1",
377
+ "single": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219036#step:16:1"
378
+ }
379
+ },
380
+ "models_apertus": {
381
+ "failed": {
382
+ "PyTorch": {
383
+ "unclassified": 0,
384
+ "single": 0,
385
+ "multi": 0
386
+ },
387
+ "Tokenizers": {
388
+ "unclassified": 0,
389
+ "single": 0,
390
+ "multi": 0
391
+ },
392
+ "Pipelines": {
393
+ "unclassified": 0,
394
+ "single": 0,
395
+ "multi": 0
396
+ },
397
+ "Trainer": {
398
+ "unclassified": 0,
399
+ "single": 0,
400
+ "multi": 0
401
+ },
402
+ "ONNX": {
403
+ "unclassified": 0,
404
+ "single": 0,
405
+ "multi": 0
406
+ },
407
+ "Auto": {
408
+ "unclassified": 0,
409
+ "single": 0,
410
+ "multi": 0
411
+ },
412
+ "Quantization": {
413
+ "unclassified": 0,
414
+ "single": 0,
415
+ "multi": 0
416
+ },
417
+ "Unclassified": {
418
+ "unclassified": 0,
419
+ "single": 0,
420
+ "multi": 0
421
+ }
422
+ },
423
+ "errors": 0,
424
+ "success": 259,
425
+ "skipped": 203,
426
+ "time_spent": [
427
+ 99.65,
428
+ 100.57
429
+ ],
430
+ "error": false,
431
+ "failures": {},
432
+ "job_link": {
433
+ "single": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219069",
434
+ "multi": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219254"
435
+ },
436
+ "captured_info": {
437
+ "single": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219069#step:16:1",
438
+ "multi": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219254#step:16:1"
439
+ }
440
+ },
441
+ "models_arcee": {
442
+ "failed": {
443
+ "PyTorch": {
444
+ "unclassified": 0,
445
+ "single": 0,
446
+ "multi": 0
447
+ },
448
+ "Tokenizers": {
449
+ "unclassified": 0,
450
+ "single": 0,
451
+ "multi": 0
452
+ },
453
+ "Pipelines": {
454
+ "unclassified": 0,
455
+ "single": 0,
456
+ "multi": 0
457
+ },
458
+ "Trainer": {
459
+ "unclassified": 0,
460
+ "single": 0,
461
+ "multi": 0
462
+ },
463
+ "ONNX": {
464
+ "unclassified": 0,
465
+ "single": 0,
466
+ "multi": 0
467
+ },
468
+ "Auto": {
469
+ "unclassified": 0,
470
+ "single": 0,
471
+ "multi": 0
472
+ },
473
+ "Quantization": {
474
+ "unclassified": 0,
475
+ "single": 0,
476
+ "multi": 0
477
+ },
478
+ "Unclassified": {
479
+ "unclassified": 0,
480
+ "single": 0,
481
+ "multi": 0
482
+ }
483
+ },
484
+ "errors": 0,
485
+ "success": 271,
486
+ "skipped": 199,
487
+ "time_spent": [
488
+ 149.94,
489
+ 149.24
490
+ ],
491
+ "error": false,
492
+ "failures": {},
493
+ "job_link": {
494
+ "multi": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219171",
495
+ "single": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219065"
496
+ },
497
+ "captured_info": {
498
+ "multi": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219171#step:16:1",
499
+ "single": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219065#step:16:1"
500
+ }
501
+ },
502
+ "models_aria": {
503
+ "failed": {
504
+ "PyTorch": {
505
+ "unclassified": 0,
506
+ "single": 0,
507
+ "multi": 0
508
+ },
509
+ "Tokenizers": {
510
+ "unclassified": 0,
511
+ "single": 0,
512
+ "multi": 0
513
+ },
514
+ "Pipelines": {
515
+ "unclassified": 0,
516
+ "single": 0,
517
+ "multi": 0
518
+ },
519
+ "Trainer": {
520
+ "unclassified": 0,
521
+ "single": 0,
522
+ "multi": 0
523
+ },
524
+ "ONNX": {
525
+ "unclassified": 0,
526
+ "single": 0,
527
+ "multi": 0
528
+ },
529
+ "Auto": {
530
+ "unclassified": 0,
531
+ "single": 0,
532
+ "multi": 0
533
+ },
534
+ "Quantization": {
535
+ "unclassified": 0,
536
+ "single": 0,
537
+ "multi": 0
538
+ },
539
+ "Unclassified": {
540
+ "unclassified": 0,
541
+ "single": 0,
542
+ "multi": 0
543
+ }
544
+ },
545
+ "errors": 0,
546
+ "success": 317,
547
+ "skipped": 185,
548
+ "time_spent": [
549
+ 193.41,
550
+ 194.7
551
+ ],
552
+ "error": false,
553
+ "failures": {},
554
+ "job_link": {
555
+ "single": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219094",
556
+ "multi": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219161"
557
+ },
558
+ "captured_info": {}
559
+ },
560
+ "models_audioflamingo3": {
561
+ "failed": {
562
+ "PyTorch": {
563
+ "unclassified": 0,
564
+ "single": 0,
565
+ "multi": 0
566
+ },
567
+ "Tokenizers": {
568
+ "unclassified": 0,
569
+ "single": 0,
570
+ "multi": 0
571
+ },
572
+ "Pipelines": {
573
+ "unclassified": 0,
574
+ "single": 0,
575
+ "multi": 0
576
+ },
577
+ "Trainer": {
578
+ "unclassified": 0,
579
+ "single": 0,
580
+ "multi": 0
581
+ },
582
+ "ONNX": {
583
+ "unclassified": 0,
584
+ "single": 0,
585
+ "multi": 0
586
+ },
587
+ "Auto": {
588
+ "unclassified": 0,
589
+ "single": 0,
590
+ "multi": 0
591
+ },
592
+ "Quantization": {
593
+ "unclassified": 0,
594
+ "single": 0,
595
+ "multi": 0
596
+ },
597
+ "Unclassified": {
598
+ "unclassified": 0,
599
+ "single": 0,
600
+ "multi": 0
601
+ }
602
+ },
603
+ "errors": 0,
604
+ "success": 271,
605
+ "skipped": 167,
606
+ "time_spent": [
607
+ 145.55,
608
+ 148.15
609
+ ],
610
+ "error": false,
611
+ "failures": {},
612
+ "job_link": {
613
+ "multi": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219220",
614
+ "single": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219089"
615
+ },
616
+ "captured_info": {}
617
+ },
618
+ "models_auto": {
619
+ "failed": {
620
+ "PyTorch": {
621
+ "unclassified": 0,
622
+ "single": 0,
623
+ "multi": 0
624
+ },
625
+ "Tokenizers": {
626
+ "unclassified": 0,
627
+ "single": 3,
628
+ "multi": 3
629
+ },
630
+ "Pipelines": {
631
+ "unclassified": 0,
632
+ "single": 0,
633
+ "multi": 0
634
+ },
635
+ "Trainer": {
636
+ "unclassified": 0,
637
+ "single": 0,
638
+ "multi": 0
639
+ },
640
+ "ONNX": {
641
+ "unclassified": 0,
642
+ "single": 0,
643
+ "multi": 0
644
+ },
645
+ "Auto": {
646
+ "unclassified": 0,
647
+ "single": 0,
648
+ "multi": 0
649
+ },
650
+ "Quantization": {
651
+ "unclassified": 0,
652
+ "single": 0,
653
+ "multi": 0
654
+ },
655
+ "Unclassified": {
656
+ "unclassified": 0,
657
+ "single": 0,
658
+ "multi": 0
659
+ }
660
+ },
661
+ "errors": 0,
662
+ "success": 232,
663
+ "skipped": 14,
664
+ "time_spent": [
665
+ 158.04,
666
+ 162.58
667
+ ],
668
+ "error": false,
669
+ "failures": {
670
+ "multi": [
671
+ {
672
+ "line": "tests/models/auto/test_tokenization_auto.py::AutoTokenizerTest::test_custom_tokenizer_from_hub",
673
+ "trace": "(line 687) AssertionError: False is not true"
674
+ },
675
+ {
676
+ "line": "tests/models/auto/test_tokenization_auto.py::AutoTokenizerTest::test_from_pretrained_dynamic_tokenizer",
677
+ "trace": "(line 163) AssertionError: ValueError not raised"
678
+ },
679
+ {
680
+ "line": "tests/models/auto/test_tokenization_auto.py::AutoTokenizerTest::test_tokenizer_identifier_non_existent",
681
+ "trace": "(line 163) AssertionError: \"julien-c/herlolip-not-exists is not a local folder and is not a valid model identifier\" does not match \"401 Client Error. (Request ID: Root=1-696521a6-373f391e5627ffae35f6db7a;41c2ad18-b15c-4904-b9d3-81092c46ef68)"
682
+ }
683
+ ],
684
+ "single": [
685
+ {
686
+ "line": "tests/models/auto/test_tokenization_auto.py::AutoTokenizerTest::test_custom_tokenizer_from_hub",
687
+ "trace": "(line 687) AssertionError: False is not true"
688
+ },
689
+ {
690
+ "line": "tests/models/auto/test_tokenization_auto.py::AutoTokenizerTest::test_from_pretrained_dynamic_tokenizer",
691
+ "trace": "(line 163) AssertionError: ValueError not raised"
692
+ },
693
+ {
694
+ "line": "tests/models/auto/test_tokenization_auto.py::AutoTokenizerTest::test_tokenizer_identifier_non_existent",
695
+ "trace": "(line 163) AssertionError: \"julien-c/herlolip-not-exists is not a local folder and is not a valid model identifier\" does not match \"401 Client Error. (Request ID: Root=1-69652113-2e76514231dddc3b27bf5e70;dcdaa291-2fbc-4259-9c05-6cefe65590cc)"
696
+ }
697
+ ]
698
+ },
699
+ "job_link": {
700
+ "multi": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219207",
701
+ "single": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219117"
702
+ },
703
+ "captured_info": {}
704
+ },
705
+ "models_bamba": {
706
+ "failed": {
707
+ "PyTorch": {
708
+ "unclassified": 0,
709
+ "single": 3,
710
+ "multi": 3
711
+ },
712
+ "Tokenizers": {
713
+ "unclassified": 0,
714
+ "single": 0,
715
+ "multi": 0
716
+ },
717
+ "Pipelines": {
718
+ "unclassified": 0,
719
+ "single": 0,
720
+ "multi": 0
721
+ },
722
+ "Trainer": {
723
+ "unclassified": 0,
724
+ "single": 0,
725
+ "multi": 0
726
+ },
727
+ "ONNX": {
728
+ "unclassified": 0,
729
+ "single": 0,
730
+ "multi": 0
731
+ },
732
+ "Auto": {
733
+ "unclassified": 0,
734
+ "single": 0,
735
+ "multi": 0
736
+ },
737
+ "Quantization": {
738
+ "unclassified": 0,
739
+ "single": 0,
740
+ "multi": 0
741
+ },
742
+ "Unclassified": {
743
+ "unclassified": 0,
744
+ "single": 0,
745
+ "multi": 0
746
+ }
747
+ },
748
+ "errors": 0,
749
+ "success": 213,
750
+ "skipped": 229,
751
+ "time_spent": [
752
+ 78.57,
753
+ 78.8
754
+ ],
755
+ "error": false,
756
+ "failures": {
757
+ "single": [
758
+ {
759
+ "line": "tests/models/bamba/test_modeling_bamba.py::BambaModelTest::test_sdpa_can_compile_dynamic",
760
+ "trace": "(line 1677) torch._dynamo.exc.TorchRuntimeError: Dynamo failed to run FX node with fake tensors: call_function <built-in function scaled_dot_product_attention>(*(FakeTensor(..., device='cuda:0', size=(s47, 4, s87, 8), dtype=torch.bfloat16), FakeTensor(..., device='cuda:0', size=(s47, 4, s87, 8), dtype=torch.bfloat16), FakeTensor(..., device='cuda:0', size=(s47, 4, s87, 8), dtype=torch.bfloat16)), **{'attn_mask': FakeTensor(..., device='cuda:0', size=(s84, 1, s34, s30), dtype=torch.bfloat16), 'dropout_p': 0.0, 'scale': 0.353553390593274, 'is_causal': False}): got RuntimeError('expand: attempting to expand a dimension of length s30 -> s87!')"
761
+ },
762
+ {
763
+ "line": "tests/models/bamba/test_modeling_bamba.py::BambaModelIntegrationTest::test_simple_batched_generate_with_padding",
764
+ "trace": "(line 901) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 8.00 GiB. GPU 0 has a total capacity of 22.30 GiB of which 3.40 GiB is free. Process 17404 has 18.90 GiB memory in use. Of the allocated memory 18.54 GiB is allocated by PyTorch, and 4.01 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)"
765
+ },
766
+ {
767
+ "line": "tests/models/bamba/test_modeling_bamba.py::BambaModelIntegrationTest::test_simple_generate",
768
+ "trace": "(line 901) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 4.00 GiB. GPU 0 has a total capacity of 22.30 GiB of which 3.25 GiB is free. Process 17404 has 19.04 GiB memory in use. Of the allocated memory 18.67 GiB is allocated by PyTorch, and 11.38 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)"
769
+ }
770
+ ],
771
+ "multi": [
772
+ {
773
+ "line": "tests/models/bamba/test_modeling_bamba.py::BambaModelTest::test_sdpa_can_compile_dynamic",
774
+ "trace": "(line 1677) torch._dynamo.exc.TorchRuntimeError: Dynamo failed to run FX node with fake tensors: call_function <built-in function scaled_dot_product_attention>(*(FakeTensor(..., device='cuda:0', size=(s47, 4, s87, 8), dtype=torch.bfloat16), FakeTensor(..., device='cuda:0', size=(s47, 4, s87, 8), dtype=torch.bfloat16), FakeTensor(..., device='cuda:0', size=(s47, 4, s87, 8), dtype=torch.bfloat16)), **{'attn_mask': FakeTensor(..., device='cuda:0', size=(s84, 1, s34, s30), dtype=torch.bfloat16), 'dropout_p': 0.0, 'scale': 0.353553390593274, 'is_causal': False}): got RuntimeError('expand: attempting to expand a dimension of length s30 -> s87!')"
775
+ },
776
+ {
777
+ "line": "tests/models/bamba/test_modeling_bamba.py::BambaModelIntegrationTest::test_simple_batched_generate_with_padding",
778
+ "trace": "(line 901) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 8.00 GiB. GPU 0 has a total capacity of 22.30 GiB of which 3.27 GiB is free. Process 31362 has 19.03 GiB memory in use. Of the allocated memory 18.55 GiB is allocated by PyTorch, and 13.88 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)"
779
+ },
780
+ {
781
+ "line": "tests/models/bamba/test_modeling_bamba.py::BambaModelIntegrationTest::test_simple_generate",
782
+ "trace": "(line 901) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 4.00 GiB. GPU 0 has a total capacity of 22.30 GiB of which 3.12 GiB is free. Process 31362 has 19.17 GiB memory in use. Of the allocated memory 18.68 GiB is allocated by PyTorch, and 19.25 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)"
783
+ }
784
+ ]
785
+ },
786
+ "job_link": {
787
+ "single": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219074",
788
+ "multi": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219148"
789
+ },
790
+ "captured_info": {}
791
+ },
792
+ "models_bark": {
793
+ "failed": {
794
+ "PyTorch": {
795
+ "unclassified": 0,
796
+ "single": 1,
797
+ "multi": 1
798
+ },
799
+ "Tokenizers": {
800
+ "unclassified": 0,
801
+ "single": 0,
802
+ "multi": 0
803
+ },
804
+ "Pipelines": {
805
+ "unclassified": 0,
806
+ "single": 0,
807
+ "multi": 0
808
+ },
809
+ "Trainer": {
810
+ "unclassified": 0,
811
+ "single": 0,
812
+ "multi": 0
813
+ },
814
+ "ONNX": {
815
+ "unclassified": 0,
816
+ "single": 0,
817
+ "multi": 0
818
+ },
819
+ "Auto": {
820
+ "unclassified": 0,
821
+ "single": 0,
822
+ "multi": 0
823
+ },
824
+ "Quantization": {
825
+ "unclassified": 0,
826
+ "single": 0,
827
+ "multi": 0
828
+ },
829
+ "Unclassified": {
830
+ "unclassified": 0,
831
+ "single": 0,
832
+ "multi": 0
833
+ }
834
+ },
835
+ "errors": 0,
836
+ "success": 416,
837
+ "skipped": 498,
838
+ "time_spent": [
839
+ 739.79,
840
+ 731.14
841
+ ],
842
+ "error": false,
843
+ "failures": {
844
+ "single": [
845
+ {
846
+ "line": "tests/models/bark/test_modeling_bark.py::BarkModelIntegrationTests::test_generate_end_to_end_with_offload",
847
+ "trace": "(line 2542) RuntimeError: Expected all tensors to be on the same device, but got index is on cuda:0, different from other tensors on cpu (when checking argument in method wrapper_CUDA__index_select)"
848
+ }
849
+ ],
850
+ "multi": [
851
+ {
852
+ "line": "tests/models/bark/test_modeling_bark.py::BarkModelIntegrationTests::test_generate_end_to_end_with_offload",
853
+ "trace": "(line 2542) RuntimeError: Expected all tensors to be on the same device, but got index is on cuda:0, different from other tensors on cpu (when checking argument in method wrapper_CUDA__index_select)"
854
+ }
855
+ ]
856
+ },
857
+ "job_link": {
858
+ "single": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219079",
859
+ "multi": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219172"
860
+ },
861
+ "captured_info": {}
862
+ },
863
+ "models_bart": {
864
+ "failed": {
865
+ "PyTorch": {
866
+ "unclassified": 0,
867
+ "single": 0,
868
+ "multi": 0
869
+ },
870
+ "Tokenizers": {
871
+ "unclassified": 0,
872
+ "single": 0,
873
+ "multi": 0
874
+ },
875
+ "Pipelines": {
876
+ "unclassified": 0,
877
+ "single": 0,
878
+ "multi": 0
879
+ },
880
+ "Trainer": {
881
+ "unclassified": 0,
882
+ "single": 0,
883
+ "multi": 0
884
+ },
885
+ "ONNX": {
886
+ "unclassified": 0,
887
+ "single": 0,
888
+ "multi": 0
889
+ },
890
+ "Auto": {
891
+ "unclassified": 0,
892
+ "single": 0,
893
+ "multi": 0
894
+ },
895
+ "Quantization": {
896
+ "unclassified": 0,
897
+ "single": 0,
898
+ "multi": 0
899
+ },
900
+ "Unclassified": {
901
+ "unclassified": 0,
902
+ "single": 0,
903
+ "multi": 0
904
+ }
905
+ },
906
+ "errors": 0,
907
+ "success": 490,
908
+ "skipped": 326,
909
+ "time_spent": [
910
+ 218.98,
911
+ 219.66
912
+ ],
913
+ "error": false,
914
+ "failures": {},
915
+ "job_link": {
916
+ "multi": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219405",
917
+ "single": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219146"
918
+ },
919
+ "captured_info": {}
920
+ },
921
+ "models_bert": {
922
+ "failed": {
923
+ "PyTorch": {
924
+ "unclassified": 0,
925
+ "single": 0,
926
+ "multi": 0
927
+ },
928
+ "Tokenizers": {
929
+ "unclassified": 0,
930
+ "single": 0,
931
+ "multi": 0
932
+ },
933
+ "Pipelines": {
934
+ "unclassified": 0,
935
+ "single": 0,
936
+ "multi": 0
937
+ },
938
+ "Trainer": {
939
+ "unclassified": 0,
940
+ "single": 0,
941
+ "multi": 0
942
+ },
943
+ "ONNX": {
944
+ "unclassified": 0,
945
+ "single": 0,
946
+ "multi": 0
947
+ },
948
+ "Auto": {
949
+ "unclassified": 0,
950
+ "single": 0,
951
+ "multi": 0
952
+ },
953
+ "Quantization": {
954
+ "unclassified": 0,
955
+ "single": 0,
956
+ "multi": 0
957
+ },
958
+ "Unclassified": {
959
+ "unclassified": 0,
960
+ "single": 0,
961
+ "multi": 0
962
+ }
963
+ },
964
+ "errors": 0,
965
+ "success": 357,
966
+ "skipped": 217,
967
+ "time_spent": [
968
+ 204.45,
969
+ 205.14
970
+ ],
971
+ "error": false,
972
+ "failures": {},
973
+ "job_link": {
974
+ "multi": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219130",
975
+ "single": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219073"
976
+ },
977
+ "captured_info": {}
978
+ },
979
+ "models_bert_generation": {
980
+ "failed": {
981
+ "PyTorch": {
982
+ "unclassified": 0,
983
+ "single": 0,
984
+ "multi": 0
985
+ },
986
+ "Tokenizers": {
987
+ "unclassified": 0,
988
+ "single": 0,
989
+ "multi": 0
990
+ },
991
+ "Pipelines": {
992
+ "unclassified": 0,
993
+ "single": 0,
994
+ "multi": 0
995
+ },
996
+ "Trainer": {
997
+ "unclassified": 0,
998
+ "single": 0,
999
+ "multi": 0
1000
+ },
1001
+ "ONNX": {
1002
+ "unclassified": 0,
1003
+ "single": 0,
1004
+ "multi": 0
1005
+ },
1006
+ "Auto": {
1007
+ "unclassified": 0,
1008
+ "single": 0,
1009
+ "multi": 0
1010
+ },
1011
+ "Quantization": {
1012
+ "unclassified": 0,
1013
+ "single": 0,
1014
+ "multi": 0
1015
+ },
1016
+ "Unclassified": {
1017
+ "unclassified": 0,
1018
+ "single": 0,
1019
+ "multi": 0
1020
+ }
1021
+ },
1022
+ "errors": 0,
1023
+ "success": 349,
1024
+ "skipped": 225,
1025
+ "time_spent": [
1026
+ 82.08,
1027
+ 78.85
1028
+ ],
1029
+ "error": false,
1030
+ "failures": {},
1031
+ "job_link": {
1032
+ "multi": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219075",
1033
+ "single": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219137"
1034
+ },
1035
+ "captured_info": {}
1036
+ },
1037
+ "models_big_bird": {
1038
+ "failed": {
1039
+ "PyTorch": {
1040
+ "unclassified": 0,
1041
+ "single": 1,
1042
+ "multi": 1
1043
+ },
1044
+ "Tokenizers": {
1045
+ "unclassified": 0,
1046
+ "single": 0,
1047
+ "multi": 0
1048
+ },
1049
+ "Pipelines": {
1050
+ "unclassified": 0,
1051
+ "single": 0,
1052
+ "multi": 0
1053
+ },
1054
+ "Trainer": {
1055
+ "unclassified": 0,
1056
+ "single": 0,
1057
+ "multi": 0
1058
+ },
1059
+ "ONNX": {
1060
+ "unclassified": 0,
1061
+ "single": 0,
1062
+ "multi": 0
1063
+ },
1064
+ "Auto": {
1065
+ "unclassified": 0,
1066
+ "single": 0,
1067
+ "multi": 0
1068
+ },
1069
+ "Quantization": {
1070
+ "unclassified": 0,
1071
+ "single": 0,
1072
+ "multi": 0
1073
+ },
1074
+ "Unclassified": {
1075
+ "unclassified": 0,
1076
+ "single": 0,
1077
+ "multi": 0
1078
+ }
1079
+ },
1080
+ "errors": 0,
1081
+ "success": 244,
1082
+ "skipped": 252,
1083
+ "time_spent": [
1084
+ 83.18,
1085
+ 81.43
1086
+ ],
1087
+ "error": false,
1088
+ "failures": {
1089
+ "single": [
1090
+ {
1091
+ "line": "tests/models/big_bird/test_modeling_big_bird.py::BigBirdModelIntegrationTest::test_fill_mask",
1092
+ "trace": "(line 911) AssertionError: '' != 'happiness'"
1093
+ }
1094
+ ],
1095
+ "multi": [
1096
+ {
1097
+ "line": "tests/models/big_bird/test_modeling_big_bird.py::BigBirdModelIntegrationTest::test_fill_mask",
1098
+ "trace": "(line 911) AssertionError: '' != 'happiness'"
1099
+ }
1100
+ ]
1101
+ },
1102
+ "job_link": {
1103
+ "single": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219225",
1104
+ "multi": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219102"
1105
+ },
1106
+ "captured_info": {
1107
+ "single": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219225#step:16:1",
1108
+ "multi": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219102#step:16:1"
1109
+ }
1110
+ },
1111
+ "models_gemma": {
1112
+ "failed": {
1113
+ "PyTorch": {
1114
+ "unclassified": 0,
1115
+ "single": 3,
1116
+ "multi": 3
1117
+ },
1118
+ "Tokenizers": {
1119
+ "unclassified": 0,
1120
+ "single": 0,
1121
+ "multi": 0
1122
+ },
1123
+ "Pipelines": {
1124
+ "unclassified": 0,
1125
+ "single": 0,
1126
+ "multi": 0
1127
+ },
1128
+ "Trainer": {
1129
+ "unclassified": 0,
1130
+ "single": 0,
1131
+ "multi": 0
1132
+ },
1133
+ "ONNX": {
1134
+ "unclassified": 0,
1135
+ "single": 0,
1136
+ "multi": 0
1137
+ },
1138
+ "Auto": {
1139
+ "unclassified": 0,
1140
+ "single": 0,
1141
+ "multi": 0
1142
+ },
1143
+ "Quantization": {
1144
+ "unclassified": 0,
1145
+ "single": 0,
1146
+ "multi": 0
1147
+ },
1148
+ "Unclassified": {
1149
+ "unclassified": 0,
1150
+ "single": 0,
1151
+ "multi": 0
1152
+ }
1153
+ },
1154
+ "errors": 0,
1155
+ "success": 385,
1156
+ "skipped": 203,
1157
+ "time_spent": [
1158
+ 945.77,
1159
+ 944.87
1160
+ ],
1161
+ "error": false,
1162
+ "failures": {
1163
+ "multi": [
1164
+ {
1165
+ "line": "tests/models/gemma/test_modeling_gemma.py::GemmaIntegrationTest::test_export_static_cache",
1166
+ "trace": "(line 436) AssertionError: Lists differ: ['Hel[87 chars] in the 1990s. I have been looking on the internet and I have'] != ['Hel[87 chars] in the 1990s. I have looked on the internet and I have found']"
1167
+ },
1168
+ {
1169
+ "line": "tests/models/gemma/test_modeling_gemma.py::GemmaIntegrationTest::test_model_7b_4bit",
1170
+ "trace": "(line 329) AssertionError: Lists differ: ['Hel[42 chars]I am using a pic16f877a. I', 'Hi today I am go[68 chars]est'] != ['Hel[42 chars]I am trying to make a program that will take a[110 chars]!:)']"
1171
+ },
1172
+ {
1173
+ "line": "tests/models/gemma/test_modeling_gemma.py::GemmaIntegrationTest::test_model_7b_fp16_static_cache",
1174
+ "trace": "(line 299) AssertionError: Lists differ: ['Hel[27 chars]a 1999 4.0L 4x4. I', 'Hi today I am going to s[51 chars] 3D'] != ['Hel[27 chars]a 1995 3000gt SL. I have a', 'Hi today I am go[59 chars] 3D']"
1175
+ }
1176
+ ],
1177
+ "single": [
1178
+ {
1179
+ "line": "tests/models/gemma/test_modeling_gemma.py::GemmaIntegrationTest::test_export_static_cache",
1180
+ "trace": "(line 436) AssertionError: Lists differ: ['Hel[87 chars] in the 1990s. I have been looking on the internet and I have'] != ['Hel[87 chars] in the 1990s. I have looked on the internet and I have found']"
1181
+ },
1182
+ {
1183
+ "line": "tests/models/gemma/test_modeling_gemma.py::GemmaIntegrationTest::test_model_7b_4bit",
1184
+ "trace": "(line 329) AssertionError: Lists differ: ['Hel[42 chars]I am using a pic16f877a. I', 'Hi today I am go[68 chars]est'] != ['Hel[42 chars]I am trying to make a program that will take a[110 chars]!:)']"
1185
+ },
1186
+ {
1187
+ "line": "tests/models/gemma/test_modeling_gemma.py::GemmaIntegrationTest::test_model_7b_fp16_static_cache",
1188
+ "trace": "(line 299) AssertionError: Lists differ: ['Hel[27 chars]a 1999 4.0L 4x4. I', 'Hi today I am going to s[51 chars] 3D'] != ['Hel[27 chars]a 1995 3000gt SL. I have a', 'Hi today I am go[59 chars] 3D']"
1189
+ }
1190
+ ]
1191
+ },
1192
+ "job_link": {
1193
+ "multi": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219106",
1194
+ "single": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219159"
1195
+ },
1196
+ "captured_info": {
1197
+ "multi": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219106#step:16:1",
1198
+ "single": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219159#step:16:1"
1199
+ }
1200
+ },
1201
+ "models_gemma2": {
1202
+ "failed": {
1203
+ "PyTorch": {
1204
+ "unclassified": 0,
1205
+ "single": 1,
1206
+ "multi": 1
1207
+ },
1208
+ "Tokenizers": {
1209
+ "unclassified": 0,
1210
+ "single": 0,
1211
+ "multi": 0
1212
+ },
1213
+ "Pipelines": {
1214
+ "unclassified": 0,
1215
+ "single": 0,
1216
+ "multi": 0
1217
+ },
1218
+ "Trainer": {
1219
+ "unclassified": 0,
1220
+ "single": 0,
1221
+ "multi": 0
1222
+ },
1223
+ "ONNX": {
1224
+ "unclassified": 0,
1225
+ "single": 0,
1226
+ "multi": 0
1227
+ },
1228
+ "Auto": {
1229
+ "unclassified": 0,
1230
+ "single": 0,
1231
+ "multi": 0
1232
+ },
1233
+ "Quantization": {
1234
+ "unclassified": 0,
1235
+ "single": 0,
1236
+ "multi": 0
1237
+ },
1238
+ "Unclassified": {
1239
+ "unclassified": 0,
1240
+ "single": 1,
1241
+ "multi": 1
1242
+ }
1243
+ },
1244
+ "errors": 0,
1245
+ "success": 280,
1246
+ "skipped": 212,
1247
+ "time_spent": [
1248
+ 522.52,
1249
+ 634.89
1250
+ ],
1251
+ "error": false,
1252
+ "failures": {
1253
+ "multi": [
1254
+ {
1255
+ "line": "[100%]",
1256
+ "trace": "(line 2779) Failed: (subprocess) torch.AcceleratorError: CUDA error: an illegal memory access was encountered"
1257
+ },
1258
+ {
1259
+ "line": "tests/models/gemma2/test_modeling_gemma2.py::Gemma2IntegrationTest::test_model_2b_pipeline_bf16_flex_attention",
1260
+ "trace": "Cannot retrieve error message."
1261
+ }
1262
+ ],
1263
+ "single": [
1264
+ {
1265
+ "line": "[100%]",
1266
+ "trace": "(line 2779) Failed: (subprocess) torch.AcceleratorError: CUDA error: an illegal memory access was encountered"
1267
+ },
1268
+ {
1269
+ "line": "tests/models/gemma2/test_modeling_gemma2.py::Gemma2IntegrationTest::test_model_2b_pipeline_bf16_flex_attention",
1270
+ "trace": "Cannot retrieve error message."
1271
+ }
1272
+ ]
1273
+ },
1274
+ "job_link": {
1275
+ "multi": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219166",
1276
+ "single": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219180"
1277
+ },
1278
+ "captured_info": {}
1279
+ },
1280
+ "models_gemma3": {
1281
+ "failed": {
1282
+ "PyTorch": {
1283
+ "unclassified": 0,
1284
+ "single": 6,
1285
+ "multi": 6
1286
+ },
1287
+ "Tokenizers": {
1288
+ "unclassified": 0,
1289
+ "single": 0,
1290
+ "multi": 0
1291
+ },
1292
+ "Pipelines": {
1293
+ "unclassified": 0,
1294
+ "single": 0,
1295
+ "multi": 0
1296
+ },
1297
+ "Trainer": {
1298
+ "unclassified": 0,
1299
+ "single": 0,
1300
+ "multi": 0
1301
+ },
1302
+ "ONNX": {
1303
+ "unclassified": 0,
1304
+ "single": 0,
1305
+ "multi": 0
1306
+ },
1307
+ "Auto": {
1308
+ "unclassified": 0,
1309
+ "single": 0,
1310
+ "multi": 0
1311
+ },
1312
+ "Quantization": {
1313
+ "unclassified": 0,
1314
+ "single": 0,
1315
+ "multi": 0
1316
+ },
1317
+ "Unclassified": {
1318
+ "unclassified": 0,
1319
+ "single": 0,
1320
+ "multi": 0
1321
+ }
1322
+ },
1323
+ "errors": 0,
1324
+ "success": 586,
1325
+ "skipped": 372,
1326
+ "time_spent": [
1327
+ 557.93,
1328
+ 553.8
1329
+ ],
1330
+ "error": false,
1331
+ "failures": {
1332
+ "single": [
1333
+ {
1334
+ "line": "tests/models/gemma3/test_modeling_gemma3.py::Gemma3TextModelTest::test_generation_beyond_sliding_window_tiny_model",
1335
+ "trace": "(line 151) AssertionError: Tensor-likes are not equal!"
1336
+ },
1337
+ {
1338
+ "line": "tests/models/gemma3/test_modeling_gemma3.py::Gemma3IntegrationTest::test_dynamic_sliding_window_is_default",
1339
+ "trace": "(line 677) AssertionError: 'DynamicSlidingWindowLayer' unexpectedly found in 'DynamicCache(layers=[DynamicSlidingWindowLayer, DynamicSlidingWindowLayer, DynamicSlidingWindowLayer, DynamicSlidingWindowLayer, DynamicSlidingWindowLayer, DynamicLayer, DynamicSlidingWindowLayer, DynamicSlidingWindowLayer, DynamicSlidingWindowLayer, DynamicSlidingWindowLayer, DynamicSlidingWindowLayer, DynamicLayer, DynamicSlidingWindowLayer, DynamicSlidingWindowLayer, DynamicSlidingWindowLayer, DynamicSlidingWindowLayer, DynamicSlidingWindowLayer, DynamicLayer, DynamicSlidingWindowLayer, DynamicSlidingWindowLayer, DynamicSlidingWindowLayer, DynamicSlidingWindowLayer, DynamicSlidingWindowLayer, DynamicLayer, DynamicSlidingWindowLayer, DynamicSlidingWindowLayer])'"
1340
+ },
1341
+ {
1342
+ "line": "tests/models/gemma3/test_modeling_gemma3.py::Gemma3IntegrationTest::test_export_text_only",
1343
+ "trace": "(line 677) AssertionError: 'What[21 chars]ance? editarviles 콘 കേന്ദ്ര続くétricasétricas em[88 chars]िमों' != 'What[21 chars]ance?ępuje തരennelimegreen materiale トレーétrica[74 chars]rens'"
1344
+ },
1345
+ {
1346
+ "line": "tests/models/gemma3/test_modeling_gemma3.py::Gemma3IntegrationTest::test_generation_beyond_sliding_window_1_sdpa",
1347
+ "trace": "(line 3400) UnboundLocalError: local variable 'tb' referenced before assignment"
1348
+ },
1349
+ {
1350
+ "line": "tests/models/gemma3/test_modeling_gemma3.py::Gemma3IntegrationTest::test_generation_beyond_sliding_window_2_eager",
1351
+ "trace": "(line 3400) UnboundLocalError: local variable 'tb' referenced before assignment"
1352
+ },
1353
+ {
1354
+ "line": "tests/models/gemma3/test_modeling_gemma3.py::Gemma3IntegrationTest::test_model_1b_text_only",
1355
+ "trace": "(line 677) AssertionError: Lists differ: [\"Write a poem about Machine Learning. sub[159 chars]ctr\"] != ['Write a poem about Machine Learning.\\n\\n[108 chars],\\n']"
1356
+ }
1357
+ ],
1358
+ "multi": [
1359
+ {
1360
+ "line": "tests/models/gemma3/test_modeling_gemma3.py::Gemma3TextModelTest::test_generation_beyond_sliding_window_tiny_model",
1361
+ "trace": "(line 151) AssertionError: Tensor-likes are not equal!"
1362
+ },
1363
+ {
1364
+ "line": "tests/models/gemma3/test_modeling_gemma3.py::Gemma3IntegrationTest::test_dynamic_sliding_window_is_default",
1365
+ "trace": "(line 677) AssertionError: 'DynamicSlidingWindowLayer' unexpectedly found in 'DynamicCache(layers=[DynamicSlidingWindowLayer, DynamicSlidingWindowLayer, DynamicSlidingWindowLayer, DynamicSlidingWindowLayer, DynamicSlidingWindowLayer, DynamicLayer, DynamicSlidingWindowLayer, DynamicSlidingWindowLayer, DynamicSlidingWindowLayer, DynamicSlidingWindowLayer, DynamicSlidingWindowLayer, DynamicLayer, DynamicSlidingWindowLayer, DynamicSlidingWindowLayer, DynamicSlidingWindowLayer, DynamicSlidingWindowLayer, DynamicSlidingWindowLayer, DynamicLayer, DynamicSlidingWindowLayer, DynamicSlidingWindowLayer, DynamicSlidingWindowLayer, DynamicSlidingWindowLayer, DynamicSlidingWindowLayer, DynamicLayer, DynamicSlidingWindowLayer, DynamicSlidingWindowLayer])'"
1366
+ },
1367
+ {
1368
+ "line": "tests/models/gemma3/test_modeling_gemma3.py::Gemma3IntegrationTest::test_export_text_only",
1369
+ "trace": "(line 677) AssertionError: 'What[21 chars]ance? editarviles 콘 കേന്ദ്ര続くétricasétricas em[88 chars]िमों' != 'What[21 chars]ance?ępuje തരennelimegreen materiale トレーétrica[74 chars]rens'"
1370
+ },
1371
+ {
1372
+ "line": "tests/models/gemma3/test_modeling_gemma3.py::Gemma3IntegrationTest::test_generation_beyond_sliding_window_1_sdpa",
1373
+ "trace": "(line 3400) UnboundLocalError: local variable 'tb' referenced before assignment"
1374
+ },
1375
+ {
1376
+ "line": "tests/models/gemma3/test_modeling_gemma3.py::Gemma3IntegrationTest::test_generation_beyond_sliding_window_2_eager",
1377
+ "trace": "(line 3400) UnboundLocalError: local variable 'tb' referenced before assignment"
1378
+ },
1379
+ {
1380
+ "line": "tests/models/gemma3/test_modeling_gemma3.py::Gemma3IntegrationTest::test_model_1b_text_only",
1381
+ "trace": "(line 677) AssertionError: Lists differ: [\"Write a poem about Machine Learning. sub[159 chars]ctr\"] != ['Write a poem about Machine Learning.\\n\\n[108 chars],\\n']"
1382
+ }
1383
+ ]
1384
+ },
1385
+ "job_link": {
1386
+ "single": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219174",
1387
+ "multi": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219169"
1388
+ },
1389
+ "captured_info": {
1390
+ "single": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219174#step:16:1",
1391
+ "multi": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219169#step:16:1"
1392
+ }
1393
+ },
1394
+ "models_llama": {
1395
+ "failed": {
1396
+ "PyTorch": {
1397
+ "unclassified": 0,
1398
+ "single": 0,
1399
+ "multi": 0
1400
+ },
1401
+ "Tokenizers": {
1402
+ "unclassified": 0,
1403
+ "single": 5,
1404
+ "multi": 5
1405
+ },
1406
+ "Pipelines": {
1407
+ "unclassified": 0,
1408
+ "single": 0,
1409
+ "multi": 0
1410
+ },
1411
+ "Trainer": {
1412
+ "unclassified": 0,
1413
+ "single": 0,
1414
+ "multi": 0
1415
+ },
1416
+ "ONNX": {
1417
+ "unclassified": 0,
1418
+ "single": 0,
1419
+ "multi": 0
1420
+ },
1421
+ "Auto": {
1422
+ "unclassified": 0,
1423
+ "single": 0,
1424
+ "multi": 0
1425
+ },
1426
+ "Quantization": {
1427
+ "unclassified": 0,
1428
+ "single": 0,
1429
+ "multi": 0
1430
+ },
1431
+ "Unclassified": {
1432
+ "unclassified": 0,
1433
+ "single": 0,
1434
+ "multi": 0
1435
+ }
1436
+ },
1437
+ "errors": 0,
1438
+ "success": 377,
1439
+ "skipped": 201,
1440
+ "time_spent": [
1441
+ 314.7,
1442
+ 310.21
1443
+ ],
1444
+ "error": false,
1445
+ "failures": {
1446
+ "single": [
1447
+ {
1448
+ "line": "tests/models/llama/test_tokenization_llama.py::LlamaTokenizationTest::test_added_tokens_serialization",
1449
+ "trace": "(line 483) OSError: You are trying to access a gated repo."
1450
+ },
1451
+ {
1452
+ "line": "tests/models/llama/test_tokenization_llama.py::LlamaTokenizationTest::test_chat_template_return_assistant_tokens_mask",
1453
+ "trace": "(line 483) OSError: You are trying to access a gated repo."
1454
+ },
1455
+ {
1456
+ "line": "tests/models/llama/test_tokenization_llama.py::LlamaTokenizationTest::test_chat_template_return_assistant_tokens_mask_truncated",
1457
+ "trace": "(line 483) OSError: You are trying to access a gated repo."
1458
+ },
1459
+ {
1460
+ "line": "tests/models/llama/test_tokenization_llama.py::LlamaTokenizationTest::test_padding_side_in_kwargs",
1461
+ "trace": "(line 483) OSError: You are trying to access a gated repo."
1462
+ },
1463
+ {
1464
+ "line": "tests/models/llama/test_tokenization_llama.py::LlamaTokenizationTest::test_truncation_side_in_kwargs",
1465
+ "trace": "(line 483) OSError: You are trying to access a gated repo."
1466
+ }
1467
+ ],
1468
+ "multi": [
1469
+ {
1470
+ "line": "tests/models/llama/test_tokenization_llama.py::LlamaTokenizationTest::test_added_tokens_serialization",
1471
+ "trace": "(line 483) OSError: You are trying to access a gated repo."
1472
+ },
1473
+ {
1474
+ "line": "tests/models/llama/test_tokenization_llama.py::LlamaTokenizationTest::test_chat_template_return_assistant_tokens_mask",
1475
+ "trace": "(line 483) OSError: You are trying to access a gated repo."
1476
+ },
1477
+ {
1478
+ "line": "tests/models/llama/test_tokenization_llama.py::LlamaTokenizationTest::test_chat_template_return_assistant_tokens_mask_truncated",
1479
+ "trace": "(line 483) OSError: You are trying to access a gated repo."
1480
+ },
1481
+ {
1482
+ "line": "tests/models/llama/test_tokenization_llama.py::LlamaTokenizationTest::test_padding_side_in_kwargs",
1483
+ "trace": "(line 483) OSError: You are trying to access a gated repo."
1484
+ },
1485
+ {
1486
+ "line": "tests/models/llama/test_tokenization_llama.py::LlamaTokenizationTest::test_truncation_side_in_kwargs",
1487
+ "trace": "(line 483) OSError: You are trying to access a gated repo."
1488
+ }
1489
+ ]
1490
+ },
1491
+ "job_link": {
1492
+ "single": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219189",
1493
+ "multi": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219202"
1494
+ },
1495
+ "captured_info": {}
1496
+ },
1497
+ "models_llava": {
1498
+ "failed": {
1499
+ "PyTorch": {
1500
+ "unclassified": 0,
1501
+ "single": 4,
1502
+ "multi": 4
1503
+ },
1504
+ "Tokenizers": {
1505
+ "unclassified": 0,
1506
+ "single": 0,
1507
+ "multi": 0
1508
+ },
1509
+ "Pipelines": {
1510
+ "unclassified": 0,
1511
+ "single": 0,
1512
+ "multi": 0
1513
+ },
1514
+ "Trainer": {
1515
+ "unclassified": 0,
1516
+ "single": 0,
1517
+ "multi": 0
1518
+ },
1519
+ "ONNX": {
1520
+ "unclassified": 0,
1521
+ "single": 0,
1522
+ "multi": 0
1523
+ },
1524
+ "Auto": {
1525
+ "unclassified": 0,
1526
+ "single": 0,
1527
+ "multi": 0
1528
+ },
1529
+ "Quantization": {
1530
+ "unclassified": 0,
1531
+ "single": 0,
1532
+ "multi": 0
1533
+ },
1534
+ "Unclassified": {
1535
+ "unclassified": 0,
1536
+ "single": 0,
1537
+ "multi": 0
1538
+ }
1539
+ },
1540
+ "errors": 0,
1541
+ "success": 371,
1542
+ "skipped": 257,
1543
+ "time_spent": [
1544
+ 293.2,
1545
+ 302.08
1546
+ ],
1547
+ "error": false,
1548
+ "failures": {
1549
+ "single": [
1550
+ {
1551
+ "line": "tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationIntegrationTest::test_generation_siglip_backbone",
1552
+ "trace": "(line 631) AssertionError: 'user[25 chars]stantConditionsent优于是非是非人家 draw人家 Acceler𝜃墉 no[81 chars]ller' != 'user[25 chars]stant The image shows two cats, one on the lef[84 chars] cat'"
1553
+ },
1554
+ {
1555
+ "line": "tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationIntegrationTest::test_pixtral",
1556
+ "trace": "(line 692) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 140.00 MiB. GPU 0 has a total capacity of 22.30 GiB of which 90.69 MiB is free. Process 17501 has 22.21 GiB memory in use. Of the allocated memory 21.82 GiB is allocated by PyTorch, and 4.11 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)"
1557
+ },
1558
+ {
1559
+ "line": "tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationIntegrationTest::test_pixtral_4bit",
1560
+ "trace": "(line 692) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 1.25 GiB. GPU 0 has a total capacity of 22.30 GiB of which 522.69 MiB is free. Process 17501 has 21.79 GiB memory in use. Of the allocated memory 21.39 GiB is allocated by PyTorch, and 8.05 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)"
1561
+ },
1562
+ {
1563
+ "line": "tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationIntegrationTest::test_tokenizer_integration",
1564
+ "trace": "(line 589) AssertionError: Lists differ: ['<|im_start|>', 'sy', 'st', 'em', '\\n', 'An', 'sw', 'er', ' [245 chars]'\\n'] != ['<|im_start|>', 'system', '\\n', 'Answer', '▁the', '▁question[173 chars]'\\n']"
1565
+ }
1566
+ ],
1567
+ "multi": [
1568
+ {
1569
+ "line": "tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationIntegrationTest::test_generation_siglip_backbone",
1570
+ "trace": "(line 631) AssertionError: 'user[25 chars]stantConditionsent优于是非是非人家 draw人家 Acceler𝜃墉 no[81 chars]ller' != 'user[25 chars]stant The image shows two cats, one on the lef[84 chars] cat'"
1571
+ },
1572
+ {
1573
+ "line": "tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationIntegrationTest::test_pixtral",
1574
+ "trace": "(line 692) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 140.00 MiB. GPU 0 has a total capacity of 22.30 GiB of which 92.69 MiB is free. Process 22420 has 22.21 GiB memory in use. Of the allocated memory 21.69 GiB is allocated by PyTorch, and 12.11 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)"
1575
+ },
1576
+ {
1577
+ "line": "tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationIntegrationTest::test_pixtral_4bit",
1578
+ "trace": "(line 692) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 1.25 GiB. GPU 0 has a total capacity of 22.30 GiB of which 300.69 MiB is free. Process 22420 has 22.00 GiB memory in use. Of the allocated memory 21.13 GiB is allocated by PyTorch, and 380.92 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)"
1579
+ },
1580
+ {
1581
+ "line": "tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationIntegrationTest::test_tokenizer_integration",
1582
+ "trace": "(line 589) AssertionError: Lists differ: ['<|im_start|>', 'sy', 'st', 'em', '\\n', 'An', 'sw', 'er', ' [245 chars]'\\n'] != ['<|im_start|>', 'system', '\\n', 'Answer', '▁the', '▁question[173 chars]'\\n']"
1583
+ }
1584
+ ]
1585
+ },
1586
+ "job_link": {
1587
+ "single": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219205",
1588
+ "multi": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219216"
1589
+ },
1590
+ "captured_info": {
1591
+ "single": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219205#step:16:1",
1592
+ "multi": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219216#step:16:1"
1593
+ }
1594
+ },
1595
+ "models_llava_next": {
1596
+ "failed": {
1597
+ "PyTorch": {
1598
+ "unclassified": 0,
1599
+ "single": 0,
1600
+ "multi": 0
1601
+ },
1602
+ "Tokenizers": {
1603
+ "unclassified": 0,
1604
+ "single": 0,
1605
+ "multi": 0
1606
+ },
1607
+ "Pipelines": {
1608
+ "unclassified": 0,
1609
+ "single": 0,
1610
+ "multi": 0
1611
+ },
1612
+ "Trainer": {
1613
+ "unclassified": 0,
1614
+ "single": 0,
1615
+ "multi": 0
1616
+ },
1617
+ "ONNX": {
1618
+ "unclassified": 0,
1619
+ "single": 0,
1620
+ "multi": 0
1621
+ },
1622
+ "Auto": {
1623
+ "unclassified": 0,
1624
+ "single": 0,
1625
+ "multi": 0
1626
+ },
1627
+ "Quantization": {
1628
+ "unclassified": 0,
1629
+ "single": 0,
1630
+ "multi": 0
1631
+ },
1632
+ "Unclassified": {
1633
+ "unclassified": 0,
1634
+ "single": 0,
1635
+ "multi": 0
1636
+ }
1637
+ },
1638
+ "errors": 0,
1639
+ "success": 367,
1640
+ "skipped": 143,
1641
+ "time_spent": [
1642
+ 956.49,
1643
+ 947.78
1644
+ ],
1645
+ "error": false,
1646
+ "failures": {},
1647
+ "job_link": {
1648
+ "single": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219289",
1649
+ "multi": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219187"
1650
+ },
1651
+ "captured_info": {}
1652
+ },
1653
+ "models_llava_next_video": {
1654
+ "failed": {
1655
+ "PyTorch": {
1656
+ "unclassified": 0,
1657
+ "single": 4,
1658
+ "multi": 4
1659
+ },
1660
+ "Tokenizers": {
1661
+ "unclassified": 0,
1662
+ "single": 0,
1663
+ "multi": 0
1664
+ },
1665
+ "Pipelines": {
1666
+ "unclassified": 0,
1667
+ "single": 0,
1668
+ "multi": 0
1669
+ },
1670
+ "Trainer": {
1671
+ "unclassified": 0,
1672
+ "single": 0,
1673
+ "multi": 0
1674
+ },
1675
+ "ONNX": {
1676
+ "unclassified": 0,
1677
+ "single": 0,
1678
+ "multi": 0
1679
+ },
1680
+ "Auto": {
1681
+ "unclassified": 0,
1682
+ "single": 0,
1683
+ "multi": 0
1684
+ },
1685
+ "Quantization": {
1686
+ "unclassified": 0,
1687
+ "single": 0,
1688
+ "multi": 0
1689
+ },
1690
+ "Unclassified": {
1691
+ "unclassified": 0,
1692
+ "single": 0,
1693
+ "multi": 0
1694
+ }
1695
+ },
1696
+ "errors": 0,
1697
+ "success": 367,
1698
+ "skipped": 111,
1699
+ "time_spent": [
1700
+ 185.86,
1701
+ 179.94
1702
+ ],
1703
+ "error": false,
1704
+ "failures": {
1705
+ "single": [
1706
+ {
1707
+ "line": "tests/models/llava_next_video/test_modeling_llava_next_video.py::LlavaNextVideoForConditionalGenerationIntegrationTest::test_small_model_integration_test",
1708
+ "trace": "(line 134) RuntimeError: expected mat1 and mat2 to have the same dtype, but got: c10::Half != c10::BFloat16"
1709
+ },
1710
+ {
1711
+ "line": "tests/models/llava_next_video/test_modeling_llava_next_video.py::LlavaNextVideoForConditionalGenerationIntegrationTest::test_small_model_integration_test_batch",
1712
+ "trace": "(line 134) RuntimeError: expected mat1 and mat2 to have the same dtype, but got: c10::Half != c10::BFloat16"
1713
+ },
1714
+ {
1715
+ "line": "tests/models/llava_next_video/test_modeling_llava_next_video.py::LlavaNextVideoForConditionalGenerationIntegrationTest::test_small_model_integration_test_batch_different_vision_types",
1716
+ "trace": "(line 134) RuntimeError: expected mat1 and mat2 to have the same dtype, but got: c10::Half != c10::BFloat16"
1717
+ },
1718
+ {
1719
+ "line": "tests/models/llava_next_video/test_modeling_llava_next_video.py::LlavaNextVideoForConditionalGenerationIntegrationTest::test_small_model_integration_test_batch_matches_single",
1720
+ "trace": "(line 134) RuntimeError: expected mat1 and mat2 to have the same dtype, but got: c10::Half != c10::BFloat16"
1721
+ }
1722
+ ],
1723
+ "multi": [
1724
+ {
1725
+ "line": "tests/models/llava_next_video/test_modeling_llava_next_video.py::LlavaNextVideoForConditionalGenerationIntegrationTest::test_small_model_integration_test",
1726
+ "trace": "(line 134) RuntimeError: expected mat1 and mat2 to have the same dtype, but got: c10::Half != c10::BFloat16"
1727
+ },
1728
+ {
1729
+ "line": "tests/models/llava_next_video/test_modeling_llava_next_video.py::LlavaNextVideoForConditionalGenerationIntegrationTest::test_small_model_integration_test_batch",
1730
+ "trace": "(line 134) RuntimeError: expected mat1 and mat2 to have the same dtype, but got: c10::Half != c10::BFloat16"
1731
+ },
1732
+ {
1733
+ "line": "tests/models/llava_next_video/test_modeling_llava_next_video.py::LlavaNextVideoForConditionalGenerationIntegrationTest::test_small_model_integration_test_batch_different_vision_types",
1734
+ "trace": "(line 134) RuntimeError: expected mat1 and mat2 to have the same dtype, but got: c10::Half != c10::BFloat16"
1735
+ },
1736
+ {
1737
+ "line": "tests/models/llava_next_video/test_modeling_llava_next_video.py::LlavaNextVideoForConditionalGenerationIntegrationTest::test_small_model_integration_test_batch_matches_single",
1738
+ "trace": "(line 134) RuntimeError: expected mat1 and mat2 to have the same dtype, but got: c10::Half != c10::BFloat16"
1739
+ }
1740
+ ]
1741
+ },
1742
+ "job_link": {
1743
+ "single": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219224",
1744
+ "multi": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219133"
1745
+ },
1746
+ "captured_info": {}
1747
+ },
1748
+ "models_mistral": {
1749
+ "failed": {
1750
+ "PyTorch": {
1751
+ "unclassified": 0,
1752
+ "single": 0,
1753
+ "multi": 0
1754
+ },
1755
+ "Tokenizers": {
1756
+ "unclassified": 0,
1757
+ "single": 0,
1758
+ "multi": 0
1759
+ },
1760
+ "Pipelines": {
1761
+ "unclassified": 0,
1762
+ "single": 0,
1763
+ "multi": 0
1764
+ },
1765
+ "Trainer": {
1766
+ "unclassified": 0,
1767
+ "single": 0,
1768
+ "multi": 0
1769
+ },
1770
+ "ONNX": {
1771
+ "unclassified": 0,
1772
+ "single": 0,
1773
+ "multi": 0
1774
+ },
1775
+ "Auto": {
1776
+ "unclassified": 0,
1777
+ "single": 0,
1778
+ "multi": 0
1779
+ },
1780
+ "Quantization": {
1781
+ "unclassified": 0,
1782
+ "single": 0,
1783
+ "multi": 0
1784
+ },
1785
+ "Unclassified": {
1786
+ "unclassified": 0,
1787
+ "single": 0,
1788
+ "multi": 0
1789
+ }
1790
+ },
1791
+ "errors": 0,
1792
+ "success": 297,
1793
+ "skipped": 193,
1794
+ "time_spent": [
1795
+ 234.31,
1796
+ 236.91
1797
+ ],
1798
+ "error": false,
1799
+ "failures": {},
1800
+ "job_link": {
1801
+ "single": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219228",
1802
+ "multi": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219122"
1803
+ },
1804
+ "captured_info": {
1805
+ "single": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219228#step:16:1",
1806
+ "multi": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219122#step:16:1"
1807
+ }
1808
+ },
1809
+ "models_paligemma": {
1810
+ "failed": {
1811
+ "PyTorch": {
1812
+ "unclassified": 0,
1813
+ "single": 5,
1814
+ "multi": 5
1815
+ },
1816
+ "Tokenizers": {
1817
+ "unclassified": 0,
1818
+ "single": 0,
1819
+ "multi": 0
1820
+ },
1821
+ "Pipelines": {
1822
+ "unclassified": 0,
1823
+ "single": 0,
1824
+ "multi": 0
1825
+ },
1826
+ "Trainer": {
1827
+ "unclassified": 0,
1828
+ "single": 0,
1829
+ "multi": 0
1830
+ },
1831
+ "ONNX": {
1832
+ "unclassified": 0,
1833
+ "single": 0,
1834
+ "multi": 0
1835
+ },
1836
+ "Auto": {
1837
+ "unclassified": 0,
1838
+ "single": 0,
1839
+ "multi": 0
1840
+ },
1841
+ "Quantization": {
1842
+ "unclassified": 0,
1843
+ "single": 0,
1844
+ "multi": 0
1845
+ },
1846
+ "Unclassified": {
1847
+ "unclassified": 0,
1848
+ "single": 0,
1849
+ "multi": 0
1850
+ }
1851
+ },
1852
+ "errors": 0,
1853
+ "success": 276,
1854
+ "skipped": 170,
1855
+ "time_spent": [
1856
+ 329.74,
1857
+ 317.98
1858
+ ],
1859
+ "error": false,
1860
+ "failures": {
1861
+ "multi": [
1862
+ {
1863
+ "line": "tests/models/paligemma/test_modeling_paligemma.py::PaliGemmaForConditionalGenerationIntegrationTest::test_integration_detection_bug",
1864
+ "trace": "(line 677) AssertionError: 'detect shoe\\n<loc0045><loc0309><loc0708><loc0646> shoe' != 'detect shoe\\n<loc0051><loc0309><loc0708><loc0646> shoe'"
1865
+ },
1866
+ {
1867
+ "line": "tests/models/paligemma/test_modeling_paligemma.py::PaliGemmaForConditionalGenerationIntegrationTest::test_paligemma_finetuning_with_suffixes_bf16",
1868
+ "trace": "(line 613) AssertionError: assert False"
1869
+ },
1870
+ {
1871
+ "line": "tests/models/paligemma/test_modeling_paligemma.py::PaliGemmaForConditionalGenerationIntegrationTest::test_small_model_integration_test_paligemma_batched",
1872
+ "trace": "(line 677) AssertionError: Lists differ: ['ans[13 chars]is the cow standing?\\nbeach', '\\n a cow standing on the beach'] != ['ans[13 chars]is the cow standing?\\nbeach', '\\ncow on the beach']"
1873
+ },
1874
+ {
1875
+ "line": "tests/models/paligemma/test_modeling_paligemma.py::PaliGemmaForConditionalGenerationIntegrationTest::test_small_model_integration_test_paligemma_batched_bf16",
1876
+ "trace": "(line 677) AssertionError: Lists differ: ['ans[13 chars]is the cow standing?\\nbeach', '\\n a cow standing on the beach'] != ['ans[13 chars]is the cow standing?\\nbeach', '\\ncow on the beach']"
1877
+ },
1878
+ {
1879
+ "line": "tests/models/paligemma/test_modeling_paligemma.py::PaliGemmaForConditionalGenerationIntegrationTest::test_small_model_integration_test_paligemma_batched_f16",
1880
+ "trace": "(line 677) AssertionError: Lists differ: ['ans[13 chars]is the cow standing?\\nbeach', '\\n a cow standing on the beach'] != ['ans[13 chars]is the cow standing?\\nbeach', '\\ncow on the beach']"
1881
+ }
1882
+ ],
1883
+ "single": [
1884
+ {
1885
+ "line": "tests/models/paligemma/test_modeling_paligemma.py::PaliGemmaForConditionalGenerationIntegrationTest::test_integration_detection_bug",
1886
+ "trace": "(line 677) AssertionError: 'detect shoe\\n<loc0045><loc0309><loc0708><loc0646> shoe' != 'detect shoe\\n<loc0051><loc0309><loc0708><loc0646> shoe'"
1887
+ },
1888
+ {
1889
+ "line": "tests/models/paligemma/test_modeling_paligemma.py::PaliGemmaForConditionalGenerationIntegrationTest::test_paligemma_finetuning_with_suffixes_bf16",
1890
+ "trace": "(line 613) AssertionError: assert False"
1891
+ },
1892
+ {
1893
+ "line": "tests/models/paligemma/test_modeling_paligemma.py::PaliGemmaForConditionalGenerationIntegrationTest::test_small_model_integration_test_paligemma_batched",
1894
+ "trace": "(line 677) AssertionError: Lists differ: ['ans[13 chars]is the cow standing?\\nbeach', '\\n a cow standing on the beach'] != ['ans[13 chars]is the cow standing?\\nbeach', '\\ncow on the beach']"
1895
+ },
1896
+ {
1897
+ "line": "tests/models/paligemma/test_modeling_paligemma.py::PaliGemmaForConditionalGenerationIntegrationTest::test_small_model_integration_test_paligemma_batched_bf16",
1898
+ "trace": "(line 677) AssertionError: Lists differ: ['ans[13 chars]is the cow standing?\\nbeach', '\\n a cow standing on the beach'] != ['ans[13 chars]is the cow standing?\\nbeach', '\\ncow on the beach']"
1899
+ },
1900
+ {
1901
+ "line": "tests/models/paligemma/test_modeling_paligemma.py::PaliGemmaForConditionalGenerationIntegrationTest::test_small_model_integration_test_paligemma_batched_f16",
1902
+ "trace": "(line 677) AssertionError: Lists differ: ['ans[13 chars]is the cow standing?\\nbeach', '\\n a cow standing on the beach'] != ['ans[13 chars]is the cow standing?\\nbeach', '\\ncow on the beach']"
1903
+ }
1904
+ ]
1905
+ },
1906
+ "job_link": {
1907
+ "multi": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219259",
1908
+ "single": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219252"
1909
+ },
1910
+ "captured_info": {
1911
+ "multi": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219259#step:16:1",
1912
+ "single": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219252#step:16:1"
1913
+ }
1914
+ },
1915
+ "models_qwen2_vl": {
1916
+ "failed": {
1917
+ "PyTorch": {
1918
+ "unclassified": 0,
1919
+ "single": 0,
1920
+ "multi": 0
1921
+ },
1922
+ "Tokenizers": {
1923
+ "unclassified": 0,
1924
+ "single": 0,
1925
+ "multi": 0
1926
+ },
1927
+ "Pipelines": {
1928
+ "unclassified": 0,
1929
+ "single": 0,
1930
+ "multi": 0
1931
+ },
1932
+ "Trainer": {
1933
+ "unclassified": 0,
1934
+ "single": 0,
1935
+ "multi": 0
1936
+ },
1937
+ "ONNX": {
1938
+ "unclassified": 0,
1939
+ "single": 0,
1940
+ "multi": 0
1941
+ },
1942
+ "Auto": {
1943
+ "unclassified": 0,
1944
+ "single": 0,
1945
+ "multi": 0
1946
+ },
1947
+ "Quantization": {
1948
+ "unclassified": 0,
1949
+ "single": 0,
1950
+ "multi": 0
1951
+ },
1952
+ "Unclassified": {
1953
+ "unclassified": 0,
1954
+ "single": 0,
1955
+ "multi": 0
1956
+ }
1957
+ },
1958
+ "errors": 0,
1959
+ "success": 417,
1960
+ "skipped": 243,
1961
+ "time_spent": [
1962
+ 309.66,
1963
+ 310.83
1964
+ ],
1965
+ "error": false,
1966
+ "failures": {},
1967
+ "job_link": {
1968
+ "single": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219230",
1969
+ "multi": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219165"
1970
+ },
1971
+ "captured_info": {}
1972
+ },
1973
+ "models_t5gemma": {
1974
+ "failed": {
1975
+ "PyTorch": {
1976
+ "unclassified": 0,
1977
+ "single": 0,
1978
+ "multi": 0
1979
+ },
1980
+ "Tokenizers": {
1981
+ "unclassified": 0,
1982
+ "single": 0,
1983
+ "multi": 0
1984
+ },
1985
+ "Pipelines": {
1986
+ "unclassified": 0,
1987
+ "single": 0,
1988
+ "multi": 0
1989
+ },
1990
+ "Trainer": {
1991
+ "unclassified": 0,
1992
+ "single": 0,
1993
+ "multi": 0
1994
+ },
1995
+ "ONNX": {
1996
+ "unclassified": 0,
1997
+ "single": 0,
1998
+ "multi": 0
1999
+ },
2000
+ "Auto": {
2001
+ "unclassified": 0,
2002
+ "single": 0,
2003
+ "multi": 0
2004
+ },
2005
+ "Quantization": {
2006
+ "unclassified": 0,
2007
+ "single": 0,
2008
+ "multi": 0
2009
+ },
2010
+ "Unclassified": {
2011
+ "unclassified": 0,
2012
+ "single": 0,
2013
+ "multi": 0
2014
+ }
2015
+ },
2016
+ "errors": 0,
2017
+ "success": 423,
2018
+ "skipped": 297,
2019
+ "time_spent": [
2020
+ 252.51,
2021
+ 255.21
2022
+ ],
2023
+ "error": false,
2024
+ "failures": {},
2025
+ "job_link": {
2026
+ "single": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219293",
2027
+ "multi": "https://github.com/huggingface/transformers/actions/runs/20926567527/job/60126219219"
2028
+ },
2029
+ "captured_info": {}
2030
+ }
2031
+ }