hf-transformers-bot commited on
Commit
3d353ed
·
verified ·
1 Parent(s): 7ff24e5

Upload 2026-04-04/runs/753-23973914236/ci_results_run_models_gpu/new_failures_against_23970187589.json with huggingface_hub

Browse files
2026-04-04/runs/753-23973914236/ci_results_run_models_gpu/new_failures_against_23970187589.json ADDED
@@ -0,0 +1,1265 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "afmoe": {
3
+ "failures": {
4
+ "multi-gpu": [
5
+ {
6
+ "line": "tests/models/afmoe/test_modeling_afmoe.py::AfmoeIntegrationTest::test_compile_static_cache",
7
+ "trace": "(line 200) AssertionError: Lists differ: ['Sim[225 chars] it on everything. I love it on my eggs, on my burgers, on'] != ['Sim[225 chars] it on everything. I love the taste of it. I love the smell']"
8
+ }
9
+ ]
10
+ },
11
+ "job_link": {
12
+ "single": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928368505",
13
+ "multi": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928368462"
14
+ }
15
+ },
16
+ "audioflamingo3": {
17
+ "failures": {
18
+ "multi-gpu": [
19
+ {
20
+ "line": "tests/models/audioflamingo3/test_modeling_audioflamingo3.py::AudioFlamingo3ForConditionalGenerationIntegrationTest::test_fixture_batched_matches",
21
+ "trace": "(line 352) AssertionError: The values for attribute 'shape' do not match: torch.Size([2, 23]) != torch.Size([2, 24])."
22
+ }
23
+ ],
24
+ "single-gpu": [
25
+ {
26
+ "line": "tests/models/audioflamingo3/test_modeling_audioflamingo3.py::AudioFlamingo3ForConditionalGenerationIntegrationTest::test_fixture_batched_matches",
27
+ "trace": "(line 352) AssertionError: The values for attribute 'shape' do not match: torch.Size([2, 23]) != torch.Size([2, 24])."
28
+ }
29
+ ]
30
+ },
31
+ "job_link": {
32
+ "multi": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928368488",
33
+ "single": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928368509"
34
+ }
35
+ },
36
+ "auto": {
37
+ "failures": {
38
+ "single-gpu": [
39
+ {
40
+ "line": "tests/models/auto/test_tokenization_auto.py::AutoTokenizerTest::test_custom_tokenizer_init",
41
+ "trace": "(line 258) ImportError: This modeling file requires the following packages that were not found in your environment: matplotlib. Run `pip install matplotlib`"
42
+ }
43
+ ],
44
+ "multi-gpu": [
45
+ {
46
+ "line": "tests/models/auto/test_tokenization_auto.py::AutoTokenizerTest::test_custom_tokenizer_init",
47
+ "trace": "(line 258) ImportError: This modeling file requires the following packages that were not found in your environment: matplotlib. Run `pip install matplotlib`"
48
+ }
49
+ ]
50
+ },
51
+ "job_link": {
52
+ "single": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928368497",
53
+ "multi": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928368474"
54
+ }
55
+ },
56
+ "blip_2": {
57
+ "failures": {
58
+ "single-gpu": [
59
+ {
60
+ "line": "tests/models/blip_2/test_modeling_blip_2.py::Blip2ModelIntegrationTest::test_inference_vision_with_projection_fp16",
61
+ "trace": "(line 687) AssertionError: False is not true"
62
+ }
63
+ ],
64
+ "multi-gpu": [
65
+ {
66
+ "line": "tests/models/blip_2/test_modeling_blip_2.py::Blip2ModelIntegrationTest::test_inference_vision_with_projection_fp16",
67
+ "trace": "(line 687) AssertionError: False is not true"
68
+ }
69
+ ]
70
+ },
71
+ "job_link": {
72
+ "single": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928368951",
73
+ "multi": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928368890"
74
+ }
75
+ },
76
+ "conditional_detr": {
77
+ "failures": {
78
+ "multi-gpu": [
79
+ {
80
+ "line": "tests/models/conditional_detr/test_image_processing_conditional_detr.py::ConditionalDetrImageProcessingTest::test_batched_coco_detection_annotations",
81
+ "trace": "(line 243) ImportError: Pycocotools is not installed in your environment."
82
+ }
83
+ ],
84
+ "single-gpu": [
85
+ {
86
+ "line": "tests/models/conditional_detr/test_image_processing_conditional_detr.py::ConditionalDetrImageProcessingTest::test_batched_coco_detection_annotations",
87
+ "trace": "(line 243) ImportError: Pycocotools is not installed in your environment."
88
+ }
89
+ ]
90
+ },
91
+ "job_link": {
92
+ "multi": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928369357",
93
+ "single": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928369402"
94
+ }
95
+ },
96
+ "deformable_detr": {
97
+ "failures": {
98
+ "multi-gpu": [
99
+ {
100
+ "line": "tests/models/deformable_detr/test_image_processing_deformable_detr.py::DeformableDetrImageProcessingTest::test_batched_coco_detection_annotations",
101
+ "trace": "(line 87) ImportError: Pycocotools is not installed in your environment."
102
+ },
103
+ {
104
+ "line": "tests/models/deformable_detr/test_modeling_deformable_detr.py::DeformableDetrModelTest::test_batching_equivalence",
105
+ "trace": "(line 1525) AssertionError: Batched and Single row outputs are not equal in DeformableDetrModel for key=encoder_last_hidden_state."
106
+ }
107
+ ],
108
+ "single-gpu": [
109
+ {
110
+ "line": "tests/models/deformable_detr/test_image_processing_deformable_detr.py::DeformableDetrImageProcessingTest::test_batched_coco_detection_annotations",
111
+ "trace": "(line 87) ImportError: Pycocotools is not installed in your environment."
112
+ }
113
+ ]
114
+ },
115
+ "job_link": {
116
+ "multi": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928369759",
117
+ "single": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928369802"
118
+ }
119
+ },
120
+ "detr": {
121
+ "failures": {
122
+ "single-gpu": [
123
+ {
124
+ "line": "tests/models/detr/test_image_processing_detr.py::DetrImageProcessingTest::test_batched_coco_detection_annotations",
125
+ "trace": "(line 241) ImportError: Pycocotools is not installed in your environment."
126
+ }
127
+ ],
128
+ "multi-gpu": [
129
+ {
130
+ "line": "tests/models/detr/test_image_processing_detr.py::DetrImageProcessingTest::test_batched_coco_detection_annotations",
131
+ "trace": "(line 241) ImportError: Pycocotools is not installed in your environment."
132
+ }
133
+ ]
134
+ },
135
+ "job_link": {
136
+ "single": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928369881",
137
+ "multi": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928369852"
138
+ }
139
+ },
140
+ "edgetam_video": {
141
+ "failures": {
142
+ "single-gpu": [
143
+ {
144
+ "line": "tests/models/edgetam_video/test_modeling_edgetam_video.py::EdgeTamVideoModelIntegrationTest::test_inference_mask_generation_video_multi_objects_multi_points",
145
+ "trace": "(line 363) AssertionError: Tensor-likes are not close!"
146
+ },
147
+ {
148
+ "line": "tests/models/edgetam_video/test_modeling_edgetam_video.py::EdgeTamVideoModelIntegrationTest::test_inference_mask_generation_video_multi_points",
149
+ "trace": "(line 195) AssertionError: Tensor-likes are not close!"
150
+ },
151
+ {
152
+ "line": "tests/models/edgetam_video/test_modeling_edgetam_video.py::EdgeTamVideoModelIntegrationTest::test_inference_mask_generation_video_one_bb",
153
+ "trace": "(line 250) AssertionError: Tensor-likes are not close!"
154
+ },
155
+ {
156
+ "line": "tests/models/edgetam_video/test_modeling_edgetam_video.py::EdgeTamVideoModelIntegrationTest::test_inference_mask_generation_video_one_point",
157
+ "trace": "(line 101) AssertionError: Tensor-likes are not close!"
158
+ },
159
+ {
160
+ "line": "tests/models/edgetam_video/test_modeling_edgetam_video.py::EdgeTamVideoModelIntegrationTest::test_inference_mask_generation_video_one_point_one_bb",
161
+ "trace": "(line 307) AssertionError: Tensor-likes are not close!"
162
+ },
163
+ {
164
+ "line": "tests/models/edgetam_video/test_modeling_edgetam_video.py::EdgeTamVideoModelIntegrationTest::test_inference_mask_generation_video_one_point_propagate_in_video_directly",
165
+ "trace": "(line 162) AssertionError: Tensor-likes are not close!"
166
+ },
167
+ {
168
+ "line": "tests/models/edgetam_video/test_modeling_edgetam_video.py::EdgeTamVideoModelIntegrationTest::test_inference_propagate_video_from_mask_input",
169
+ "trace": "(line 452) AssertionError: Tensor-likes are not close!"
170
+ }
171
+ ]
172
+ },
173
+ "job_link": {
174
+ "single": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928370178",
175
+ "multi": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928370188"
176
+ }
177
+ },
178
+ "eomt_dinov3": {
179
+ "failures": {
180
+ "single-gpu": [
181
+ {
182
+ "line": "tests/models/eomt_dinov3/test_modeling_eomt_dinov3.py::EomtDinov3ForUniversalSegmentationIntegrationTest::test_inference_bf16",
183
+ "trace": "(line 310) AssertionError: Tensor-likes are not close!"
184
+ }
185
+ ],
186
+ "multi-gpu": [
187
+ {
188
+ "line": "tests/models/eomt_dinov3/test_modeling_eomt_dinov3.py::EomtDinov3ForUniversalSegmentationIntegrationTest::test_inference_bf16",
189
+ "trace": "(line 310) AssertionError: Tensor-likes are not close!"
190
+ }
191
+ ]
192
+ },
193
+ "job_link": {
194
+ "single": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928370329",
195
+ "multi": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928370353"
196
+ }
197
+ },
198
+ "ernie4_5_vl_moe": {
199
+ "failures": {
200
+ "single-gpu": [
201
+ {
202
+ "line": "tests/models/ernie4_5_vl_moe/test_modeling_ernie4_5_vl_moe.py::Ernie4_5_VLMoeSmallIntegrationTest::test_small_model_integration_test_expand",
203
+ "trace": "(line 713) AssertionError: Lists differ: ['不是啊[43 chars]是啊不是啊不是啊错的错的错的错的错的错的错的错的错的错的错的错的', '不是啊不是啊不是啊不[67 chars]是这样'] != ['不是啊[43 chars]是啊不是啊错的错的错的错的错的错的错的错的错的错的错的错的错的', '不是啊不是啊不是啊不是[65 chars]是这样']"
204
+ }
205
+ ]
206
+ },
207
+ "job_link": {
208
+ "single": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928370404"
209
+ }
210
+ },
211
+ "florence2": {
212
+ "failures": {
213
+ "single-gpu": [
214
+ {
215
+ "line": "tests/models/florence2/test_modeling_florence2.py::Florence2ForConditionalGenerationIntegrationTest::test_base_model_batching_inference_eager",
216
+ "trace": "(line 333) AssertionError: Lists differ: [[2, [23 chars] 50981, 50269, 50269, 50688, 50942, 50269, 503[412 chars], 1]] != [[2, [23 chars] 50980, 50269, 50269, 50688, 50942, 50269, 503[412 chars], 1]]"
217
+ },
218
+ {
219
+ "line": "tests/models/florence2/test_modeling_florence2.py::Florence2ForConditionalGenerationIntegrationTest::test_large_model_batching_inference_sdpa",
220
+ "trace": "(line 584) AssertionError: Lists differ: [[2, [178 chars] 50447, 50457, 50447, 50483, 50401, 50482, 401[727 chars], 1]] != [[2, [178 chars] 50446, 50457, 50446, 50483, 50401, 50482, 401[727 chars], 1]]"
221
+ }
222
+ ],
223
+ "multi-gpu": [
224
+ {
225
+ "line": "tests/models/florence2/test_modeling_florence2.py::Florence2ForConditionalGenerationIntegrationTest::test_base_model_batching_inference_eager",
226
+ "trace": "(line 333) AssertionError: Lists differ: [[2, [23 chars] 50981, 50269, 50269, 50688, 50942, 50269, 503[412 chars], 1]] != [[2, [23 chars] 50980, 50269, 50269, 50688, 50942, 50269, 503[412 chars], 1]]"
227
+ },
228
+ {
229
+ "line": "tests/models/florence2/test_modeling_florence2.py::Florence2ForConditionalGenerationIntegrationTest::test_large_model_batching_inference_sdpa",
230
+ "trace": "(line 584) AssertionError: Lists differ: [[2, [178 chars] 50447, 50457, 50447, 50483, 50401, 50482, 401[727 chars], 1]] != [[2, [178 chars] 50446, 50457, 50446, 50483, 50401, 50482, 401[727 chars], 1]]"
231
+ }
232
+ ]
233
+ },
234
+ "job_link": {
235
+ "single": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928370623",
236
+ "multi": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928370567"
237
+ }
238
+ },
239
+ "gemma4": {
240
+ "failures": {
241
+ "multi-gpu": [
242
+ {
243
+ "line": "tests/models/gemma4/test_modeling_gemma4.py::Gemma4Audio2TextModelTest::test_sdpa_can_compile_dynamic",
244
+ "trace": "(line 271) torch._inductor.exc.InductorError: AssertionError:"
245
+ }
246
+ ],
247
+ "single-gpu": [
248
+ {
249
+ "line": "tests/models/gemma4/test_modeling_gemma4.py::Gemma4TextModelTest::test_torch_compile_for_training",
250
+ "trace": "(line 8201) torch._dynamo.exc.BackendCompilerFailed: backend='inductor' raised:"
251
+ },
252
+ {
253
+ "line": "tests/models/gemma4/test_modeling_gemma4.py::Gemma4Audio2TextModelTest::test_sdpa_can_compile_dynamic",
254
+ "trace": "(line 271) torch._inductor.exc.InductorError: AssertionError:"
255
+ },
256
+ {
257
+ "line": "tests/models/gemma4/test_modeling_gemma4.py::Gemma4Vision2TextModelTest::test_sdpa_can_dispatch_on_flash",
258
+ "trace": "(line 92) RuntimeError: No available kernel. Aborting execution."
259
+ }
260
+ ]
261
+ },
262
+ "job_link": {
263
+ "multi": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928370739",
264
+ "single": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928370824"
265
+ }
266
+ },
267
+ "glm": {
268
+ "failures": {
269
+ "single-gpu": [
270
+ {
271
+ "line": "tests/models/glm/test_modeling_glm.py::GlmIntegrationTest::test_model_9b_eager",
272
+ "trace": "(line 133) AssertionError: Lists differ: ['Hel[140 chars]ou how to make a simple and easy to make a DIY air freshener'] != ['Hel[140 chars]ou how to make a simple and easy to make a DIY paper lantern.']"
273
+ },
274
+ {
275
+ "line": "tests/models/glm/test_modeling_glm.py::GlmIntegrationTest::test_model_9b_fp16",
276
+ "trace": "(line 1369) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 214.00 MiB. GPU 0 has a total capacity of 22.30 GiB of which 78.69 MiB is free. Process 193084 has 22.22 GiB memory in use. Of the allocated memory 21.83 GiB is allocated by PyTorch, and 8.56 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://docs.pytorch.org/docs/stable/notes/cuda.html#optimizing-memory-usage-with-pytorch-cuda-alloc-conf)"
277
+ },
278
+ {
279
+ "line": "tests/models/glm/test_modeling_glm.py::GlmIntegrationTest::test_model_9b_sdpa",
280
+ "trace": "(line 1369) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 1.16 GiB. GPU 0 has a total capacity of 22.30 GiB of which 78.69 MiB is free. Process 193084 has 22.22 GiB memory in use. Of the allocated memory 21.83 GiB is allocated by PyTorch, and 8.56 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://docs.pytorch.org/docs/stable/notes/cuda.html#optimizing-memory-usage-with-pytorch-cuda-alloc-conf)"
281
+ }
282
+ ]
283
+ },
284
+ "job_link": {
285
+ "single": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928370836"
286
+ }
287
+ },
288
+ "glm4": {
289
+ "failures": {
290
+ "single-gpu": [
291
+ {
292
+ "line": "tests/models/glm4/test_modeling_glm4.py::Glm4IntegrationTest::test_model_9b_bf16",
293
+ "trace": "(line 120) AssertionError: Lists differ: ['Hel[95 chars]e was. I', 'Hi today I am going to tell you ab[70 chars]tes'] != ['Hel[95 chars]e was and what', 'Hi today I am going to tell [76 chars]tes']"
294
+ },
295
+ {
296
+ "line": "tests/models/glm4/test_modeling_glm4.py::Glm4IntegrationTest::test_model_9b_eager",
297
+ "trace": "(line 1369) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 214.00 MiB. GPU 0 has a total capacity of 22.30 GiB of which 76.69 MiB is free. Process 245193 has 22.22 GiB memory in use. Of the allocated memory 21.83 GiB is allocated by PyTorch, and 9.81 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://docs.pytorch.org/docs/stable/notes/cuda.html#optimizing-memory-usage-with-pytorch-cuda-alloc-conf)"
298
+ },
299
+ {
300
+ "line": "tests/models/glm4/test_modeling_glm4.py::Glm4IntegrationTest::test_model_9b_sdpa",
301
+ "trace": "(line 182) AssertionError: Lists differ: ['Hel[95 chars]e was. I', 'Hi today I am going to tell you ab[70 chars]tes'] != ['Hel[95 chars]e was and what', 'Hi today I am going to tell [76 chars]tes']"
302
+ }
303
+ ],
304
+ "multi-gpu": [
305
+ {
306
+ "line": "tests/models/glm4/test_modeling_glm4.py::Glm4IntegrationTest::test_model_9b_bf16",
307
+ "trace": "(line 120) AssertionError: Lists differ: ['Hel[95 chars]e was. I', 'Hi today I am going to tell you ab[70 chars]tes'] != ['Hel[95 chars]e was and what', 'Hi today I am going to tell [76 chars]tes']"
308
+ },
309
+ {
310
+ "line": "tests/models/glm4/test_modeling_glm4.py::Glm4IntegrationTest::test_model_9b_eager",
311
+ "trace": "(line 1369) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 32.00 MiB. GPU 0 has a total capacity of 22.30 GiB of which 26.69 MiB is free. Process 104148 has 22.27 GiB memory in use. Of the allocated memory 21.78 GiB is allocated by PyTorch, and 5.69 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://docs.pytorch.org/docs/stable/notes/cuda.html#optimizing-memory-usage-with-pytorch-cuda-alloc-conf)"
312
+ },
313
+ {
314
+ "line": "tests/models/glm4/test_modeling_glm4.py::Glm4IntegrationTest::test_model_9b_sdpa",
315
+ "trace": "(line 182) AssertionError: Lists differ: ['Hel[95 chars]e was. I', 'Hi today I am going to tell you ab[70 chars]tes'] != ['Hel[95 chars]e was and what', 'Hi today I am going to tell [76 chars]tes']"
316
+ }
317
+ ]
318
+ },
319
+ "job_link": {
320
+ "single": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928370854",
321
+ "multi": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928370863"
322
+ }
323
+ },
324
+ "glm_moe_dsa": {
325
+ "failures": {
326
+ "single-gpu": [
327
+ {
328
+ "line": "tests/models/glm_moe_dsa/test_modeling_glm_moe_dsa.py::GlmMoeDsaModelTest::test_eager_matches_batched_and_grouped_inference_2_bf16",
329
+ "trace": "(line 626) AssertionError: Tensor-likes are not close!"
330
+ }
331
+ ]
332
+ },
333
+ "job_link": {
334
+ "single": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928370992",
335
+ "multi": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928370917"
336
+ }
337
+ },
338
+ "granitemoe": {
339
+ "failures": {
340
+ "multi-gpu": [
341
+ {
342
+ "line": "tests/models/granitemoe/test_modeling_granitemoe.py::GraniteMoeModelTest::test_sdpa_can_compile_dynamic",
343
+ "trace": "(line 2511) torch._inductor.exc.InductorError: AssertionError: both a fallback and a decomp for same op: aten.index_add.default"
344
+ }
345
+ ]
346
+ },
347
+ "job_link": {
348
+ "multi": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928371215"
349
+ }
350
+ },
351
+ "granitemoehybrid": {
352
+ "failures": {
353
+ "multi-gpu": [
354
+ {
355
+ "line": "tests/models/granitemoehybrid/test_modeling_granitemoehybrid.py::GraniteMoeHybridModelTest::test_sdpa_can_compile_dynamic",
356
+ "trace": "(line 2511) torch._inductor.exc.InductorError: AssertionError: both a fallback and a decomp for same op: aten.index_add.default"
357
+ }
358
+ ]
359
+ },
360
+ "job_link": {
361
+ "multi": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928371218"
362
+ }
363
+ },
364
+ "granitemoeshared": {
365
+ "failures": {
366
+ "multi-gpu": [
367
+ {
368
+ "line": "tests/models/granitemoeshared/test_modeling_granitemoeshared.py::GraniteMoeSharedModelTest::test_sdpa_can_compile_dynamic",
369
+ "trace": "(line 2511) torch._inductor.exc.InductorError: AssertionError: both a fallback and a decomp for same op: aten.index_add.default"
370
+ }
371
+ ]
372
+ },
373
+ "job_link": {
374
+ "multi": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928371221"
375
+ }
376
+ },
377
+ "grounding_dino": {
378
+ "failures": {
379
+ "single-gpu": [
380
+ {
381
+ "line": "tests/models/grounding_dino/test_image_processing_grounding_dino.py::GroundingDinoImageProcessingTest::test_batched_coco_detection_annotations",
382
+ "trace": "(line 96) ImportError: Pycocotools is not installed in your environment."
383
+ }
384
+ ],
385
+ "multi-gpu": [
386
+ {
387
+ "line": "tests/models/grounding_dino/test_image_processing_grounding_dino.py::GroundingDinoImageProcessingTest::test_batched_coco_detection_annotations",
388
+ "trace": "(line 96) ImportError: Pycocotools is not installed in your environment."
389
+ }
390
+ ]
391
+ },
392
+ "job_link": {
393
+ "single": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928371316",
394
+ "multi": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928371243"
395
+ }
396
+ },
397
+ "janus": {
398
+ "failures": {
399
+ "multi-gpu": [
400
+ {
401
+ "line": "tests/models/janus/test_modeling_janus.py::JanusIntegrationTest::test_model_generate_images",
402
+ "trace": "(line 687) AssertionError: False is not true"
403
+ },
404
+ {
405
+ "line": "tests/models/janus/test_modeling_janus.py::JanusIntegrationTest::test_model_text_generation_batched",
406
+ "trace": "(line 482) AssertionError: Lists differ: ['You[334 chars] and is one', 'You are a helpful language and [293 chars]lso'] != ['You[334 chars] and the constellation', 'You are a helpful la[310 chars]Peg']"
407
+ },
408
+ {
409
+ "line": "tests/models/janus/test_modeling_janus.py::JanusIntegrationTest::test_model_text_generation_with_multi_image",
410
+ "trace": "(line 505) AssertionError: Lists differ: ['You[276 chars]same star constellation. The first image shows[112 chars]art'] != ['You[276 chars]same constellation. The first image shows the [123 chars]ape']"
411
+ }
412
+ ],
413
+ "single-gpu": [
414
+ {
415
+ "line": "tests/models/janus/test_modeling_janus.py::JanusIntegrationTest::test_model_generate_images",
416
+ "trace": "(line 687) AssertionError: False is not true"
417
+ },
418
+ {
419
+ "line": "tests/models/janus/test_modeling_janus.py::JanusIntegrationTest::test_model_text_generation_batched",
420
+ "trace": "(line 482) AssertionError: Lists differ: ['You[334 chars] and is one', 'You are a helpful language and [293 chars]lso'] != ['You[334 chars] and the constellation', 'You are a helpful la[310 chars]Peg']"
421
+ },
422
+ {
423
+ "line": "tests/models/janus/test_modeling_janus.py::JanusIntegrationTest::test_model_text_generation_with_multi_image",
424
+ "trace": "(line 505) AssertionError: Lists differ: ['You[276 chars]same star constellation. The first image shows[112 chars]art'] != ['You[276 chars]same constellation. The first image shows the [123 chars]ape']"
425
+ }
426
+ ]
427
+ },
428
+ "job_link": {
429
+ "multi": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928371754",
430
+ "single": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928371775"
431
+ }
432
+ },
433
+ "jetmoe": {
434
+ "failures": {
435
+ "multi-gpu": [
436
+ {
437
+ "line": "tests/models/jetmoe/test_modeling_jetmoe.py::JetMoeModelTest::test_sdpa_can_compile_dynamic",
438
+ "trace": "(line 2511) torch._inductor.exc.InductorError: AssertionError: both a fallback and a decomp for same op: aten.index_add.default"
439
+ }
440
+ ],
441
+ "single-gpu": [
442
+ {
443
+ "line": "tests/models/jetmoe/test_modeling_jetmoe.py::JetMoeModelTest::test_sdpa_can_compile_dynamic",
444
+ "trace": "(line 2511) torch._inductor.exc.InductorError: AssertionError: both a fallback and a decomp for same op: aten.index_add.default"
445
+ }
446
+ ]
447
+ },
448
+ "job_link": {
449
+ "multi": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928371783",
450
+ "single": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928371784"
451
+ }
452
+ },
453
+ "kosmos2_5": {
454
+ "failures": {
455
+ "single-gpu": [
456
+ {
457
+ "line": "tests/models/kosmos2_5/test_modeling_kosmos2_5.py::Kosmos2_5ModelIntegrationTest::test_sdpa",
458
+ "trace": "(line 604) AssertionError: Lists differ: ['<bb[216 chars]<y_650></bbox>COOKIE DOH SAUCES\\n<bbox><x_788>[452 chars]0\\n'] != ['<bb[216 chars]<y_651></bbox>COOKIE DOH SAUCES\\n<bbox><x_788>[452 chars]0\\n']"
459
+ }
460
+ ],
461
+ "multi-gpu": [
462
+ {
463
+ "line": "tests/models/kosmos2_5/test_modeling_kosmos2_5.py::Kosmos2_5ModelIntegrationTest::test_sdpa",
464
+ "trace": "(line 604) AssertionError: Lists differ: ['<bb[216 chars]<y_650></bbox>COOKIE DOH SAUCES\\n<bbox><x_788>[452 chars]0\\n'] != ['<bb[216 chars]<y_651></bbox>COOKIE DOH SAUCES\\n<bbox><x_788>[452 chars]0\\n']"
465
+ }
466
+ ]
467
+ },
468
+ "job_link": {
469
+ "single": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928371801",
470
+ "multi": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928371825"
471
+ }
472
+ },
473
+ "lfm2_vl": {
474
+ "failures": {
475
+ "single-gpu": [
476
+ {
477
+ "line": "tests/models/lfm2_vl/test_modeling_lfm2_vl.py::Lfm2_5VlForConditionalGenerationIntegrationTest::test_integration_test_batched",
478
+ "trace": "(line 369) AssertionError: Lists differ: ['In [90 chars]ands tall on a small', 'In this image, we see [69 chars]s a'] != ['In [90 chars]ands on Liberty Island in', 'In this image, we[74 chars]s a']"
479
+ },
480
+ {
481
+ "line": "tests/models/lfm2_vl/test_modeling_lfm2_vl.py::Lfm2_5VlForConditionalGenerationIntegrationTest::test_integration_test_high_resolution",
482
+ "trace": "(line 347) AssertionError: 'In t[52 chars]ymbol of freedom and democracy. It stands tall on a small' != 'In t[52 chars]ymbol of freedom and democracy. It stands on Liberty Island in'"
483
+ }
484
+ ],
485
+ "multi-gpu": [
486
+ {
487
+ "line": "tests/models/lfm2_vl/test_modeling_lfm2_vl.py::Lfm2_5VlForConditionalGenerationIntegrationTest::test_integration_test_batched",
488
+ "trace": "(line 369) AssertionError: Lists differ: ['In [90 chars]ands tall on a small', 'In this image, we see [69 chars]s a'] != ['In [90 chars]ands on Liberty Island in', 'In this image, we[74 chars]s a']"
489
+ },
490
+ {
491
+ "line": "tests/models/lfm2_vl/test_modeling_lfm2_vl.py::Lfm2_5VlForConditionalGenerationIntegrationTest::test_integration_test_high_resolution",
492
+ "trace": "(line 347) AssertionError: 'In t[52 chars]ymbol of freedom and democracy. It stands tall on a small' != 'In t[52 chars]ymbol of freedom and democracy. It stands on Liberty Island in'"
493
+ }
494
+ ]
495
+ },
496
+ "job_link": {
497
+ "single": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928372064",
498
+ "multi": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928371990"
499
+ }
500
+ },
501
+ "luke": {
502
+ "failures": {
503
+ "multi-gpu": [
504
+ {
505
+ "line": "tests/models/luke/test_modeling_luke.py::LukeModelTest::test_multi_gpu_data_parallel_forward",
506
+ "trace": "(line 774) StopIteration: Caught StopIteration in replica 1 on device 1."
507
+ },
508
+ {
509
+ "line": "tests/models/luke/test_modeling_luke.py::LukeModelIntegrationTests::test_inference_base_model",
510
+ "trace": "(line 905) AssertionError: Tensor-likes are not close!"
511
+ },
512
+ {
513
+ "line": "tests/models/luke/test_modeling_luke.py::LukeModelIntegrationTests::test_inference_large_model",
514
+ "trace": "(line 940) AssertionError: Tensor-likes are not close!"
515
+ },
516
+ {
517
+ "line": "tests/models/luke/test_tokenization_luke.py::LukeTokenizerTest::test_sequence_builders",
518
+ "trace": "(line 55) AssertionError: Lists differ: [0, 46665, 21043, 2] != [46665, 21043]"
519
+ }
520
+ ]
521
+ },
522
+ "job_link": {
523
+ "single": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928372279",
524
+ "multi": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928372247"
525
+ }
526
+ },
527
+ "maskformer": {
528
+ "failures": {
529
+ "multi-gpu": [
530
+ {
531
+ "line": "tests/models/maskformer/test_modeling_maskformer.py::MaskFormerModelIntegrationTest::test_inference_instance_segmentation_head",
532
+ "trace": "(line 625) AssertionError: Tensor-likes are not close!"
533
+ },
534
+ {
535
+ "line": "tests/models/maskformer/test_modeling_maskformer.py::MaskFormerModelIntegrationTest::test_inference_instance_segmentation_head_resnet_backbone",
536
+ "trace": "(line 659) AssertionError: Tensor-likes are not close!"
537
+ }
538
+ ]
539
+ },
540
+ "job_link": {
541
+ "multi": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928372400",
542
+ "single": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928372451"
543
+ }
544
+ },
545
+ "minimax": {
546
+ "failures": {
547
+ "single-gpu": [
548
+ {
549
+ "line": "tests/models/minimax/test_modeling_minimax.py::MiniMaxIntegrationTest::test_small_model_logits",
550
+ "trace": "(line 233) AssertionError: Tensor-likes are not close!"
551
+ }
552
+ ]
553
+ },
554
+ "job_link": {
555
+ "single": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928372628"
556
+ }
557
+ },
558
+ "moonshine": {
559
+ "failures": {
560
+ "multi-gpu": [
561
+ {
562
+ "line": "tests/models/moonshine/test_modeling_moonshine.py::MoonshineModelTest::test_sdpa_can_compile_dynamic",
563
+ "trace": "(line 271) torch._inductor.exc.InductorError: AssertionError:"
564
+ }
565
+ ],
566
+ "single-gpu": [
567
+ {
568
+ "line": "tests/models/moonshine/test_modeling_moonshine.py::MoonshineModelTest::test_sdpa_can_compile_dynamic",
569
+ "trace": "(line 271) torch._inductor.exc.InductorError: AssertionError:"
570
+ }
571
+ ]
572
+ },
573
+ "job_link": {
574
+ "multi": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928368482",
575
+ "single": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928368431"
576
+ }
577
+ },
578
+ "moonshine_streaming": {
579
+ "failures": {
580
+ "multi-gpu": [
581
+ {
582
+ "line": "tests/models/moonshine_streaming/test_modeling_moonshine_streaming.py::MoonshineStreamingModelTest::test_sdpa_can_compile_dynamic",
583
+ "trace": "(line 271) torch._inductor.exc.InductorError: AssertionError:"
584
+ }
585
+ ],
586
+ "single-gpu": [
587
+ {
588
+ "line": "tests/models/moonshine_streaming/test_modeling_moonshine_streaming.py::MoonshineStreamingModelTest::test_sdpa_can_compile_dynamic",
589
+ "trace": "(line 271) torch._inductor.exc.InductorError: AssertionError:"
590
+ }
591
+ ]
592
+ },
593
+ "job_link": {
594
+ "multi": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928368666",
595
+ "single": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928368631"
596
+ }
597
+ },
598
+ "parakeet": {
599
+ "failures": {
600
+ "single-gpu": [
601
+ {
602
+ "line": "tests/models/parakeet/test_modeling_parakeet.py::ParakeetForCTCIntegrationTest::test_1b_model_integration_batched",
603
+ "trace": "(line 373) AssertionError: Tensor-likes are not equal!"
604
+ }
605
+ ],
606
+ "multi-gpu": [
607
+ {
608
+ "line": "tests/models/parakeet/test_modeling_parakeet.py::ParakeetForCTCIntegrationTest::test_1b_model_integration_batched",
609
+ "trace": "(line 373) AssertionError: Tensor-likes are not equal!"
610
+ }
611
+ ]
612
+ },
613
+ "job_link": {
614
+ "single": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928369266",
615
+ "multi": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928369298"
616
+ }
617
+ },
618
+ "qwen2": {
619
+ "failures": {
620
+ "multi-gpu": [
621
+ {
622
+ "line": "tests/models/qwen2/test_modeling_qwen2.py::Qwen2IntegrationTest::test_model_450m_logits",
623
+ "trace": "(line 88) AssertionError: Tensor-likes are not close!"
624
+ },
625
+ {
626
+ "line": "tests/models/qwen2/test_modeling_qwen2.py::Qwen2IntegrationTest::test_speculative_generation",
627
+ "trace": "(line 205) AssertionError: 'My f[22 chars]100% real, organic, vegan and gluten free. I u[19 chars] and' != 'My f[22 chars]100% natural, organic, gluten-free, vegan, and[27 chars]king'"
628
+ }
629
+ ]
630
+ },
631
+ "job_link": {
632
+ "multi": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928369979",
633
+ "single": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928369963"
634
+ }
635
+ },
636
+ "qwen2_5_omni": {
637
+ "failures": {
638
+ "multi-gpu": [
639
+ {
640
+ "line": "tests/models/qwen2_5_omni/test_modeling_qwen2_5_omni.py::Qwen2_5OmniThinkerForConditionalGenerationModelTest::test_multi_gpu_data_parallel_forward",
641
+ "trace": "(line 774) StopIteration: Caught StopIteration in replica 1 on device 1."
642
+ },
643
+ {
644
+ "line": "tests/models/qwen2_5_omni/test_modeling_qwen2_5_omni.py::Qwen2_5OmniModelIntegrationTest::test_small_model_integration_test_batch",
645
+ "trace": "(line 734) AssertionError: Lists differ: [\"sys[109 chars]d is glass shattering, and the dog is a Labrad[185 chars]er.\"] != [\"sys[109 chars]d is a glass shattering. The dog in the pictur[211 chars]er.\"]"
646
+ }
647
+ ],
648
+ "single-gpu": [
649
+ {
650
+ "line": "tests/models/qwen2_5_omni/test_modeling_qwen2_5_omni.py::Qwen2_5OmniModelIntegrationTest::test_small_model_integration_test_batch",
651
+ "trace": "(line 734) AssertionError: Lists differ: [\"sys[109 chars]d is glass shattering, and the dog is a Labrad[185 chars]er.\"] != [\"sys[109 chars]d is a glass shattering. The dog in the pictur[211 chars]er.\"]"
652
+ }
653
+ ]
654
+ },
655
+ "job_link": {
656
+ "multi": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928369999",
657
+ "single": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928369985"
658
+ }
659
+ },
660
+ "qwen2_5_vl": {
661
+ "failures": {
662
+ "multi-gpu": [
663
+ {
664
+ "line": "tests/models/qwen2_5_vl/test_modeling_qwen2_5_vl.py::Qwen2_5_VLIntegrationTest::test_small_model_integration_test_batch",
665
+ "trace": "(line 524) AssertionError: Lists differ: ['sys[176 chars] and gentle nature, which is evident in', 'sys[212 chars] in'] != ['sys[176 chars] and energetic nature, which is evident in', '[218 chars] in']"
666
+ },
667
+ {
668
+ "line": "tests/models/qwen2_5_vl/test_modeling_qwen2_5_vl.py::Qwen2_5_VLIntegrationTest::test_small_model_integration_test_expand",
669
+ "trace": "(line 545) AssertionError: Lists differ: ['sys[176 chars] and gentle nature, which is evident in', 'sys[437 chars] in'] != ['sys[176 chars] and energetic nature, which is evident in', '[443 chars] in']"
670
+ }
671
+ ],
672
+ "single-gpu": [
673
+ {
674
+ "line": "tests/models/qwen2_5_vl/test_modeling_qwen2_5_vl.py::Qwen2_5_VLIntegrationTest::test_small_model_integration_test_batch",
675
+ "trace": "(line 524) AssertionError: Lists differ: ['sys[176 chars] and gentle nature, which is evident in', 'sys[212 chars] in'] != ['sys[176 chars] and energetic nature, which is evident in', '[218 chars] in']"
676
+ },
677
+ {
678
+ "line": "tests/models/qwen2_5_vl/test_modeling_qwen2_5_vl.py::Qwen2_5_VLIntegrationTest::test_small_model_integration_test_expand",
679
+ "trace": "(line 545) AssertionError: Lists differ: ['sys[176 chars] and gentle nature, which is evident in', 'sys[437 chars] in'] != ['sys[176 chars] and energetic nature, which is evident in', '[443 chars] in']"
680
+ }
681
+ ]
682
+ },
683
+ "job_link": {
684
+ "multi": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928370011",
685
+ "single": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928369971"
686
+ }
687
+ },
688
+ "qwen2_vl": {
689
+ "failures": {
690
+ "single-gpu": [
691
+ {
692
+ "line": "tests/models/qwen2_vl/test_modeling_qwen2_vl.py::Qwen2VLIntegrationTest::test_small_model_integration_test_batch_different_resolutions",
693
+ "trace": "(line 615) AssertionError: Lists differ: ['sys[348 chars]be a Golden Retriever. Golden Retrievers are k[58 chars]are'] != ['sys[348 chars]be a Labrador Retriever. Labradors are known f[66 chars]ces']"
694
+ }
695
+ ],
696
+ "multi-gpu": [
697
+ {
698
+ "line": "tests/models/qwen2_vl/test_modeling_qwen2_vl.py::Qwen2VLIntegrationTest::test_small_model_integration_test_batch_different_resolutions",
699
+ "trace": "(line 615) AssertionError: Lists differ: ['sys[348 chars]be a Golden Retriever. Golden Retrievers are k[58 chars]are'] != ['sys[348 chars]be a Labrador Retriever. Labradors are known f[66 chars]ces']"
700
+ }
701
+ ]
702
+ },
703
+ "job_link": {
704
+ "single": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928369987",
705
+ "multi": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928370075"
706
+ }
707
+ },
708
+ "qwen3_5": {
709
+ "failures": {
710
+ "multi-gpu": [
711
+ {
712
+ "line": "tests/models/qwen3_5/test_modeling_qwen3_5.py::Qwen3_5ModelTest::test_model_parallel_beam_search",
713
+ "trace": "(line 2567) RuntimeError: Expected all tensors to be on the same device, but got index is on cuda:0, different from other tensors on cuda:1 (when checking argument in method wrapper_CUDA__index_select)"
714
+ },
715
+ {
716
+ "line": "tests/models/qwen3_5/test_modeling_qwen3_5.py::Qwen3_5ModelTest::test_model_parallelism",
717
+ "trace": "(line 1967) AttributeError: 'Qwen3_5Model' object has no attribute 'hf_device_map'"
718
+ }
719
+ ]
720
+ },
721
+ "job_link": {
722
+ "single": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928370009",
723
+ "multi": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928370120"
724
+ }
725
+ },
726
+ "qwen3_omni_moe": {
727
+ "failures": {
728
+ "multi-gpu": [
729
+ {
730
+ "line": "tests/models/qwen3_omni_moe/test_modeling_qwen3_omni_moe.py::Qwen3OmniMoeThinkerForConditionalGenerationModelTest::test_model_parallelism",
731
+ "trace": "(line 1967) AttributeError: 'Qwen3OmniMoeThinkerForConditionalGeneration' object has no attribute 'hf_device_map'"
732
+ },
733
+ {
734
+ "line": "tests/models/qwen3_omni_moe/test_modeling_qwen3_omni_moe.py::Qwen3OmniMoeThinkerForConditionalGenerationModelTest::test_multi_gpu_data_parallel_forward",
735
+ "trace": "(line 774) StopIteration: Caught StopIteration in replica 1 on device 1."
736
+ },
737
+ {
738
+ "line": "tests/models/qwen3_omni_moe/test_modeling_qwen3_omni_moe.py::Qwen3OmniModelIntegrationTest::test_small_model_integration_test",
739
+ "trace": "(line 273) RuntimeError: We encountered some issues during automatic conversion of the weights. For details look at the `CONVERSION` entries of the above report!"
740
+ },
741
+ {
742
+ "line": "tests/models/qwen3_omni_moe/test_modeling_qwen3_omni_moe.py::Qwen3OmniModelIntegrationTest::test_small_model_integration_test_batch",
743
+ "trace": "(line 273) RuntimeError: We encountered some issues during automatic conversion of the weights. For details look at the `CONVERSION` entries of the above report!"
744
+ },
745
+ {
746
+ "line": "tests/models/qwen3_omni_moe/test_modeling_qwen3_omni_moe.py::Qwen3OmniModelIntegrationTest::test_small_model_integration_test_multiturn",
747
+ "trace": "(line 273) RuntimeError: We encountered some issues during automatic conversion of the weights. For details look at the `CONVERSION` entries of the above report!"
748
+ },
749
+ {
750
+ "line": "tests/models/qwen3_omni_moe/test_modeling_qwen3_omni_moe.py::Qwen3OmniModelIntegrationTest::test_small_model_integration_test_w_audio",
751
+ "trace": "(line 273) RuntimeError: We encountered some issues during automatic conversion of the weights. For details look at the `CONVERSION` entries of the above report!"
752
+ }
753
+ ]
754
+ },
755
+ "job_link": {
756
+ "multi": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928370220"
757
+ }
758
+ },
759
+ "qwen3_vl": {
760
+ "failures": {
761
+ "multi-gpu": [
762
+ {
763
+ "line": "tests/models/qwen3_vl/test_modeling_qwen3_vl.py::Qwen3VLModelTest::test_model_parallel_beam_search",
764
+ "trace": "(line 2567) RuntimeError: Expected all tensors to be on the same device, but got index is on cuda:0, different from other tensors on cuda:1 (when checking argument in method wrapper_CUDA__index_select)"
765
+ },
766
+ {
767
+ "line": "tests/models/qwen3_vl/test_modeling_qwen3_vl.py::Qwen3VLModelTest::test_model_parallelism",
768
+ "trace": "(line 1967) AttributeError: 'Qwen3VLModel' object has no attribute 'hf_device_map'"
769
+ },
770
+ {
771
+ "line": "tests/models/qwen3_vl/test_modeling_qwen3_vl.py::Qwen3VLModelTest::test_multi_gpu_data_parallel_forward",
772
+ "trace": "(line 774) StopIteration: Caught StopIteration in replica 1 on device 1."
773
+ }
774
+ ]
775
+ },
776
+ "job_link": {
777
+ "single": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928370184",
778
+ "multi": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928370217"
779
+ }
780
+ },
781
+ "qwen3_vl_moe": {
782
+ "failures": {
783
+ "multi-gpu": [
784
+ {
785
+ "line": "tests/models/qwen3_vl_moe/test_modeling_qwen3_vl_moe.py::Qwen3VLMoeModelTest::test_multi_gpu_data_parallel_forward",
786
+ "trace": "(line 774) StopIteration: Caught StopIteration in replica 1 on device 1."
787
+ },
788
+ {
789
+ "line": "tests/models/qwen3_vl_moe/test_modeling_qwen3_vl_moe.py::Qwen3VLMoeIntegrationTest::test_small_model_integration_test",
790
+ "trace": "(line 273) RuntimeError: We encountered some issues during automatic conversion of the weights. For details look at the `CONVERSION` entries of the above report!"
791
+ },
792
+ {
793
+ "line": "tests/models/qwen3_vl_moe/test_modeling_qwen3_vl_moe.py::Qwen3VLMoeIntegrationTest::test_small_model_integration_test_batch",
794
+ "trace": "(line 273) RuntimeError: We encountered some issues during automatic conversion of the weights. For details look at the `CONVERSION` entries of the above report!"
795
+ },
796
+ {
797
+ "line": "tests/models/qwen3_vl_moe/test_modeling_qwen3_vl_moe.py::Qwen3VLMoeIntegrationTest::test_small_model_integration_test_batch_different_resolutions",
798
+ "trace": "(line 273) RuntimeError: We encountered some issues during automatic conversion of the weights. For details look at the `CONVERSION` entries of the above report!"
799
+ },
800
+ {
801
+ "line": "tests/models/qwen3_vl_moe/test_modeling_qwen3_vl_moe.py::Qwen3VLMoeIntegrationTest::test_small_model_integration_test_batch_wo_image",
802
+ "trace": "(line 273) RuntimeError: We encountered some issues during automatic conversion of the weights. For details look at the `CONVERSION` entries of the above report!"
803
+ },
804
+ {
805
+ "line": "tests/models/qwen3_vl_moe/test_modeling_qwen3_vl_moe.py::Qwen3VLMoeIntegrationTest::test_small_model_integration_test_expand",
806
+ "trace": "(line 273) RuntimeError: We encountered some issues during automatic conversion of the weights. For details look at the `CONVERSION` entries of the above report!"
807
+ },
808
+ {
809
+ "line": "tests/models/qwen3_vl_moe/test_modeling_qwen3_vl_moe.py::Qwen3VLMoeIntegrationTest::test_small_model_integration_test_expand_with_video",
810
+ "trace": "(line 273) RuntimeError: We encountered some issues during automatic conversion of the weights. For details look at the `CONVERSION` entries of the above report!"
811
+ },
812
+ {
813
+ "line": "tests/models/qwen3_vl_moe/test_modeling_qwen3_vl_moe.py::Qwen3VLMoeIntegrationTest::test_small_model_integration_test_with_video",
814
+ "trace": "(line 528) OSError: Qwen/Qwen3-VL-30B-A3B-Instruct does not appear to have a file named model-00001-of-00013.safetensors. Checkout 'https://huggingface.co/Qwen/Qwen3-VL-30B-A3B-Instruct/tree/main' for available files."
815
+ }
816
+ ]
817
+ },
818
+ "job_link": {
819
+ "multi": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928370231"
820
+ }
821
+ },
822
+ "rag": {
823
+ "failures": {
824
+ "single-gpu": [
825
+ {
826
+ "line": "tests/models/rag/test_modeling_rag.py::RagModelIntegrationTests::test_rag_sequence_generate_batch",
827
+ "trace": "(line 419) AttributeError: 'RagConfig' object has no attribute 'num_return_sequences'"
828
+ },
829
+ {
830
+ "line": "tests/models/rag/test_modeling_rag.py::RagModelIntegrationTests::test_rag_sequence_generate_batch_from_context_input_ids",
831
+ "trace": "(line 419) AttributeError: 'RagConfig' object has no attribute 'num_return_sequences'"
832
+ },
833
+ {
834
+ "line": "tests/models/rag/test_modeling_rag.py::RagModelIntegrationTests::test_rag_sequence_generate_beam",
835
+ "trace": "(line 892) AssertionError: '\" in the United States. \"People Need Love\"[155 chars]hit.' != '\"She\\'s My Kind of Girl\" was released thro[257 chars]nts.'"
836
+ },
837
+ {
838
+ "line": "tests/models/rag/test_modeling_rag.py::RagModelIntegrationTests::test_rag_token_generate_beam",
839
+ "trace": "(line 854) AssertionError: '\"She[14 chars] Girl' != '\"She[14 chars] Girl\" was released through Epic Records in Ja[179 chars]ses\"'"
840
+ },
841
+ {
842
+ "line": "tests/models/rag/test_modeling_rag.py::RagModelSaveLoadTests::test_rag_token_from_pretrained",
843
+ "trace": "(line 4089) TypeError: DPRQuestionEncoder.__init__() got an unexpected keyword argument 'max_length'"
844
+ }
845
+ ]
846
+ },
847
+ "job_link": {
848
+ "single": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928370209",
849
+ "multi": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928370225"
850
+ }
851
+ },
852
+ "seamless_m4t": {
853
+ "failures": {
854
+ "multi-gpu": [
855
+ {
856
+ "line": "tests/models/seamless_m4t/test_modeling_seamless_m4t.py::SeamlessM4TModelWithTextInputTest::test_model_parallelism",
857
+ "trace": "(line 157) RuntimeError: indices should be either on cpu or on the same device as the indexed tensor (cuda:0)"
858
+ }
859
+ ]
860
+ },
861
+ "job_link": {
862
+ "multi": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928370568",
863
+ "single": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928370536"
864
+ }
865
+ },
866
+ "smolvlm": {
867
+ "failures": {
868
+ "single-gpu": [
869
+ {
870
+ "line": "tests/models/smolvlm/test_video_processing_smolvlm.py::SmolVLMVideoProcessingTest::test_can_compile_fast_video_processor",
871
+ "trace": "(line 215) AssertionError: Tensor-likes are not close!"
872
+ }
873
+ ],
874
+ "multi-gpu": [
875
+ {
876
+ "line": "tests/models/smolvlm/test_video_processing_smolvlm.py::SmolVLMVideoProcessingTest::test_can_compile_fast_video_processor",
877
+ "trace": "(line 215) AssertionError: Tensor-likes are not close!"
878
+ }
879
+ ]
880
+ },
881
+ "job_link": {
882
+ "single": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928370746",
883
+ "multi": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928370812"
884
+ }
885
+ },
886
+ "stablelm": {
887
+ "failures": {
888
+ "single-gpu": [
889
+ {
890
+ "line": "tests/models/stablelm/test_modeling_stablelm.py::StableLmModelIntegrationTest::test_model_stablelm_3b_4e1t_logits",
891
+ "trace": "(line 65) AssertionError: Tensor-likes are not close!"
892
+ },
893
+ {
894
+ "line": "tests/models/stablelm/test_modeling_stablelm.py::StableLmModelIntegrationTest::test_model_tiny_random_stablelm_2_logits",
895
+ "trace": "(line 98) AssertionError: Tensor-likes are not close!"
896
+ }
897
+ ]
898
+ },
899
+ "job_link": {
900
+ "multi": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928370897",
901
+ "single": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928370904"
902
+ }
903
+ },
904
+ "swin2sr": {
905
+ "failures": {
906
+ "multi-gpu": [
907
+ {
908
+ "line": "tests/models/swin2sr/test_modeling_swin2sr.py::Swin2SRModelIntegrationTest::test_inference_fp16",
909
+ "trace": "(line 332) AssertionError: Tensor-likes are not close!"
910
+ }
911
+ ]
912
+ },
913
+ "job_link": {
914
+ "multi": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928371028",
915
+ "single": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928371027"
916
+ }
917
+ },
918
+ "switch_transformers": {
919
+ "failures": {
920
+ "single-gpu": [
921
+ {
922
+ "line": "tests/models/switch_transformers/test_modeling_switch_transformers.py::SwitchTransformersModelTest::test_generate_with_past_key_values",
923
+ "trace": "(line 687) AssertionError: tensor(False, device='cuda:0') is not true"
924
+ }
925
+ ],
926
+ "multi-gpu": [
927
+ {
928
+ "line": "tests/models/switch_transformers/test_modeling_switch_transformers.py::SwitchTransformersModelTest::test_multi_gpu_data_parallel_forward",
929
+ "trace": "(line 774) StopIteration: Caught StopIteration in replica 1 on device 1."
930
+ }
931
+ ]
932
+ },
933
+ "job_link": {
934
+ "single": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928371043",
935
+ "multi": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928371035"
936
+ }
937
+ },
938
+ "unispeech": {
939
+ "failures": {
940
+ "single-gpu": [
941
+ {
942
+ "line": "tests/models/unispeech/test_modeling_unispeech.py::UniSpeechRobustModelTest::test_batched_inference",
943
+ "trace": "(line 687) AssertionError: False is not true"
944
+ }
945
+ ]
946
+ },
947
+ "job_link": {
948
+ "multi": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928371376",
949
+ "single": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928371329"
950
+ }
951
+ },
952
+ "upernet": {
953
+ "failures": {
954
+ "single-gpu": [
955
+ {
956
+ "line": "tests/models/upernet/test_modeling_upernet.py::UperNetModelIntegrationTest::test_inference_convnext_backbone",
957
+ "trace": "(line 306) AssertionError: Tensor-likes are not close!"
958
+ }
959
+ ],
960
+ "multi-gpu": [
961
+ {
962
+ "line": "tests/models/upernet/test_modeling_upernet.py::UperNetModelIntegrationTest::test_inference_convnext_backbone",
963
+ "trace": "(line 306) AssertionError: Tensor-likes are not close!"
964
+ }
965
+ ]
966
+ },
967
+ "job_link": {
968
+ "single": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928371377",
969
+ "multi": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928371403"
970
+ }
971
+ },
972
+ "vibevoice_acoustic_tokenizer": {
973
+ "failures": {
974
+ "multi-gpu": [
975
+ {
976
+ "line": "tests/models/vibevoice_acoustic_tokenizer/test_modeling_vibevoice_acoustic_tokenizer.py::VibeVoiceAcousticTokenizerIntegrationTest::test_batch_integration",
977
+ "trace": "(line 343) AssertionError: Tensor-likes are not close!"
978
+ }
979
+ ],
980
+ "single-gpu": [
981
+ {
982
+ "line": "tests/models/vibevoice_acoustic_tokenizer/test_modeling_vibevoice_acoustic_tokenizer.py::VibeVoiceAcousticTokenizerIntegrationTest::test_batch_integration",
983
+ "trace": "(line 343) AssertionError: Tensor-likes are not close!"
984
+ }
985
+ ]
986
+ },
987
+ "job_link": {
988
+ "multi": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928371457",
989
+ "single": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928371422"
990
+ }
991
+ },
992
+ "videomt": {
993
+ "failures": {
994
+ "single-gpu": [
995
+ {
996
+ "line": "tests/models/videomt/test_modeling_videomt.py::VideomtForUniversalSegmentationTest::test_sdpa_can_compile_dynamic",
997
+ "trace": "(line 271) torch._inductor.exc.InductorError: AssertionError:"
998
+ }
999
+ ],
1000
+ "multi-gpu": [
1001
+ {
1002
+ "line": "tests/models/videomt/test_modeling_videomt.py::VideomtForUniversalSegmentationTest::test_sdpa_can_compile_dynamic",
1003
+ "trace": "(line 271) torch._inductor.exc.InductorError: AssertionError:"
1004
+ }
1005
+ ]
1006
+ },
1007
+ "job_link": {
1008
+ "single": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928371624",
1009
+ "multi": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928371635"
1010
+ }
1011
+ },
1012
+ "xcodec": {
1013
+ "failures": {
1014
+ "single-gpu": [
1015
+ {
1016
+ "line": "tests/models/xcodec/test_modeling_xcodec.py::XcodecIntegrationTest::test_integration_06_hf_audio_xcodec_hubert_general_1_0",
1017
+ "trace": "(line 331) AssertionError: Tensor-likes are not equal!"
1018
+ },
1019
+ {
1020
+ "line": "tests/models/xcodec/test_modeling_xcodec.py::XcodecIntegrationTest::test_integration_07_hf_audio_xcodec_hubert_general_1_5",
1021
+ "trace": "(line 331) AssertionError: Tensor-likes are not equal!"
1022
+ },
1023
+ {
1024
+ "line": "tests/models/xcodec/test_modeling_xcodec.py::XcodecIntegrationTest::test_integration_08_hf_audio_xcodec_hubert_general_2_0",
1025
+ "trace": "(line 331) AssertionError: Tensor-likes are not equal!"
1026
+ },
1027
+ {
1028
+ "line": "tests/models/xcodec/test_modeling_xcodec.py::XcodecIntegrationTest::test_integration_12_hf_audio_xcodec_hubert_general_balanced_1_5",
1029
+ "trace": "(line 342) AssertionError: Tensor-likes are not close!"
1030
+ }
1031
+ ],
1032
+ "multi-gpu": [
1033
+ {
1034
+ "line": "tests/models/xcodec/test_modeling_xcodec.py::XcodecIntegrationTest::test_integration_06_hf_audio_xcodec_hubert_general_1_0",
1035
+ "trace": "(line 331) AssertionError: Tensor-likes are not equal!"
1036
+ },
1037
+ {
1038
+ "line": "tests/models/xcodec/test_modeling_xcodec.py::XcodecIntegrationTest::test_integration_07_hf_audio_xcodec_hubert_general_1_5",
1039
+ "trace": "(line 331) AssertionError: Tensor-likes are not equal!"
1040
+ },
1041
+ {
1042
+ "line": "tests/models/xcodec/test_modeling_xcodec.py::XcodecIntegrationTest::test_integration_08_hf_audio_xcodec_hubert_general_2_0",
1043
+ "trace": "(line 331) AssertionError: Tensor-likes are not equal!"
1044
+ },
1045
+ {
1046
+ "line": "tests/models/xcodec/test_modeling_xcodec.py::XcodecIntegrationTest::test_integration_12_hf_audio_xcodec_hubert_general_balanced_1_5",
1047
+ "trace": "(line 342) AssertionError: Tensor-likes are not close!"
1048
+ }
1049
+ ]
1050
+ },
1051
+ "job_link": {
1052
+ "single": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928372034",
1053
+ "multi": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928372063"
1054
+ }
1055
+ },
1056
+ "yolos": {
1057
+ "failures": {
1058
+ "single-gpu": [
1059
+ {
1060
+ "line": "tests/models/yolos/test_image_processing_yolos.py::YolosImageProcessingTest::test_batched_coco_detection_annotations",
1061
+ "trace": "(line 69) ImportError: Pycocotools is not installed in your environment."
1062
+ }
1063
+ ],
1064
+ "multi-gpu": [
1065
+ {
1066
+ "line": "tests/models/yolos/test_image_processing_yolos.py::YolosImageProcessingTest::test_batched_coco_detection_annotations",
1067
+ "trace": "(line 69) ImportError: Pycocotools is not installed in your environment."
1068
+ }
1069
+ ]
1070
+ },
1071
+ "job_link": {
1072
+ "single": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928372211",
1073
+ "multi": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928372181"
1074
+ }
1075
+ },
1076
+ "pipelines": {
1077
+ "failures": {
1078
+ "multi-gpu": [
1079
+ {
1080
+ "line": "tests/pipelines/test_pipelines_any_to_any.py::AnyToAnyPipelineTests::test_small_model_pt_chat_with_response_parsing",
1081
+ "trace": "(line 330) AssertionError: 'last_word' not found in {'first_word': 'The'}"
1082
+ },
1083
+ {
1084
+ "line": "tests/pipelines/test_pipelines_any_to_any.py::AnyToAnyPipelineTests::test_small_model_pt_image_gen",
1085
+ "trace": "(line 514) ValueError: PIL.Image.Image is not a valid TensorType, please select one of ['pt', 'np', 'mlx']"
1086
+ },
1087
+ {
1088
+ "line": "tests/pipelines/test_pipelines_any_to_any.py::AnyToAnyPipelineTests::test_small_model_pt_token_audio_gen",
1089
+ "trace": "(line 134) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 3.76 GiB. GPU 0 has a total capacity of 22.30 GiB of which 3.15 GiB is free. Process 161692 has 19.14 GiB memory in use. Of the allocated memory 13.66 GiB is allocated by PyTorch, and 5.17 GiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://docs.pytorch.org/docs/stable/notes/cuda.html#optimizing-memory-usage-with-pytorch-cuda-alloc-conf)"
1090
+ },
1091
+ {
1092
+ "line": "tests/pipelines/test_pipelines_any_to_any.py::AnyToAnyPipelineTests::test_small_model_pt_token_audio_input",
1093
+ "trace": "(line 352) AssertionError: Lists differ: [{'in[242 chars]ated_text': \"user\\nWhat do you hear in this au[163 chars]s.\"}] != [{'in[242 chars]ated_None': \"user\\nWhat do you hear in this au[147 chars]n.\"}]"
1094
+ },
1095
+ {
1096
+ "line": "tests/pipelines/test_pipelines_any_to_any.py::AnyToAnyPipelineTests::test_small_model_pt_token_text_only",
1097
+ "trace": "(line 199) AssertionError: Lists differ: [{'in[90 chars]ital of France? Assistant: The capital of France is Paris.\\n'}] != [{'in[90 chars]ital of France? Assistant: The capital of France is Paris.'}]"
1098
+ },
1099
+ {
1100
+ "line": "tests/pipelines/test_pipelines_automatic_speech_recognition.py::AutomaticSpeechRecognitionPipelineTests::test_pipeline_assisted_generation",
1101
+ "trace": "(line 163) AssertionError: TypeError not raised"
1102
+ },
1103
+ {
1104
+ "line": "tests/pipelines/test_pipelines_automatic_speech_recognition.py::AutomaticSpeechRecognitionPipelineTests::test_simple_s2t",
1105
+ "trace": "(line 1511) OSError: Could not load this library: /usr/local/lib/python3.10/dist-packages/torchaudio/lib/_torchaudio.abi3.so"
1106
+ },
1107
+ {
1108
+ "line": "tests/pipelines/test_pipelines_automatic_speech_recognition.py::AutomaticSpeechRecognitionPipelineTests::test_small_model_pt",
1109
+ "trace": "(line 1511) OSError: Could not load this library: /usr/local/lib/python3.10/dist-packages/torchaudio/lib/_torchaudio.abi3.so"
1110
+ },
1111
+ {
1112
+ "line": "tests/pipelines/test_pipelines_automatic_speech_recognition.py::AutomaticSpeechRecognitionPipelineTests::test_small_model_pt_bf16",
1113
+ "trace": "(line 1511) OSError: Could not load this library: /usr/local/lib/python3.10/dist-packages/torchaudio/lib/_torchaudio.abi3.so"
1114
+ },
1115
+ {
1116
+ "line": "tests/pipelines/test_pipelines_automatic_speech_recognition.py::AutomaticSpeechRecognitionPipelineTests::test_small_model_pt_fp16",
1117
+ "trace": "(line 1511) OSError: Could not load this library: /usr/local/lib/python3.10/dist-packages/torchaudio/lib/_torchaudio.abi3.so"
1118
+ },
1119
+ {
1120
+ "line": "tests/pipelines/test_pipelines_automatic_speech_recognition.py::AutomaticSpeechRecognitionPipelineTests::test_speculative_decoding_whisper_distil",
1121
+ "trace": "(line 326) UnboundLocalError: local variable 'is_updated' referenced before assignment"
1122
+ },
1123
+ {
1124
+ "line": "tests/pipelines/test_pipelines_automatic_speech_recognition.py::AutomaticSpeechRecognitionPipelineTests::test_whisper_longform",
1125
+ "trace": "(line 1478) AssertionError: assert ' Folks, if y...s my segment.' == ' Folks, if y...IC Meanwhile!'"
1126
+ },
1127
+ {
1128
+ "line": "tests/pipelines/test_pipelines_common.py::PipelineUtilsTest::test_bc_torch_device",
1129
+ "trace": "(line 753) ValueError: not enough values to unpack (expected 2, got 1)"
1130
+ },
1131
+ {
1132
+ "line": "tests/pipelines/test_pipelines_common.py::PipelineUtilsTest::test_load_default_pipelines_pt",
1133
+ "trace": "(line 692) RuntimeError: Expected all tensors to be on the same device, but found at least two devices, cuda:0 and cpu!"
1134
+ },
1135
+ {
1136
+ "line": "tests/pipelines/test_pipelines_common.py::PipelineUtilsTest::test_load_default_pipelines_pt_table_qa",
1137
+ "trace": "(line 692) RuntimeError: Expected all tensors to be on the same device, but found at least two devices, cuda:0 and cpu!"
1138
+ },
1139
+ {
1140
+ "line": "tests/pipelines/test_pipelines_common.py::PipelineUtilsTest::test_pipeline_no_device",
1141
+ "trace": "(line 163) AssertionError: ValueError not raised"
1142
+ },
1143
+ {
1144
+ "line": "tests/pipelines/test_pipelines_common.py::CustomPipelineTest::test_cached_pipeline_has_minimum_calls_to_head",
1145
+ "trace": "(line 812) AssertionError: 0 != 1"
1146
+ },
1147
+ {
1148
+ "line": "tests/pipelines/test_pipelines_depth_estimation.py::DepthEstimationPipelineTests::test_large_model_pt",
1149
+ "trace": "(line 136) AssertionError: 2.666 != 2.662"
1150
+ },
1151
+ {
1152
+ "line": "tests/pipelines/test_pipelines_depth_estimation.py::DepthEstimationPipelineTests::test_multiprocess",
1153
+ "trace": "(line 268) ValueError: Could not load model hf-internal-testing/tiny-random-DepthAnythingForDepthEstimation with any of the following classes: (<class 'transformers.models.auto.modeling_auto.AutoModelForDepthEstimation'>, <class 'transformers.models.depth_anything.modeling_depth_anything.DepthAnythingForDepthEstimation'>). See the original errors:"
1154
+ },
1155
+ {
1156
+ "line": "tests/pipelines/test_pipelines_fill_mask.py::FillMaskPipelineTests::test_small_model_pt",
1157
+ "trace": "(line 50) AssertionError: Lists differ: [{'score': 2.2e-05, 'token': 35676, 'token_st[126 chars]LS'}] != [{'sequence': 'My name is Maul', 'score': 2.2[124 chars]LS'}]"
1158
+ },
1159
+ {
1160
+ "line": "tests/pipelines/test_pipelines_image_classification.py::ImageClassificationPipelineTests::test_perceiver",
1161
+ "trace": "(line 216) AssertionError: Lists differ: [{'label': 'tabby, tabby cat', 'score': 0.43[164 chars]072}] != [{'score': 0.4385, 'label': 'tabby, tabby ca[163 chars]or'}]"
1162
+ },
1163
+ {
1164
+ "line": "tests/pipelines/test_pipelines_image_feature_extraction.py::ImageFeatureExtractionPipelineTests::test_image_processing_small_model_pt",
1165
+ "trace": "(line 146) huggingface_hub.errors.StrictDataclassFieldValidationError: Validation error for field 'image_mean':"
1166
+ },
1167
+ {
1168
+ "line": "tests/pipelines/test_pipelines_image_feature_extraction.py::ImageFeatureExtractionPipelineTests::test_small_model_pt",
1169
+ "trace": "(line 53) AssertionError: Lists differ: [-1.417, -0.392, -1.265, -1.196, 1.648, 0.885, 0.56, -0.606, -1[169 chars].981] != [-1.417, -0.392, -1.264, -1.196, 1.648, 0.885, 0.56, -0.606, -1[169 chars].981]"
1170
+ },
1171
+ {
1172
+ "line": "tests/pipelines/test_pipelines_image_segmentation.py::ImageSegmentationPipelineTests::test_integration_torch_image_segmentation",
1173
+ "trace": "(line 4125) UnboundLocalError: local variable 'output' referenced before assignment"
1174
+ },
1175
+ {
1176
+ "line": "tests/pipelines/test_pipelines_image_segmentation.py::ImageSegmentationPipelineTests::test_maskformer",
1177
+ "trace": "(line 4125) UnboundLocalError: local variable 'output' referenced before assignment"
1178
+ },
1179
+ {
1180
+ "line": "tests/pipelines/test_pipelines_image_segmentation.py::ImageSegmentationPipelineTests::test_oneformer",
1181
+ "trace": "(line 4125) UnboundLocalError: local variable 'output' referenced before assignment"
1182
+ },
1183
+ {
1184
+ "line": "tests/pipelines/test_pipelines_image_segmentation.py::ImageSegmentationPipelineTests::test_small_model_pt",
1185
+ "trace": "(line 4125) UnboundLocalError: local variable 'output' referenced before assignment"
1186
+ },
1187
+ {
1188
+ "line": "tests/pipelines/test_pipelines_image_segmentation.py::ImageSegmentationPipelineTests::test_small_model_pt_semantic",
1189
+ "trace": "(line 146) huggingface_hub.errors.StrictDataclassFieldValidationError: Validation error for field 'vocab_size':"
1190
+ },
1191
+ {
1192
+ "line": "tests/pipelines/test_pipelines_image_segmentation.py::ImageSegmentationPipelineTests::test_threshold",
1193
+ "trace": "(line 4125) UnboundLocalError: local variable 'output' referenced before assignment"
1194
+ },
1195
+ {
1196
+ "line": "tests/pipelines/test_pipelines_image_text_to_text.py::ImageTextToTextPipelineTests::test_small_model_pt_token",
1197
+ "trace": "(line 216) AssertionError: Lists differ: [{'in[68 chars]xt': '<image> What this is? Assistant: This is[500 chars]s.'}] != [{'in[68 chars]xt': \"<image> What this is? Assistant: This is[836 chars]r.\"}]"
1198
+ },
1199
+ {
1200
+ "line": "tests/pipelines/test_pipelines_mask_generation.py::MaskGenerationPipelineTests::test_small_model_pt",
1201
+ "trace": "(line 54) AttributeError: 'Tensor' object has no attribute 'tobytes'"
1202
+ },
1203
+ {
1204
+ "line": "tests/pipelines/test_pipelines_mask_generation.py::MaskGenerationPipelineTests::test_threshold",
1205
+ "trace": "(line 54) AttributeError: 'Tensor' object has no attribute 'tobytes'"
1206
+ },
1207
+ {
1208
+ "line": "tests/pipelines/test_pipelines_table_question_answering.py::TQAPipelineTests::test_large_model_pt_tapex",
1209
+ "trace": "(line 724) AttributeError: 'NoneType' object has no attribute 'from_pretrained'"
1210
+ },
1211
+ {
1212
+ "line": "tests/pipelines/test_pipelines_text_generation.py::TextGenerationPipelineTests::test_pipeline_length_setting_warning",
1213
+ "trace": "(line 502) AssertionError: 'Both `max_new_tokens`' unexpectedly found in 'Both `max_new_tokens` (=1) and `max_length`(=20) seem to have been set. `max_new_tokens` will take precedence. Please refer to the documentation for more information. (https://huggingface.co/docs/transformers/main/en/main_classes/text_generation)\\n'"
1214
+ },
1215
+ {
1216
+ "line": "tests/pipelines/test_pipelines_text_to_audio.py::TextToAudioPipelineTests::test_csm_model_pt",
1217
+ "trace": "(line 146) huggingface_hub.errors.StrictDataclassFieldValidationError: Validation error for field 'use_cache':"
1218
+ },
1219
+ {
1220
+ "line": "tests/pipelines/test_pipelines_text_to_audio.py::TextToAudioPipelineTests::test_dia_model",
1221
+ "trace": "(line 146) huggingface_hub.errors.StrictDataclassFieldValidationError: Validation error for field 'use_cache':"
1222
+ },
1223
+ {
1224
+ "line": "tests/pipelines/test_pipelines_text_to_audio.py::TextToAudioPipelineTests::test_forward_model_kwargs",
1225
+ "trace": "(line 163) AssertionError: TypeError not raised"
1226
+ },
1227
+ {
1228
+ "line": "tests/pipelines/test_pipelines_text_to_audio.py::TextToAudioPipelineTests::test_medium_seamless_m4t_pt",
1229
+ "trace": "(line 146) huggingface_hub.errors.StrictDataclassFieldValidationError: Validation error for field 'use_cache':"
1230
+ },
1231
+ {
1232
+ "line": "tests/pipelines/test_pipelines_token_classification.py::TokenClassificationPipelineTests::test_no_offset_tokenizer",
1233
+ "trace": "(line 4125) UnboundLocalError: local variable 'output' referenced before assignment"
1234
+ },
1235
+ {
1236
+ "line": "tests/pipelines/test_pipelines_token_classification.py::TokenClassificationPipelineTests::test_pt_ignore_subwords_slow_tokenizer_raises",
1237
+ "trace": "(line 163) AssertionError: ValueError not raised"
1238
+ },
1239
+ {
1240
+ "line": "tests/pipelines/test_pipelines_zero_shot_image_classification.py::ZeroShotImageClassificationPipelineTests::test_large_model_pt",
1241
+ "trace": "(line 151) AssertionError: Lists differ: [{'score': 0.512, 'label': 'remote'}, {'score': 0.484, '[46 chars]ne'}] != [{'score': 0.511, 'label': 'remote'}, {'score': 0.485, '[46 chars]ne'}]"
1242
+ },
1243
+ {
1244
+ "line": "tests/pipelines/test_pipelines_zero_shot_object_detection.py::ZeroShotObjectDetectionPipelineTests::test_large_model_pt",
1245
+ "trace": "(line 155) AssertionError: Lists differ: [{'score': 0.2869, 'label': 'cat', 'box': {'xmin': 325, '[415 chars]77}}] != [{'score': 0.2868, 'label': 'cat', 'box': {'xmin': 324, '[414 chars]76}}]"
1246
+ },
1247
+ {
1248
+ "line": "tests/pipelines/test_pipelines_zero_shot_object_detection.py::ZeroShotObjectDetectionPipelineTests::test_small_model_pt",
1249
+ "trace": "(line 104) AssertionError: Lists differ: [{'score': 0.7236, 'label': 'cat', 'box': {'xmin': 204, '[813 chars]27}}] != [{'score': 0.7235, 'label': 'cat', 'box': {'xmin': 204, '[812 chars]27}}]"
1250
+ },
1251
+ {
1252
+ "line": "tests/pipelines/test_pipelines_zero_shot_object_detection.py::ZeroShotObjectDetectionPipelineTests::test_threshold",
1253
+ "trace": "(line 209) AssertionError: Lists differ: [{'score': 0.2869, 'label': 'cat', 'box': {'xmin': 325, '[223 chars]72}}] != [{'score': 0.2868, 'label': 'cat', 'box': {'xmin': 324, '[222 chars]72}}]"
1254
+ },
1255
+ {
1256
+ "line": "tests/pipelines/test_pipelines_zero_shot_object_detection.py::ZeroShotObjectDetectionPipelineTests::test_top_k",
1257
+ "trace": "(line 229) AssertionError: Lists differ: [{'score': 0.2869, 'label': 'cat', 'box': {'xmin': 325, '[130 chars]15}}] != [{'score': 0.2868, 'label': 'cat', 'box': {'xmin': 324, '[129 chars]15}}]"
1258
+ }
1259
+ ]
1260
+ },
1261
+ "job_link": {
1262
+ "multi": "https://github.com/huggingface/transformers/actions/runs/23973914236/job/69928372457"
1263
+ }
1264
+ }
1265
+ }