hf-transformers-bot commited on
Commit
af12b8d
·
verified ·
1 Parent(s): be630ba

Upload 2026-01-29/ci_results_run_models_gpu/new_failures_with_bad_commit_grouped_by_authors.json with huggingface_hub

Browse files
2026-01-29/ci_results_run_models_gpu/new_failures_with_bad_commit_grouped_by_authors.json ADDED
@@ -0,0 +1,313 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "remi-or": {
3
+ "cli": {
4
+ "single-gpu": [
5
+ {
6
+ "test": "tests/cli/test_serve.py::ServeCompletionsContinuousBatchingIntegrationTest::test_generation_config_in_request",
7
+ "commit": "b207f38e0d754d6addda307c187d4478a44b276a",
8
+ "status": "git bisect found the bad commit.",
9
+ "pr_number": 43521,
10
+ "author": "remi-or",
11
+ "merged_by": "ArthurZucker",
12
+ "parent": "be87564082c7d0fc1c6508652b4490ecc89ad9ff",
13
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/21464126403/job/61822958473"
14
+ },
15
+ {
16
+ "test": "tests/cli/test_serve.py::ServeCompletionsContinuousBatchingIntegrationTest::test_requests_1_one_token",
17
+ "commit": "b207f38e0d754d6addda307c187d4478a44b276a",
18
+ "status": "git bisect found the bad commit.",
19
+ "pr_number": 43521,
20
+ "author": "remi-or",
21
+ "merged_by": "ArthurZucker",
22
+ "parent": "be87564082c7d0fc1c6508652b4490ecc89ad9ff",
23
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/21464126403/job/61822958473"
24
+ }
25
+ ]
26
+ }
27
+ },
28
+ "null": {
29
+ "auto": {
30
+ "single-gpu": [
31
+ {
32
+ "test": "tests/models/auto/test_tokenization_auto.py::AutoTokenizerTest::test_from_pretrained_use_fast_toggle",
33
+ "commit": null,
34
+ "status": "flaky: test fails on the current CI run (commit: e46d21afe2aaf942fd9515b2fba1d6f3affd61a6) but passes during the check.",
35
+ "pr_number": null,
36
+ "author": null,
37
+ "merged_by": null,
38
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/21464126403/job/61822953534"
39
+ }
40
+ ]
41
+ },
42
+ "aya_vision": {
43
+ "single-gpu": [
44
+ {
45
+ "test": "tests/models/aya_vision/test_modeling_aya_vision.py::AyaVisionIntegrationTest::test_small_model_integration_batched_generate_multi_image",
46
+ "commit": null,
47
+ "status": "flaky: test fails on the current CI run (commit: e46d21afe2aaf942fd9515b2fba1d6f3affd61a6) but passes during the check.",
48
+ "pr_number": null,
49
+ "author": null,
50
+ "merged_by": null,
51
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/21464126403/job/61822953492"
52
+ }
53
+ ]
54
+ },
55
+ "blip_2": {
56
+ "single-gpu": [
57
+ {
58
+ "test": "tests/models/blip_2/test_modeling_blip_2.py::Blip2ModelIntegrationTest::test_inference_itm_fp16",
59
+ "commit": null,
60
+ "status": "flaky: test fails on the current CI run (commit: e46d21afe2aaf942fd9515b2fba1d6f3affd61a6) but passes during the check.",
61
+ "pr_number": null,
62
+ "author": null,
63
+ "merged_by": null,
64
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/21464126403/job/61822954063"
65
+ }
66
+ ]
67
+ },
68
+ "electra": {
69
+ "single-gpu": [
70
+ {
71
+ "test": "tests/models/electra/test_modeling_electra.py::ElectraModelTest::test_sdpa_padding_matches_padding_free_with_position_ids",
72
+ "commit": null,
73
+ "status": "flaky: test passed in the previous run (commit: be0115e42fbf5d7ffe915e01b0cf9a68b29958c6) but failed (on the same commit) during the check of the current run.",
74
+ "pr_number": null,
75
+ "author": null,
76
+ "merged_by": null,
77
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/21464126403/job/61822955835"
78
+ }
79
+ ]
80
+ },
81
+ "exaone4": {
82
+ "single-gpu": [
83
+ {
84
+ "test": "tests/models/exaone4/test_modeling_exaone4.py::Exaone4IntegrationTest::test_export_static_cache",
85
+ "commit": null,
86
+ "status": "flaky: test fails on the current CI run (commit: e46d21afe2aaf942fd9515b2fba1d6f3affd61a6) but passes during the check.",
87
+ "pr_number": null,
88
+ "author": null,
89
+ "merged_by": null,
90
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/21464126403/job/61822956151"
91
+ }
92
+ ]
93
+ },
94
+ "flex_olmo": {
95
+ "single-gpu": [
96
+ {
97
+ "test": "tests/models/flex_olmo/test_modeling_flex_olmo.py::FlexOlmoModelTest::test_eager_padding_matches_padding_free_with_position_ids",
98
+ "commit": null,
99
+ "status": "flaky: test fails on the current CI run (commit: e46d21afe2aaf942fd9515b2fba1d6f3affd61a6) but passes during the check.",
100
+ "pr_number": null,
101
+ "author": null,
102
+ "merged_by": null,
103
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/21464126403/job/61822956355"
104
+ }
105
+ ]
106
+ },
107
+ "gemma3n": {
108
+ "single-gpu": [
109
+ {
110
+ "test": "tests/models/gemma3n/test_modeling_gemma3n.py::Gemma3nIntegrationTest::test_generation_beyond_sliding_window_0_sdpa",
111
+ "commit": null,
112
+ "status": "flaky: test passed in the previous run (commit: be0115e42fbf5d7ffe915e01b0cf9a68b29958c6) but failed (on the same commit) during the check of the current run.",
113
+ "pr_number": null,
114
+ "author": null,
115
+ "merged_by": null,
116
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/21464126403/job/61822956591"
117
+ }
118
+ ]
119
+ },
120
+ "internvl": {
121
+ "single-gpu": [
122
+ {
123
+ "test": "tests/models/internvl/test_modeling_internvl.py::InternVLLlamaIntegrationTest::test_llama_small_model_integration_forward",
124
+ "commit": null,
125
+ "status": "flaky: test fails on the current CI run (commit: e46d21afe2aaf942fd9515b2fba1d6f3affd61a6) but passes during the check.",
126
+ "pr_number": null,
127
+ "author": null,
128
+ "merged_by": null,
129
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/21464126403/job/61822957820"
130
+ }
131
+ ]
132
+ },
133
+ "luke": {
134
+ "single-gpu": [
135
+ {
136
+ "test": "tests/models/luke/test_tokenization_luke.py::LukeTokenizerIntegrationTests::test_text_pair_padding_pytorch_tensors",
137
+ "commit": null,
138
+ "status": "flaky: test fails on the current CI run (commit: e46d21afe2aaf942fd9515b2fba1d6f3affd61a6) but passes during the check.",
139
+ "pr_number": null,
140
+ "author": null,
141
+ "merged_by": null,
142
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/21464126403/job/61822958603"
143
+ }
144
+ ]
145
+ },
146
+ "m2m_100": {
147
+ "single-gpu": [
148
+ {
149
+ "test": "tests/models/m2m_100/test_tokenization_m2m_100.py::M2M100TokenizationTest::test_truncation_side_in_kwargs",
150
+ "commit": null,
151
+ "status": "flaky: test fails on the current CI run (commit: e46d21afe2aaf942fd9515b2fba1d6f3affd61a6) but passes during the check.",
152
+ "pr_number": null,
153
+ "author": null,
154
+ "merged_by": null,
155
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/21464126403/job/61822958655"
156
+ }
157
+ ]
158
+ },
159
+ "markuplm": {
160
+ "single-gpu": [
161
+ {
162
+ "test": "tests/models/markuplm/test_processing_markuplm.py::MarkupLMProcessorIntegrationTests::test_processor_case_1",
163
+ "commit": null,
164
+ "status": "flaky: test fails on the current CI run (commit: e46d21afe2aaf942fd9515b2fba1d6f3affd61a6) but passes during the check.",
165
+ "pr_number": null,
166
+ "author": null,
167
+ "merged_by": null,
168
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/21464126403/job/61822958769"
169
+ }
170
+ ]
171
+ },
172
+ "mpnet": {
173
+ "single-gpu": [
174
+ {
175
+ "test": "tests/models/mpnet/test_tokenization_mpnet.py::MPNetTokenizerTest::test_integration_from_extractor",
176
+ "commit": null,
177
+ "status": "flaky: test fails on the current CI run (commit: e46d21afe2aaf942fd9515b2fba1d6f3affd61a6) but passes during the check.",
178
+ "pr_number": null,
179
+ "author": null,
180
+ "merged_by": null,
181
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/21464126403/job/61822953753"
182
+ }
183
+ ]
184
+ },
185
+ "openai": {
186
+ "single-gpu": [
187
+ {
188
+ "test": "tests/models/openai/test_tokenization_openai.py::OpenAIGPTTokenizationTest::test_integration_from_extractor",
189
+ "commit": null,
190
+ "status": "flaky: test fails on the current CI run (commit: e46d21afe2aaf942fd9515b2fba1d6f3affd61a6) but passes during the check.",
191
+ "pr_number": null,
192
+ "author": null,
193
+ "merged_by": null,
194
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/21464126403/job/61822954364"
195
+ },
196
+ {
197
+ "test": "tests/models/openai/test_tokenization_openai.py::OpenAIGPTTokenizationTest::test_internal_consistency",
198
+ "commit": null,
199
+ "status": "flaky: test fails on the current CI run (commit: e46d21afe2aaf942fd9515b2fba1d6f3affd61a6) but passes during the check.",
200
+ "pr_number": null,
201
+ "author": null,
202
+ "merged_by": null,
203
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/21464126403/job/61822954364"
204
+ },
205
+ {
206
+ "test": "tests/models/openai/test_tokenization_openai.py::OpenAIGPTTokenizationTest::test_maximum_encoding_length_pair_input",
207
+ "commit": null,
208
+ "status": "flaky: test fails on the current CI run (commit: e46d21afe2aaf942fd9515b2fba1d6f3affd61a6) but passes during the check.",
209
+ "pr_number": null,
210
+ "author": null,
211
+ "merged_by": null,
212
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/21464126403/job/61822954364"
213
+ },
214
+ {
215
+ "test": "tests/models/openai/test_tokenization_openai.py::OpenAIGPTTokenizationTest::test_maximum_encoding_length_single_input",
216
+ "commit": null,
217
+ "status": "flaky: test fails on the current CI run (commit: e46d21afe2aaf942fd9515b2fba1d6f3affd61a6) but passes during the check.",
218
+ "pr_number": null,
219
+ "author": null,
220
+ "merged_by": null,
221
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/21464126403/job/61822954364"
222
+ }
223
+ ]
224
+ },
225
+ "qwen2_audio": {
226
+ "single-gpu": [
227
+ {
228
+ "test": "tests/models/qwen2_audio/test_modeling_qwen2_audio.py::Qwen2AudioForConditionalGenerationIntegrationTest::test_small_model_integration_test_multiurn",
229
+ "commit": null,
230
+ "status": "flaky: test fails on the current CI run (commit: e46d21afe2aaf942fd9515b2fba1d6f3affd61a6) but passes during the check.",
231
+ "pr_number": null,
232
+ "author": null,
233
+ "merged_by": null,
234
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/21464126403/job/61822955378"
235
+ }
236
+ ]
237
+ },
238
+ "reformer": {
239
+ "single-gpu": [
240
+ {
241
+ "test": "tests/models/reformer/test_tokenization_reformer.py::ReformerTokenizationTest::test_chat_template_return_assistant_tokens_mask_truncated",
242
+ "commit": null,
243
+ "status": "flaky: test fails on the current CI run (commit: e46d21afe2aaf942fd9515b2fba1d6f3affd61a6) but passes during the check.",
244
+ "pr_number": null,
245
+ "author": null,
246
+ "merged_by": null,
247
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/21464126403/job/61822955606"
248
+ }
249
+ ]
250
+ },
251
+ "t5": {
252
+ "single-gpu": [
253
+ {
254
+ "test": "tests/models/t5/test_tokenization_t5.py::T5TokenizationTest::test_chat_template_return_assistant_tokens_mask",
255
+ "commit": null,
256
+ "status": "flaky: test fails on the current CI run (commit: e46d21afe2aaf942fd9515b2fba1d6f3affd61a6) but passes during the check.",
257
+ "pr_number": null,
258
+ "author": null,
259
+ "merged_by": null,
260
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/21464126403/job/61822956905"
261
+ }
262
+ ]
263
+ },
264
+ "unispeech": {
265
+ "single-gpu": [
266
+ {
267
+ "test": "tests/models/unispeech/test_modeling_unispeech.py::UniSpeechRobustModelTest::test_batched_inference",
268
+ "commit": null,
269
+ "status": "flaky: test passed in the previous run (commit: be0115e42fbf5d7ffe915e01b0cf9a68b29958c6) but failed (on the same commit) during the check of the current run.",
270
+ "pr_number": null,
271
+ "author": null,
272
+ "merged_by": null,
273
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/21464126403/job/61822957279"
274
+ }
275
+ ]
276
+ },
277
+ "whisper": {
278
+ "single-gpu": [
279
+ {
280
+ "test": "tests/models/whisper/test_modeling_whisper.py::WhisperModelIntegrationTests::test_tiny_static_generation",
281
+ "commit": null,
282
+ "status": "flaky: test fails on the current CI run (commit: e46d21afe2aaf942fd9515b2fba1d6f3affd61a6) but passes during the check.",
283
+ "pr_number": null,
284
+ "author": null,
285
+ "merged_by": null,
286
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/21464126403/job/61822958102"
287
+ },
288
+ {
289
+ "test": "tests/models/whisper/test_tokenization_whisper.py::WhisperTokenizerTest::test_decode_asr_with_word_level_timestamps",
290
+ "commit": null,
291
+ "status": "flaky: test fails on the current CI run (commit: e46d21afe2aaf942fd9515b2fba1d6f3affd61a6) but passes during the check.",
292
+ "pr_number": null,
293
+ "author": null,
294
+ "merged_by": null,
295
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/21464126403/job/61822958102"
296
+ }
297
+ ]
298
+ },
299
+ "pipelines": {
300
+ "single-gpu": [
301
+ {
302
+ "test": "tests/pipelines/test_pipelines_automatic_speech_recognition.py::AutomaticSpeechRecognitionPipelineTests::test_torch_large_with_input_features",
303
+ "commit": null,
304
+ "status": "flaky: test fails on the current CI run (commit: e46d21afe2aaf942fd9515b2fba1d6f3affd61a6) but passes during the check.",
305
+ "pr_number": null,
306
+ "author": null,
307
+ "merged_by": null,
308
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/21464126403/job/61822958740"
309
+ }
310
+ ]
311
+ }
312
+ }
313
+ }