hf-transformers-bot commited on
Commit
5497be8
·
verified ·
1 Parent(s): b304b22

Upload 2025-11-13/ci_results_run_models_gpu/new_failures_with_bad_commit_grouped_by_authors.json with huggingface_hub

Browse files
2025-11-13/ci_results_run_models_gpu/new_failures_with_bad_commit_grouped_by_authors.json ADDED
@@ -0,0 +1,1540 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "ydshieh": {
3
+ "audioflamingo3": {
4
+ "single-gpu": [
5
+ {
6
+ "test": "tests/models/audioflamingo3/test_modeling_audioflamingo3.py::AudioFlamingo3ForConditionalGenerationModelTest::test_eager_matches_fa2_generate",
7
+ "commit": "1709ed96e47f36fe926e2cd2556fa839b41c2afd",
8
+ "status": "git bisect found the bad commit.",
9
+ "pr_number": 40290,
10
+ "author": "lashahub",
11
+ "merged_by": "ydshieh",
12
+ "parent": "fd36275be2f3e56bc20da01f1f320b623b413957",
13
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256262527"
14
+ }
15
+ ]
16
+ }
17
+ },
18
+ "null": {
19
+ "bamba": {
20
+ "single-gpu": [
21
+ {
22
+ "test": "tests/models/bamba/test_modeling_bamba.py::BambaModelTest::test_flash_attn_2_inference_equivalence",
23
+ "commit": null,
24
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
25
+ "pr_number": null,
26
+ "author": null,
27
+ "merged_by": null,
28
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256262537"
29
+ },
30
+ {
31
+ "test": "tests/models/bamba/test_modeling_bamba.py::BambaModelTest::test_flash_attn_2_inference_equivalence_right_padding",
32
+ "commit": null,
33
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
34
+ "pr_number": null,
35
+ "author": null,
36
+ "merged_by": null,
37
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256262537"
38
+ }
39
+ ]
40
+ },
41
+ "bark": {
42
+ "single-gpu": [
43
+ {
44
+ "test": "tests/models/bark/test_modeling_bark.py::BarkSemanticModelTest::test_eager_matches_fa2_generate",
45
+ "commit": null,
46
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
47
+ "pr_number": null,
48
+ "author": null,
49
+ "merged_by": null,
50
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256262543"
51
+ },
52
+ {
53
+ "test": "tests/models/bark/test_modeling_bark.py::BarkSemanticModelTest::test_flash_attention_2_continue_generate_with_position_ids",
54
+ "commit": null,
55
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
56
+ "pr_number": null,
57
+ "author": null,
58
+ "merged_by": null,
59
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256262543"
60
+ },
61
+ {
62
+ "test": "tests/models/bark/test_modeling_bark.py::BarkCoarseModelTest::test_eager_matches_fa2_generate",
63
+ "commit": null,
64
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
65
+ "pr_number": null,
66
+ "author": null,
67
+ "merged_by": null,
68
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256262543"
69
+ },
70
+ {
71
+ "test": "tests/models/bark/test_modeling_bark.py::BarkCoarseModelTest::test_flash_attention_2_continue_generate_with_position_ids",
72
+ "commit": null,
73
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
74
+ "pr_number": null,
75
+ "author": null,
76
+ "merged_by": null,
77
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256262543"
78
+ }
79
+ ]
80
+ },
81
+ "blt": {
82
+ "single-gpu": [
83
+ {
84
+ "test": "tests/models/blt/test_modeling_blt.py::BltModelTest::test_flash_attn_2_can_compile_with_attention_mask_None_without_graph_break",
85
+ "commit": null,
86
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
87
+ "pr_number": null,
88
+ "author": null,
89
+ "merged_by": null,
90
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256263058"
91
+ }
92
+ ]
93
+ },
94
+ "chameleon": {
95
+ "single-gpu": [
96
+ {
97
+ "test": "tests/models/chameleon/test_modeling_chameleon.py::ChameleonModelTest::test_flash_attention_2_padding_matches_padding_free_with_position_ids",
98
+ "commit": null,
99
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
100
+ "pr_number": null,
101
+ "author": null,
102
+ "merged_by": null,
103
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256263264"
104
+ },
105
+ {
106
+ "test": "tests/models/chameleon/test_modeling_chameleon.py::ChameleonVision2SeqModelTest::test_flash_attention_2_padding_matches_padding_free_with_position_ids_and_fa_kwargs",
107
+ "commit": null,
108
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
109
+ "pr_number": null,
110
+ "author": null,
111
+ "merged_by": null,
112
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256263264"
113
+ }
114
+ ]
115
+ },
116
+ "clap": {
117
+ "single-gpu": [
118
+ {
119
+ "test": "tests/models/clap/test_modeling_clap.py::ClapAudioModelTest::test_flash_attn_2_inference_equivalence",
120
+ "commit": null,
121
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
122
+ "pr_number": null,
123
+ "author": null,
124
+ "merged_by": null,
125
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256263271"
126
+ },
127
+ {
128
+ "test": "tests/models/clap/test_modeling_clap.py::ClapAudioModelTest::test_flash_attn_2_inference_equivalence_right_padding",
129
+ "commit": null,
130
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
131
+ "pr_number": null,
132
+ "author": null,
133
+ "merged_by": null,
134
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256263271"
135
+ },
136
+ {
137
+ "test": "tests/models/clap/test_modeling_clap.py::ClapModelTest::test_flash_attn_2_inference_equivalence",
138
+ "commit": null,
139
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
140
+ "pr_number": null,
141
+ "author": null,
142
+ "merged_by": null,
143
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256263271"
144
+ },
145
+ {
146
+ "test": "tests/models/clap/test_modeling_clap.py::ClapModelTest::test_flash_attn_2_inference_equivalence_right_padding",
147
+ "commit": null,
148
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
149
+ "pr_number": null,
150
+ "author": null,
151
+ "merged_by": null,
152
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256263271"
153
+ }
154
+ ]
155
+ },
156
+ "colpali": {
157
+ "single-gpu": [
158
+ {
159
+ "test": "tests/models/colpali/test_modeling_colpali.py::ColPaliForRetrievalModelTest::test_flash_attn_2_inference_equivalence",
160
+ "commit": null,
161
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
162
+ "pr_number": null,
163
+ "author": null,
164
+ "merged_by": null,
165
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256263606"
166
+ },
167
+ {
168
+ "test": "tests/models/colpali/test_modeling_colpali.py::ColPaliForRetrievalModelTest::test_flash_attn_2_inference_equivalence_right_padding",
169
+ "commit": null,
170
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
171
+ "pr_number": null,
172
+ "author": null,
173
+ "merged_by": null,
174
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256263606"
175
+ }
176
+ ]
177
+ },
178
+ "deepseek_v2": {
179
+ "single-gpu": [
180
+ {
181
+ "test": "tests/models/deepseek_v2/test_modeling_deepseek_v2.py::DeepseekV2ModelTest::test_flash_attention_2_padding_matches_padding_free_with_position_ids",
182
+ "commit": null,
183
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
184
+ "pr_number": null,
185
+ "author": null,
186
+ "merged_by": null,
187
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256263996"
188
+ },
189
+ {
190
+ "test": "tests/models/deepseek_v2/test_modeling_deepseek_v2.py::DeepseekV2ModelTest::test_flash_attn_2_can_compile_with_attention_mask_None_without_graph_break",
191
+ "commit": null,
192
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
193
+ "pr_number": null,
194
+ "author": null,
195
+ "merged_by": null,
196
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256263996"
197
+ },
198
+ {
199
+ "test": "tests/models/deepseek_v2/test_modeling_deepseek_v2.py::DeepseekV2ModelTest::test_flash_attn_2_fp32_ln",
200
+ "commit": null,
201
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
202
+ "pr_number": null,
203
+ "author": null,
204
+ "merged_by": null,
205
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256263996"
206
+ }
207
+ ]
208
+ },
209
+ "deepseek_vl_hybrid": {
210
+ "single-gpu": [
211
+ {
212
+ "test": "tests/models/deepseek_vl_hybrid/test_modeling_deepseek_vl_hybrid.py::DeepseekVLHybridModelTest::test_flash_attention_2_continue_generate_with_position_ids",
213
+ "commit": null,
214
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
215
+ "pr_number": null,
216
+ "author": null,
217
+ "merged_by": null,
218
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256264025"
219
+ },
220
+ {
221
+ "test": "tests/models/deepseek_vl_hybrid/test_modeling_deepseek_vl_hybrid.py::DeepseekVLHybridModelTest::test_flash_attn_2_fp32_ln",
222
+ "commit": null,
223
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
224
+ "pr_number": null,
225
+ "author": null,
226
+ "merged_by": null,
227
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256264025"
228
+ },
229
+ {
230
+ "test": "tests/models/deepseek_vl_hybrid/test_modeling_deepseek_vl_hybrid.py::DeepseekVLHybridModelTest::test_flash_attn_2_from_config",
231
+ "commit": null,
232
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
233
+ "pr_number": null,
234
+ "author": null,
235
+ "merged_by": null,
236
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256264025"
237
+ }
238
+ ]
239
+ },
240
+ "diffllama": {
241
+ "single-gpu": [
242
+ {
243
+ "test": "tests/models/diffllama/test_modeling_diffllama.py::DiffLlamaModelTest::test_flash_attn_2_generate_padding_right",
244
+ "commit": null,
245
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
246
+ "pr_number": null,
247
+ "author": null,
248
+ "merged_by": null,
249
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256264149"
250
+ },
251
+ {
252
+ "test": "tests/models/diffllama/test_modeling_diffllama.py::DiffLlamaModelTest::test_flash_attn_2_inference_equivalence",
253
+ "commit": null,
254
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
255
+ "pr_number": null,
256
+ "author": null,
257
+ "merged_by": null,
258
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256264149"
259
+ },
260
+ {
261
+ "test": "tests/models/diffllama/test_modeling_diffllama.py::DiffLlamaModelTest::test_flash_attn_2_inference_equivalence_right_padding",
262
+ "commit": null,
263
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
264
+ "pr_number": null,
265
+ "author": null,
266
+ "merged_by": null,
267
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256264149"
268
+ }
269
+ ]
270
+ },
271
+ "donut": {
272
+ "single-gpu": [
273
+ {
274
+ "test": "tests/models/donut/test_modeling_donut_swin.py::DonutSwinModelTest::test_flash_attn_2_inference_equivalence",
275
+ "commit": null,
276
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
277
+ "pr_number": null,
278
+ "author": null,
279
+ "merged_by": null,
280
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256264372"
281
+ },
282
+ {
283
+ "test": "tests/models/donut/test_modeling_donut_swin.py::DonutSwinModelTest::test_flash_attn_2_inference_equivalence_right_padding",
284
+ "commit": null,
285
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
286
+ "pr_number": null,
287
+ "author": null,
288
+ "merged_by": null,
289
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256264372"
290
+ }
291
+ ]
292
+ },
293
+ "efficientloftr": {
294
+ "single-gpu": [
295
+ {
296
+ "test": "tests/models/efficientloftr/test_modeling_efficientloftr.py::EfficientLoFTRModelTest::test_flash_attn_2_inference_equivalence",
297
+ "commit": null,
298
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
299
+ "pr_number": null,
300
+ "author": null,
301
+ "merged_by": null,
302
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256264426"
303
+ },
304
+ {
305
+ "test": "tests/models/efficientloftr/test_modeling_efficientloftr.py::EfficientLoFTRModelTest::test_flash_attn_2_inference_equivalence_right_padding",
306
+ "commit": null,
307
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
308
+ "pr_number": null,
309
+ "author": null,
310
+ "merged_by": null,
311
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256264426"
312
+ }
313
+ ]
314
+ },
315
+ "emu3": {
316
+ "single-gpu": [
317
+ {
318
+ "test": "tests/models/emu3/test_modeling_emu3.py::Emu3Vision2TextModelTest::test_flash_attn_2_inference_equivalence",
319
+ "commit": null,
320
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
321
+ "pr_number": null,
322
+ "author": null,
323
+ "merged_by": null,
324
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256264594"
325
+ },
326
+ {
327
+ "test": "tests/models/emu3/test_modeling_emu3.py::Emu3Vision2TextModelTest::test_flash_attn_2_inference_equivalence_right_padding",
328
+ "commit": null,
329
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
330
+ "pr_number": null,
331
+ "author": null,
332
+ "merged_by": null,
333
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256264594"
334
+ }
335
+ ]
336
+ },
337
+ "exaone4": {
338
+ "single-gpu": [
339
+ {
340
+ "test": "tests/models/exaone4/test_modeling_exaone4.py::Exaone4ModelTest::test_flash_attention_2_padding_matches_padding_free_with_position_ids",
341
+ "commit": null,
342
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
343
+ "pr_number": null,
344
+ "author": null,
345
+ "merged_by": null,
346
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256264724"
347
+ }
348
+ ]
349
+ },
350
+ "falcon": {
351
+ "single-gpu": [
352
+ {
353
+ "test": "tests/models/falcon/test_modeling_falcon.py::FalconModelTest::test_flash_attn_2_inference_equivalence",
354
+ "commit": null,
355
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
356
+ "pr_number": null,
357
+ "author": null,
358
+ "merged_by": null,
359
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256264803"
360
+ },
361
+ {
362
+ "test": "tests/models/falcon/test_modeling_falcon.py::FalconModelTest::test_flash_attn_2_inference_equivalence_right_padding",
363
+ "commit": null,
364
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
365
+ "pr_number": null,
366
+ "author": null,
367
+ "merged_by": null,
368
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256264803"
369
+ }
370
+ ]
371
+ },
372
+ "flex_olmo": {
373
+ "single-gpu": [
374
+ {
375
+ "test": "tests/models/flex_olmo/test_modeling_flex_olmo.py::FlexOlmoModelTest::test_flash_attention_2_padding_matches_padding_free_with_position_ids",
376
+ "commit": null,
377
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
378
+ "pr_number": null,
379
+ "author": null,
380
+ "merged_by": null,
381
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256264910"
382
+ },
383
+ {
384
+ "test": "tests/models/flex_olmo/test_modeling_flex_olmo.py::FlexOlmoModelTest::test_flash_attention_2_padding_matches_padding_free_with_position_ids_and_fa_kwargs",
385
+ "commit": null,
386
+ "status": "flaky: test fails on the current CI run (commit: ffb35fe14283d61cd5434a32d04d07993e66477a) but passes during the check.",
387
+ "pr_number": null,
388
+ "author": null,
389
+ "merged_by": null,
390
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256264910"
391
+ },
392
+ {
393
+ "test": "tests/models/flex_olmo/test_modeling_flex_olmo.py::FlexOlmoModelTest::test_flash_attn_2_can_compile_with_attention_mask_None_without_graph_break",
394
+ "commit": null,
395
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
396
+ "pr_number": null,
397
+ "author": null,
398
+ "merged_by": null,
399
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256264910"
400
+ }
401
+ ]
402
+ },
403
+ "gemma3n": {
404
+ "single-gpu": [
405
+ {
406
+ "test": "tests/models/gemma3n/test_modeling_gemma3n.py::Gemma3nTextModelTest::test_flash_attn_2_equivalence",
407
+ "commit": null,
408
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
409
+ "pr_number": null,
410
+ "author": null,
411
+ "merged_by": null,
412
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256265090"
413
+ },
414
+ {
415
+ "test": "tests/models/gemma3n/test_modeling_gemma3n.py::Gemma3nTextModelTest::test_flash_attn_2_inference_equivalence",
416
+ "commit": null,
417
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
418
+ "pr_number": null,
419
+ "author": null,
420
+ "merged_by": null,
421
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256265090"
422
+ },
423
+ {
424
+ "test": "tests/models/gemma3n/test_modeling_gemma3n.py::Gemma3nTextModelTest::test_flash_attn_2_inference_equivalence_right_padding",
425
+ "commit": null,
426
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
427
+ "pr_number": null,
428
+ "author": null,
429
+ "merged_by": null,
430
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256265090"
431
+ }
432
+ ]
433
+ },
434
+ "glm4_moe": {
435
+ "single-gpu": [
436
+ {
437
+ "test": "tests/models/glm4_moe/test_modeling_glm4_moe.py::Glm4MoeModelTest::test_flash_attn_2_can_compile_with_attention_mask_None_without_graph_break",
438
+ "commit": null,
439
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
440
+ "pr_number": null,
441
+ "author": null,
442
+ "merged_by": null,
443
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256265226"
444
+ }
445
+ ]
446
+ },
447
+ "gpt_neo": {
448
+ "single-gpu": [
449
+ {
450
+ "test": "tests/models/gpt_neo/test_modeling_gpt_neo.py::GPTNeoModelTest::test_eager_matches_fa2_generate",
451
+ "commit": null,
452
+ "status": "flaky: test fails on the current CI run (commit: ffb35fe14283d61cd5434a32d04d07993e66477a) but passes during the check.",
453
+ "pr_number": null,
454
+ "author": null,
455
+ "merged_by": null,
456
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256265344"
457
+ }
458
+ ]
459
+ },
460
+ "gpt_oss": {
461
+ "single-gpu": [
462
+ {
463
+ "test": "tests/models/gpt_oss/test_modeling_gpt_oss.py::GptOssModelTest::test_flash_attn_2_inference_equivalence",
464
+ "commit": null,
465
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
466
+ "pr_number": null,
467
+ "author": null,
468
+ "merged_by": null,
469
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256265453"
470
+ },
471
+ {
472
+ "test": "tests/models/gpt_oss/test_modeling_gpt_oss.py::GptOssModelTest::test_flash_attn_2_inference_equivalence_right_padding",
473
+ "commit": null,
474
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
475
+ "pr_number": null,
476
+ "author": null,
477
+ "merged_by": null,
478
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256265453"
479
+ }
480
+ ]
481
+ },
482
+ "granitemoe": {
483
+ "single-gpu": [
484
+ {
485
+ "test": "tests/models/granitemoe/test_modeling_granitemoe.py::GraniteMoeModelTest::test_flash_attention_2_padding_matches_padding_free_with_position_ids_and_fa_kwargs",
486
+ "commit": null,
487
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
488
+ "pr_number": null,
489
+ "author": null,
490
+ "merged_by": null,
491
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256265505"
492
+ }
493
+ ]
494
+ },
495
+ "granitemoehybrid": {
496
+ "single-gpu": [
497
+ {
498
+ "test": "tests/models/granitemoehybrid/test_modeling_granitemoehybrid.py::BambaModelTest::test_flash_attn_2_inference_equivalence",
499
+ "commit": null,
500
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
501
+ "pr_number": null,
502
+ "author": null,
503
+ "merged_by": null,
504
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256265535"
505
+ },
506
+ {
507
+ "test": "tests/models/granitemoehybrid/test_modeling_granitemoehybrid.py::BambaModelTest::test_flash_attn_2_inference_equivalence_right_padding",
508
+ "commit": null,
509
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
510
+ "pr_number": null,
511
+ "author": null,
512
+ "merged_by": null,
513
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256265535"
514
+ },
515
+ {
516
+ "test": "tests/models/granitemoehybrid/test_modeling_granitemoehybrid.py::GraniteMoeHybridModelTest::test_flash_attn_2_inference_equivalence",
517
+ "commit": null,
518
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
519
+ "pr_number": null,
520
+ "author": null,
521
+ "merged_by": null,
522
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256265535"
523
+ },
524
+ {
525
+ "test": "tests/models/granitemoehybrid/test_modeling_granitemoehybrid.py::GraniteMoeHybridModelTest::test_flash_attn_2_inference_equivalence_right_padding",
526
+ "commit": null,
527
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
528
+ "pr_number": null,
529
+ "author": null,
530
+ "merged_by": null,
531
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256265535"
532
+ }
533
+ ]
534
+ },
535
+ "granitemoeshared": {
536
+ "single-gpu": [
537
+ {
538
+ "test": "tests/models/granitemoeshared/test_modeling_granitemoeshared.py::GraniteMoeSharedModelTest::test_flash_attention_2_padding_matches_padding_free_with_position_ids_and_fa_kwargs",
539
+ "commit": null,
540
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
541
+ "pr_number": null,
542
+ "author": null,
543
+ "merged_by": null,
544
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256265588"
545
+ }
546
+ ]
547
+ },
548
+ "grounding_dino": {
549
+ "single-gpu": [
550
+ {
551
+ "test": "tests/models/grounding_dino/test_modeling_grounding_dino.py::GroundingDinoModelTest::test_flash_attn_2_inference_equivalence",
552
+ "commit": null,
553
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
554
+ "pr_number": null,
555
+ "author": null,
556
+ "merged_by": null,
557
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256265608"
558
+ },
559
+ {
560
+ "test": "tests/models/grounding_dino/test_modeling_grounding_dino.py::GroundingDinoModelTest::test_flash_attn_2_inference_equivalence_right_padding",
561
+ "commit": null,
562
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
563
+ "pr_number": null,
564
+ "author": null,
565
+ "merged_by": null,
566
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256265608"
567
+ }
568
+ ]
569
+ },
570
+ "instructblip": {
571
+ "single-gpu": [
572
+ {
573
+ "test": "tests/models/instructblip/test_modeling_instructblip.py::InstructBlipForConditionalGenerationDecoderOnlyTest::test_flash_attn_2_fp32_ln",
574
+ "commit": null,
575
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
576
+ "pr_number": null,
577
+ "author": null,
578
+ "merged_by": null,
579
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256265897"
580
+ },
581
+ {
582
+ "test": "tests/models/instructblip/test_modeling_instructblip.py::InstructBlipForConditionalGenerationDecoderOnlyTest::test_flash_attn_2_from_config",
583
+ "commit": null,
584
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
585
+ "pr_number": null,
586
+ "author": null,
587
+ "merged_by": null,
588
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256265897"
589
+ }
590
+ ]
591
+ },
592
+ "instructblipvideo": {
593
+ "single-gpu": [
594
+ {
595
+ "test": "tests/models/instructblipvideo/test_modeling_instructblipvideo.py::InstructBlipVideoForConditionalGenerationDecoderOnlyTest::test_flash_attn_2_fp32_ln",
596
+ "commit": null,
597
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
598
+ "pr_number": null,
599
+ "author": null,
600
+ "merged_by": null,
601
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256265919"
602
+ },
603
+ {
604
+ "test": "tests/models/instructblipvideo/test_modeling_instructblipvideo.py::InstructBlipVideoForConditionalGenerationDecoderOnlyTest::test_flash_attn_2_from_config",
605
+ "commit": null,
606
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
607
+ "pr_number": null,
608
+ "author": null,
609
+ "merged_by": null,
610
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256265919"
611
+ }
612
+ ]
613
+ },
614
+ "janus": {
615
+ "single-gpu": [
616
+ {
617
+ "test": "tests/models/janus/test_modeling_janus.py::JanusVisionText2TextModelTest::test_flash_attn_2_inference_equivalence",
618
+ "commit": null,
619
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
620
+ "pr_number": null,
621
+ "author": null,
622
+ "merged_by": null,
623
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256265979"
624
+ },
625
+ {
626
+ "test": "tests/models/janus/test_modeling_janus.py::JanusVisionText2TextModelTest::test_flash_attn_2_inference_equivalence_right_padding",
627
+ "commit": null,
628
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
629
+ "pr_number": null,
630
+ "author": null,
631
+ "merged_by": null,
632
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256265979"
633
+ }
634
+ ]
635
+ },
636
+ "jetmoe": {
637
+ "single-gpu": [
638
+ {
639
+ "test": "tests/models/jetmoe/test_modeling_jetmoe.py::JetMoeModelTest::test_flash_attn_2_fp32_ln",
640
+ "commit": null,
641
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
642
+ "pr_number": null,
643
+ "author": null,
644
+ "merged_by": null,
645
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256266023"
646
+ }
647
+ ]
648
+ },
649
+ "kosmos2": {
650
+ "single-gpu": [
651
+ {
652
+ "test": "tests/models/kosmos2/test_modeling_kosmos2.py::Kosmos2ModelTest::test_eager_matches_fa2_generate",
653
+ "commit": null,
654
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
655
+ "pr_number": null,
656
+ "author": null,
657
+ "merged_by": null,
658
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256266038"
659
+ },
660
+ {
661
+ "test": "tests/models/kosmos2/test_modeling_kosmos2.py::Kosmos2ModelTest::test_flash_attention_2_continue_generate_with_position_ids",
662
+ "commit": null,
663
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
664
+ "pr_number": null,
665
+ "author": null,
666
+ "merged_by": null,
667
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256266038"
668
+ },
669
+ {
670
+ "test": "tests/models/kosmos2/test_modeling_kosmos2.py::Kosmos2ModelTest::test_flash_attn_2_can_dispatch_composite_models",
671
+ "commit": null,
672
+ "status": "flaky: test fails on the current CI run (commit: ffb35fe14283d61cd5434a32d04d07993e66477a) but passes during the check.",
673
+ "pr_number": null,
674
+ "author": null,
675
+ "merged_by": null,
676
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256266038"
677
+ },
678
+ {
679
+ "test": "tests/models/kosmos2/test_modeling_kosmos2.py::Kosmos2ModelTest::test_flash_attn_2_fp32_ln",
680
+ "commit": null,
681
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
682
+ "pr_number": null,
683
+ "author": null,
684
+ "merged_by": null,
685
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256266038"
686
+ },
687
+ {
688
+ "test": "tests/models/kosmos2/test_modeling_kosmos2.py::Kosmos2ModelTest::test_flash_attn_2_from_config",
689
+ "commit": null,
690
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
691
+ "pr_number": null,
692
+ "author": null,
693
+ "merged_by": null,
694
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256266038"
695
+ },
696
+ {
697
+ "test": "tests/models/kosmos2/test_modeling_kosmos2.py::Kosmos2ModelTest::test_flash_attn_2_inference_equivalence",
698
+ "commit": null,
699
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
700
+ "pr_number": null,
701
+ "author": null,
702
+ "merged_by": null,
703
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256266038"
704
+ },
705
+ {
706
+ "test": "tests/models/kosmos2/test_modeling_kosmos2.py::Kosmos2ModelTest::test_flash_attn_2_inference_equivalence_right_padding",
707
+ "commit": null,
708
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
709
+ "pr_number": null,
710
+ "author": null,
711
+ "merged_by": null,
712
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256266038"
713
+ }
714
+ ]
715
+ },
716
+ "kosmos2_5": {
717
+ "single-gpu": [
718
+ {
719
+ "test": "tests/models/kosmos2_5/test_modeling_kosmos2_5.py::Kosmos2_5ModelTest::test_flash_attn_2_can_dispatch_composite_models",
720
+ "commit": null,
721
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
722
+ "pr_number": null,
723
+ "author": null,
724
+ "merged_by": null,
725
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256266067"
726
+ }
727
+ ]
728
+ },
729
+ "kyutai_speech_to_text": {
730
+ "single-gpu": [
731
+ {
732
+ "test": "tests/models/kyutai_speech_to_text/test_modeling_kyutai_speech_to_text.py::KyutaiSpeechToTextModelTest::test_eager_matches_fa2_generate",
733
+ "commit": null,
734
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
735
+ "pr_number": null,
736
+ "author": null,
737
+ "merged_by": null,
738
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256266060"
739
+ },
740
+ {
741
+ "test": "tests/models/kyutai_speech_to_text/test_modeling_kyutai_speech_to_text.py::KyutaiSpeechToTextModelTest::test_flash_attn_2_inference_equivalence",
742
+ "commit": null,
743
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
744
+ "pr_number": null,
745
+ "author": null,
746
+ "merged_by": null,
747
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256266060"
748
+ },
749
+ {
750
+ "test": "tests/models/kyutai_speech_to_text/test_modeling_kyutai_speech_to_text.py::KyutaiSpeechToTextModelTest::test_flash_attn_2_inference_equivalence_right_padding",
751
+ "commit": null,
752
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
753
+ "pr_number": null,
754
+ "author": null,
755
+ "merged_by": null,
756
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256266060"
757
+ }
758
+ ]
759
+ },
760
+ "lfm2": {
761
+ "single-gpu": [
762
+ {
763
+ "test": "tests/models/lfm2/test_modeling_lfm2.py::Lfm2ModelTest::test_flash_attention_2_padding_matches_padding_free_with_position_ids",
764
+ "commit": null,
765
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
766
+ "pr_number": null,
767
+ "author": null,
768
+ "merged_by": null,
769
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256266197"
770
+ },
771
+ {
772
+ "test": "tests/models/lfm2/test_modeling_lfm2.py::Lfm2ModelTest::test_flash_attn_2_can_compile_with_attention_mask_None_without_graph_break",
773
+ "commit": null,
774
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
775
+ "pr_number": null,
776
+ "author": null,
777
+ "merged_by": null,
778
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256266197"
779
+ }
780
+ ]
781
+ },
782
+ "lfm2_moe": {
783
+ "single-gpu": [
784
+ {
785
+ "test": "tests/models/lfm2_moe/test_modeling_lfm2_moe.py::Lfm2MoeModelTest::test_flash_attention_2_padding_matches_padding_free_with_position_ids",
786
+ "commit": null,
787
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
788
+ "pr_number": null,
789
+ "author": null,
790
+ "merged_by": null,
791
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256266241"
792
+ },
793
+ {
794
+ "test": "tests/models/lfm2_moe/test_modeling_lfm2_moe.py::Lfm2MoeModelTest::test_flash_attn_2_can_compile_with_attention_mask_None_without_graph_break",
795
+ "commit": null,
796
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
797
+ "pr_number": null,
798
+ "author": null,
799
+ "merged_by": null,
800
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256266241"
801
+ }
802
+ ]
803
+ },
804
+ "lfm2_vl": {
805
+ "single-gpu": [
806
+ {
807
+ "test": "tests/models/lfm2_vl/test_modeling_lfm2_vl.py::Lfm2VlModelTest::test_flash_attn_2_inference_equivalence",
808
+ "commit": null,
809
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
810
+ "pr_number": null,
811
+ "author": null,
812
+ "merged_by": null,
813
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256266254"
814
+ },
815
+ {
816
+ "test": "tests/models/lfm2_vl/test_modeling_lfm2_vl.py::Lfm2VlModelTest::test_flash_attn_2_inference_equivalence_right_padding",
817
+ "commit": null,
818
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
819
+ "pr_number": null,
820
+ "author": null,
821
+ "merged_by": null,
822
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256266254"
823
+ }
824
+ ]
825
+ },
826
+ "llava_next": {
827
+ "single-gpu": [
828
+ {
829
+ "test": "tests/models/llava_next/test_modeling_llava_next.py::LlavaNextForConditionalGenerationModelTest::test_flash_attn_2_inference_equivalence",
830
+ "commit": null,
831
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
832
+ "pr_number": null,
833
+ "author": null,
834
+ "merged_by": null,
835
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256266347"
836
+ },
837
+ {
838
+ "test": "tests/models/llava_next/test_modeling_llava_next.py::LlavaNextForConditionalGenerationModelTest::test_flash_attn_2_inference_equivalence_right_padding",
839
+ "commit": null,
840
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
841
+ "pr_number": null,
842
+ "author": null,
843
+ "merged_by": null,
844
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256266347"
845
+ }
846
+ ]
847
+ },
848
+ "llava_next_video": {
849
+ "single-gpu": [
850
+ {
851
+ "test": "tests/models/llava_next_video/test_modeling_llava_next_video.py::LlavaNextVideoForConditionalGenerationModelTest::test_flash_attn_2_inference_equivalence",
852
+ "commit": null,
853
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
854
+ "pr_number": null,
855
+ "author": null,
856
+ "merged_by": null,
857
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256266334"
858
+ },
859
+ {
860
+ "test": "tests/models/llava_next_video/test_modeling_llava_next_video.py::LlavaNextVideoForConditionalGenerationModelTest::test_flash_attn_2_inference_equivalence_right_padding",
861
+ "commit": null,
862
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
863
+ "pr_number": null,
864
+ "author": null,
865
+ "merged_by": null,
866
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256266334"
867
+ }
868
+ ]
869
+ },
870
+ "llava_onevision": {
871
+ "single-gpu": [
872
+ {
873
+ "test": "tests/models/llava_onevision/test_modeling_llava_onevision.py::LlavaOnevisionForConditionalGenerationModelTest::test_flash_attn_2_inference_equivalence",
874
+ "commit": null,
875
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
876
+ "pr_number": null,
877
+ "author": null,
878
+ "merged_by": null,
879
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256266412"
880
+ },
881
+ {
882
+ "test": "tests/models/llava_onevision/test_modeling_llava_onevision.py::LlavaOnevisionForConditionalGenerationModelTest::test_flash_attn_2_inference_equivalence_right_padding",
883
+ "commit": null,
884
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
885
+ "pr_number": null,
886
+ "author": null,
887
+ "merged_by": null,
888
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256266412"
889
+ }
890
+ ]
891
+ },
892
+ "mask2former": {
893
+ "single-gpu": [
894
+ {
895
+ "test": "tests/models/mask2former/test_modeling_mask2former.py::Mask2FormerModelTest::test_flash_attn_2_inference_equivalence",
896
+ "commit": null,
897
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
898
+ "pr_number": null,
899
+ "author": null,
900
+ "merged_by": null,
901
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256266649"
902
+ },
903
+ {
904
+ "test": "tests/models/mask2former/test_modeling_mask2former.py::Mask2FormerModelTest::test_flash_attn_2_inference_equivalence_right_padding",
905
+ "commit": null,
906
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
907
+ "pr_number": null,
908
+ "author": null,
909
+ "merged_by": null,
910
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256266649"
911
+ }
912
+ ]
913
+ },
914
+ "maskformer": {
915
+ "single-gpu": [
916
+ {
917
+ "test": "tests/models/maskformer/test_modeling_maskformer.py::MaskFormerModelTest::test_flash_attn_2_inference_equivalence",
918
+ "commit": null,
919
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
920
+ "pr_number": null,
921
+ "author": null,
922
+ "merged_by": null,
923
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256266666"
924
+ },
925
+ {
926
+ "test": "tests/models/maskformer/test_modeling_maskformer.py::MaskFormerModelTest::test_flash_attn_2_inference_equivalence_right_padding",
927
+ "commit": null,
928
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
929
+ "pr_number": null,
930
+ "author": null,
931
+ "merged_by": null,
932
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256266666"
933
+ },
934
+ {
935
+ "test": "tests/models/maskformer/test_modeling_maskformer_swin.py::MaskFormerSwinModelTest::test_flash_attn_2_inference_equivalence",
936
+ "commit": null,
937
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
938
+ "pr_number": null,
939
+ "author": null,
940
+ "merged_by": null,
941
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256266666"
942
+ },
943
+ {
944
+ "test": "tests/models/maskformer/test_modeling_maskformer_swin.py::MaskFormerSwinModelTest::test_flash_attn_2_inference_equivalence_right_padding",
945
+ "commit": null,
946
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
947
+ "pr_number": null,
948
+ "author": null,
949
+ "merged_by": null,
950
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256266666"
951
+ }
952
+ ]
953
+ },
954
+ "mllama": {
955
+ "single-gpu": [
956
+ {
957
+ "test": "tests/models/mllama/test_modeling_mllama.py::MllamaForConditionalGenerationModelTest::test_eager_matches_fa2_generate",
958
+ "commit": null,
959
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
960
+ "pr_number": null,
961
+ "author": null,
962
+ "merged_by": null,
963
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256262497"
964
+ },
965
+ {
966
+ "test": "tests/models/mllama/test_modeling_mllama.py::MllamaForConditionalGenerationModelTest::test_flash_attn_2_inference_equivalence",
967
+ "commit": null,
968
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
969
+ "pr_number": null,
970
+ "author": null,
971
+ "merged_by": null,
972
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256262497"
973
+ },
974
+ {
975
+ "test": "tests/models/mllama/test_modeling_mllama.py::MllamaForConditionalGenerationModelTest::test_flash_attn_2_inference_equivalence_right_padding",
976
+ "commit": null,
977
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
978
+ "pr_number": null,
979
+ "author": null,
980
+ "merged_by": null,
981
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256262497"
982
+ }
983
+ ]
984
+ },
985
+ "mm_grounding_dino": {
986
+ "single-gpu": [
987
+ {
988
+ "test": "tests/models/mm_grounding_dino/test_modeling_mm_grounding_dino.py::MMGroundingDinoModelTest::test_flash_attn_2_inference_equivalence",
989
+ "commit": null,
990
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
991
+ "pr_number": null,
992
+ "author": null,
993
+ "merged_by": null,
994
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256262500"
995
+ },
996
+ {
997
+ "test": "tests/models/mm_grounding_dino/test_modeling_mm_grounding_dino.py::MMGroundingDinoModelTest::test_flash_attn_2_inference_equivalence_right_padding",
998
+ "commit": null,
999
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1000
+ "pr_number": null,
1001
+ "author": null,
1002
+ "merged_by": null,
1003
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256262500"
1004
+ }
1005
+ ]
1006
+ },
1007
+ "moshi": {
1008
+ "single-gpu": [
1009
+ {
1010
+ "test": "tests/models/moshi/test_modeling_moshi.py::MoshiDecoderTest::test_flash_attn_2_inference_equivalence",
1011
+ "commit": null,
1012
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1013
+ "pr_number": null,
1014
+ "author": null,
1015
+ "merged_by": null,
1016
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256262704"
1017
+ },
1018
+ {
1019
+ "test": "tests/models/moshi/test_modeling_moshi.py::MoshiDecoderTest::test_flash_attn_2_inference_equivalence_right_padding",
1020
+ "commit": null,
1021
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1022
+ "pr_number": null,
1023
+ "author": null,
1024
+ "merged_by": null,
1025
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256262704"
1026
+ },
1027
+ {
1028
+ "test": "tests/models/moshi/test_modeling_moshi.py::MoshiTest::test_eager_matches_fa2_generate",
1029
+ "commit": null,
1030
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1031
+ "pr_number": null,
1032
+ "author": null,
1033
+ "merged_by": null,
1034
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256262704"
1035
+ },
1036
+ {
1037
+ "test": "tests/models/moshi/test_modeling_moshi.py::MoshiTest::test_flash_attn_2_fp32_ln",
1038
+ "commit": null,
1039
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1040
+ "pr_number": null,
1041
+ "author": null,
1042
+ "merged_by": null,
1043
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256262704"
1044
+ },
1045
+ {
1046
+ "test": "tests/models/moshi/test_modeling_moshi.py::MoshiTest::test_flash_attn_2_from_config",
1047
+ "commit": null,
1048
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1049
+ "pr_number": null,
1050
+ "author": null,
1051
+ "merged_by": null,
1052
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256262704"
1053
+ },
1054
+ {
1055
+ "test": "tests/models/moshi/test_modeling_moshi.py::MoshiTest::test_flash_attn_2_inference_equivalence",
1056
+ "commit": null,
1057
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1058
+ "pr_number": null,
1059
+ "author": null,
1060
+ "merged_by": null,
1061
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256262704"
1062
+ },
1063
+ {
1064
+ "test": "tests/models/moshi/test_modeling_moshi.py::MoshiTest::test_flash_attn_2_inference_equivalence_right_padding",
1065
+ "commit": null,
1066
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1067
+ "pr_number": null,
1068
+ "author": null,
1069
+ "merged_by": null,
1070
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256262704"
1071
+ }
1072
+ ]
1073
+ },
1074
+ "nemotron": {
1075
+ "single-gpu": [
1076
+ {
1077
+ "test": "tests/models/nemotron/test_modeling_nemotron.py::NemotronModelTest::test_flash_attn_2_equivalence",
1078
+ "commit": null,
1079
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1080
+ "pr_number": null,
1081
+ "author": null,
1082
+ "merged_by": null,
1083
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256262902"
1084
+ }
1085
+ ]
1086
+ },
1087
+ "olmo": {
1088
+ "single-gpu": [
1089
+ {
1090
+ "test": "tests/models/olmo/test_modeling_olmo.py::OlmoModelTest::test_flash_attention_2_padding_matches_padding_free_with_position_ids",
1091
+ "commit": null,
1092
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1093
+ "pr_number": null,
1094
+ "author": null,
1095
+ "merged_by": null,
1096
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256263145"
1097
+ }
1098
+ ]
1099
+ },
1100
+ "olmo2": {
1101
+ "single-gpu": [
1102
+ {
1103
+ "test": "tests/models/olmo2/test_modeling_olmo2.py::Olmo2ModelTest::test_flash_attention_2_padding_matches_padding_free_with_position_ids",
1104
+ "commit": null,
1105
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1106
+ "pr_number": null,
1107
+ "author": null,
1108
+ "merged_by": null,
1109
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256263087"
1110
+ }
1111
+ ]
1112
+ },
1113
+ "omdet_turbo": {
1114
+ "single-gpu": [
1115
+ {
1116
+ "test": "tests/models/omdet_turbo/test_modeling_omdet_turbo.py::OmDetTurboModelTest::test_flash_attn_2_inference_equivalence",
1117
+ "commit": null,
1118
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1119
+ "pr_number": null,
1120
+ "author": null,
1121
+ "merged_by": null,
1122
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256263331"
1123
+ },
1124
+ {
1125
+ "test": "tests/models/omdet_turbo/test_modeling_omdet_turbo.py::OmDetTurboModelTest::test_flash_attn_2_inference_equivalence_right_padding",
1126
+ "commit": null,
1127
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1128
+ "pr_number": null,
1129
+ "author": null,
1130
+ "merged_by": null,
1131
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256263331"
1132
+ }
1133
+ ]
1134
+ },
1135
+ "oneformer": {
1136
+ "single-gpu": [
1137
+ {
1138
+ "test": "tests/models/oneformer/test_modeling_oneformer.py::OneFormerModelTest::test_flash_attn_2_inference_equivalence",
1139
+ "commit": null,
1140
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1141
+ "pr_number": null,
1142
+ "author": null,
1143
+ "merged_by": null,
1144
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256263301"
1145
+ },
1146
+ {
1147
+ "test": "tests/models/oneformer/test_modeling_oneformer.py::OneFormerModelTest::test_flash_attn_2_inference_equivalence_right_padding",
1148
+ "commit": null,
1149
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1150
+ "pr_number": null,
1151
+ "author": null,
1152
+ "merged_by": null,
1153
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256263301"
1154
+ }
1155
+ ]
1156
+ },
1157
+ "paligemma": {
1158
+ "single-gpu": [
1159
+ {
1160
+ "test": "tests/models/paligemma/test_modeling_paligemma.py::PaliGemmaForConditionalGenerationModelTest::test_flash_attn_2_from_config",
1161
+ "commit": null,
1162
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1163
+ "pr_number": null,
1164
+ "author": null,
1165
+ "merged_by": null,
1166
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256263449"
1167
+ },
1168
+ {
1169
+ "test": "tests/models/paligemma/test_modeling_paligemma.py::PaliGemmaForConditionalGenerationModelTest::test_flash_attn_2_inference_equivalence",
1170
+ "commit": null,
1171
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1172
+ "pr_number": null,
1173
+ "author": null,
1174
+ "merged_by": null,
1175
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256263449"
1176
+ },
1177
+ {
1178
+ "test": "tests/models/paligemma/test_modeling_paligemma.py::PaliGemmaForConditionalGenerationModelTest::test_flash_attn_2_inference_equivalence_right_padding",
1179
+ "commit": null,
1180
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1181
+ "pr_number": null,
1182
+ "author": null,
1183
+ "merged_by": null,
1184
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256263449"
1185
+ }
1186
+ ]
1187
+ },
1188
+ "paligemma2": {
1189
+ "single-gpu": [
1190
+ {
1191
+ "test": "tests/models/paligemma2/test_modeling_paligemma2.py::PaliGemma2ForConditionalGenerationModelTest::test_flash_attn_2_from_config",
1192
+ "commit": null,
1193
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1194
+ "pr_number": null,
1195
+ "author": null,
1196
+ "merged_by": null,
1197
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256263448"
1198
+ },
1199
+ {
1200
+ "test": "tests/models/paligemma2/test_modeling_paligemma2.py::PaliGemma2ForConditionalGenerationModelTest::test_flash_attn_2_inference_equivalence",
1201
+ "commit": null,
1202
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1203
+ "pr_number": null,
1204
+ "author": null,
1205
+ "merged_by": null,
1206
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256263448"
1207
+ },
1208
+ {
1209
+ "test": "tests/models/paligemma2/test_modeling_paligemma2.py::PaliGemma2ForConditionalGenerationModelTest::test_flash_attn_2_inference_equivalence_right_padding",
1210
+ "commit": null,
1211
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1212
+ "pr_number": null,
1213
+ "author": null,
1214
+ "merged_by": null,
1215
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256263448"
1216
+ }
1217
+ ]
1218
+ },
1219
+ "pegasus_x": {
1220
+ "single-gpu": [
1221
+ {
1222
+ "test": "tests/models/pegasus_x/test_modeling_pegasus_x.py::PegasusXModelTest::test_flash_attn_2_inference_equivalence",
1223
+ "commit": null,
1224
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1225
+ "pr_number": null,
1226
+ "author": null,
1227
+ "merged_by": null,
1228
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256263633"
1229
+ },
1230
+ {
1231
+ "test": "tests/models/pegasus_x/test_modeling_pegasus_x.py::PegasusXModelTest::test_flash_attn_2_inference_equivalence_right_padding",
1232
+ "commit": null,
1233
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1234
+ "pr_number": null,
1235
+ "author": null,
1236
+ "merged_by": null,
1237
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256263633"
1238
+ },
1239
+ {
1240
+ "test": "tests/models/pegasus_x/test_modeling_pegasus_x.py::PegasusXStandaloneDecoderModelTest::test_flash_attn_2_inference_equivalence",
1241
+ "commit": null,
1242
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1243
+ "pr_number": null,
1244
+ "author": null,
1245
+ "merged_by": null,
1246
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256263633"
1247
+ },
1248
+ {
1249
+ "test": "tests/models/pegasus_x/test_modeling_pegasus_x.py::PegasusXStandaloneDecoderModelTest::test_flash_attn_2_inference_equivalence_right_padding",
1250
+ "commit": null,
1251
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1252
+ "pr_number": null,
1253
+ "author": null,
1254
+ "merged_by": null,
1255
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256263633"
1256
+ }
1257
+ ]
1258
+ },
1259
+ "perception_lm": {
1260
+ "single-gpu": [
1261
+ {
1262
+ "test": "tests/models/perception_lm/test_modeling_perception_lm.py::PerceptionLMForConditionalGenerationModelTest::test_flash_attention_2_continue_generate_with_position_ids",
1263
+ "commit": null,
1264
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1265
+ "pr_number": null,
1266
+ "author": null,
1267
+ "merged_by": null,
1268
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256263665"
1269
+ }
1270
+ ]
1271
+ },
1272
+ "phi": {
1273
+ "single-gpu": [
1274
+ {
1275
+ "test": "tests/models/phi/test_modeling_phi.py::PhiModelTest::test_flash_attention_2_padding_matches_padding_free_with_position_ids",
1276
+ "commit": null,
1277
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1278
+ "pr_number": null,
1279
+ "author": null,
1280
+ "merged_by": null,
1281
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256263676"
1282
+ }
1283
+ ]
1284
+ },
1285
+ "phimoe": {
1286
+ "single-gpu": [
1287
+ {
1288
+ "test": "tests/models/phimoe/test_modeling_phimoe.py::PhimoeModelTest::test_flash_attn_2_equivalence",
1289
+ "commit": null,
1290
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1291
+ "pr_number": null,
1292
+ "author": null,
1293
+ "merged_by": null,
1294
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256263814"
1295
+ }
1296
+ ]
1297
+ },
1298
+ "pixtral": {
1299
+ "single-gpu": [
1300
+ {
1301
+ "test": "tests/models/pixtral/test_modeling_pixtral.py::PixtralVisionModelModelTest::test_flash_attn_2_inference_equivalence",
1302
+ "commit": null,
1303
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1304
+ "pr_number": null,
1305
+ "author": null,
1306
+ "merged_by": null,
1307
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256263835"
1308
+ },
1309
+ {
1310
+ "test": "tests/models/pixtral/test_modeling_pixtral.py::PixtralVisionModelModelTest::test_flash_attn_2_inference_equivalence_right_padding",
1311
+ "commit": null,
1312
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1313
+ "pr_number": null,
1314
+ "author": null,
1315
+ "merged_by": null,
1316
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256263835"
1317
+ }
1318
+ ]
1319
+ },
1320
+ "qwen2_5_vl": {
1321
+ "single-gpu": [
1322
+ {
1323
+ "test": "tests/models/qwen2_5_vl/test_modeling_qwen2_5_vl.py::Qwen2_5_VLIntegrationTest::test_small_model_integration_test_batch_wo_image_flashatt2",
1324
+ "commit": null,
1325
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1326
+ "pr_number": null,
1327
+ "author": null,
1328
+ "merged_by": null,
1329
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256264041"
1330
+ }
1331
+ ]
1332
+ },
1333
+ "qwen3_omni_moe": {
1334
+ "single-gpu": [
1335
+ {
1336
+ "test": "tests/models/qwen3_omni_moe/test_modeling_qwen3_omni_moe.py::Qwen2_5OmniThinkerForConditionalGenerationModelTest::test_flash_attention_2_padding_matches_padding_free_with_position_ids",
1337
+ "commit": null,
1338
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1339
+ "pr_number": null,
1340
+ "author": null,
1341
+ "merged_by": null,
1342
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256264215"
1343
+ },
1344
+ {
1345
+ "test": "tests/models/qwen3_omni_moe/test_modeling_qwen3_omni_moe.py::Qwen2_5OmniThinkerForConditionalGenerationModelTest::test_flash_attention_2_padding_matches_padding_free_with_position_ids_and_fa_kwargs",
1346
+ "commit": null,
1347
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1348
+ "pr_number": null,
1349
+ "author": null,
1350
+ "merged_by": null,
1351
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256264215"
1352
+ },
1353
+ {
1354
+ "test": "tests/models/qwen3_omni_moe/test_modeling_qwen3_omni_moe.py::Qwen2_5OmniModelIntegrationTest::test_small_model_integration_test_batch_flashatt2",
1355
+ "commit": null,
1356
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1357
+ "pr_number": null,
1358
+ "author": null,
1359
+ "merged_by": null,
1360
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256264215"
1361
+ }
1362
+ ]
1363
+ },
1364
+ "qwen3_vl_moe": {
1365
+ "single-gpu": [
1366
+ {
1367
+ "test": "tests/models/qwen3_vl_moe/test_modeling_qwen3_vl_moe.py::Qwen3VLMoeIntegrationTest::test_small_model_integration_test_batch_flashatt2",
1368
+ "commit": null,
1369
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1370
+ "pr_number": null,
1371
+ "author": null,
1372
+ "merged_by": null,
1373
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256264218"
1374
+ },
1375
+ {
1376
+ "test": "tests/models/qwen3_vl_moe/test_modeling_qwen3_vl_moe.py::Qwen3VLMoeIntegrationTest::test_small_model_integration_test_batch_wo_image_flashatt2",
1377
+ "commit": null,
1378
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1379
+ "pr_number": null,
1380
+ "author": null,
1381
+ "merged_by": null,
1382
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256264218"
1383
+ }
1384
+ ]
1385
+ },
1386
+ "sam2": {
1387
+ "single-gpu": [
1388
+ {
1389
+ "test": "tests/models/sam2/test_modeling_sam2.py::Sam2ModelTest::test_flash_attn_2_can_dispatch_composite_models",
1390
+ "commit": null,
1391
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1392
+ "pr_number": null,
1393
+ "author": null,
1394
+ "merged_by": null,
1395
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256264556"
1396
+ }
1397
+ ]
1398
+ },
1399
+ "smollm3": {
1400
+ "single-gpu": [
1401
+ {
1402
+ "test": "tests/models/smollm3/test_modeling_smollm3.py::SmolLM3IntegrationTest::test_model_3b_long_prompt",
1403
+ "commit": null,
1404
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1405
+ "pr_number": null,
1406
+ "author": null,
1407
+ "merged_by": null,
1408
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256264885"
1409
+ }
1410
+ ]
1411
+ },
1412
+ "squeezebert": {
1413
+ "single-gpu": [
1414
+ {
1415
+ "test": "tests/models/squeezebert/test_modeling_squeezebert.py::SqueezeBertModelTest::test_flash_attn_2_inference_equivalence",
1416
+ "commit": null,
1417
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1418
+ "pr_number": null,
1419
+ "author": null,
1420
+ "merged_by": null,
1421
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256264948"
1422
+ },
1423
+ {
1424
+ "test": "tests/models/squeezebert/test_modeling_squeezebert.py::SqueezeBertModelTest::test_flash_attn_2_inference_equivalence_right_padding",
1425
+ "commit": null,
1426
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1427
+ "pr_number": null,
1428
+ "author": null,
1429
+ "merged_by": null,
1430
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256264948"
1431
+ }
1432
+ ]
1433
+ },
1434
+ "starcoder2": {
1435
+ "single-gpu": [
1436
+ {
1437
+ "test": "tests/models/starcoder2/test_modeling_starcoder2.py::Starcoder2IntegrationTest::test_starcoder2_batched_generation_fa2",
1438
+ "commit": null,
1439
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1440
+ "pr_number": null,
1441
+ "author": null,
1442
+ "merged_by": null,
1443
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256265091"
1444
+ }
1445
+ ]
1446
+ },
1447
+ "swin": {
1448
+ "single-gpu": [
1449
+ {
1450
+ "test": "tests/models/swin/test_modeling_swin.py::SwinModelTest::test_flash_attn_2_inference_equivalence",
1451
+ "commit": null,
1452
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1453
+ "pr_number": null,
1454
+ "author": null,
1455
+ "merged_by": null,
1456
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256265137"
1457
+ },
1458
+ {
1459
+ "test": "tests/models/swin/test_modeling_swin.py::SwinModelTest::test_flash_attn_2_inference_equivalence_right_padding",
1460
+ "commit": null,
1461
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1462
+ "pr_number": null,
1463
+ "author": null,
1464
+ "merged_by": null,
1465
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256265137"
1466
+ }
1467
+ ]
1468
+ },
1469
+ "swin2sr": {
1470
+ "single-gpu": [
1471
+ {
1472
+ "test": "tests/models/swin2sr/test_modeling_swin2sr.py::Swin2SRModelTest::test_flash_attn_2_inference_equivalence",
1473
+ "commit": null,
1474
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1475
+ "pr_number": null,
1476
+ "author": null,
1477
+ "merged_by": null,
1478
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256265136"
1479
+ },
1480
+ {
1481
+ "test": "tests/models/swin2sr/test_modeling_swin2sr.py::Swin2SRModelTest::test_flash_attn_2_inference_equivalence_right_padding",
1482
+ "commit": null,
1483
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1484
+ "pr_number": null,
1485
+ "author": null,
1486
+ "merged_by": null,
1487
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256265136"
1488
+ }
1489
+ ]
1490
+ },
1491
+ "swinv2": {
1492
+ "single-gpu": [
1493
+ {
1494
+ "test": "tests/models/swinv2/test_modeling_swinv2.py::Swinv2ModelTest::test_flash_attn_2_inference_equivalence",
1495
+ "commit": null,
1496
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1497
+ "pr_number": null,
1498
+ "author": null,
1499
+ "merged_by": null,
1500
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256265145"
1501
+ },
1502
+ {
1503
+ "test": "tests/models/swinv2/test_modeling_swinv2.py::Swinv2ModelTest::test_flash_attn_2_inference_equivalence_right_padding",
1504
+ "commit": null,
1505
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1506
+ "pr_number": null,
1507
+ "author": null,
1508
+ "merged_by": null,
1509
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256265145"
1510
+ }
1511
+ ]
1512
+ },
1513
+ "t5gemma": {
1514
+ "single-gpu": [
1515
+ {
1516
+ "test": "tests/models/t5gemma/test_modeling_t5gemma.py::T5GemmaModelTest::test_flash_attn_2_can_compile_with_attention_mask_None_without_graph_break",
1517
+ "commit": null,
1518
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1519
+ "pr_number": null,
1520
+ "author": null,
1521
+ "merged_by": null,
1522
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256265203"
1523
+ }
1524
+ ]
1525
+ },
1526
+ "zamba": {
1527
+ "single-gpu": [
1528
+ {
1529
+ "test": "tests/models/zamba/test_modeling_zamba.py::ZambaModelTest::test_flash_attn_2_fp32_ln",
1530
+ "commit": null,
1531
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1532
+ "pr_number": null,
1533
+ "author": null,
1534
+ "merged_by": null,
1535
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256266375"
1536
+ }
1537
+ ]
1538
+ }
1539
+ }
1540
+ }