hf-transformers-bot commited on
Commit
1d98d11
·
verified ·
1 Parent(s): 3e15fde

Upload 2025-11-04/ci_results_run_models_gpu/new_failures_with_bad_commit_grouped_by_authors.json with huggingface_hub

Browse files
2025-11-04/ci_results_run_models_gpu/new_failures_with_bad_commit_grouped_by_authors.json ADDED
@@ -0,0 +1,1550 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "null": {
3
+ "bamba": {
4
+ "single-gpu": [
5
+ {
6
+ "test": "tests/models/bamba/test_modeling_bamba.py::BambaModelTest::test_flash_attn_2_inference_equivalence",
7
+ "commit": null,
8
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
9
+ "pr_number": null,
10
+ "author": null,
11
+ "merged_by": null,
12
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426979005"
13
+ },
14
+ {
15
+ "test": "tests/models/bamba/test_modeling_bamba.py::BambaModelTest::test_flash_attn_2_inference_equivalence_right_padding",
16
+ "commit": null,
17
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
18
+ "pr_number": null,
19
+ "author": null,
20
+ "merged_by": null,
21
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426979005"
22
+ }
23
+ ]
24
+ },
25
+ "bark": {
26
+ "single-gpu": [
27
+ {
28
+ "test": "tests/models/bark/test_modeling_bark.py::BarkSemanticModelTest::test_eager_matches_fa2_generate",
29
+ "commit": null,
30
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
31
+ "pr_number": null,
32
+ "author": null,
33
+ "merged_by": null,
34
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426979038"
35
+ },
36
+ {
37
+ "test": "tests/models/bark/test_modeling_bark.py::BarkSemanticModelTest::test_flash_attention_2_continue_generate_with_position_ids",
38
+ "commit": null,
39
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
40
+ "pr_number": null,
41
+ "author": null,
42
+ "merged_by": null,
43
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426979038"
44
+ },
45
+ {
46
+ "test": "tests/models/bark/test_modeling_bark.py::BarkCoarseModelTest::test_eager_matches_fa2_generate",
47
+ "commit": null,
48
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
49
+ "pr_number": null,
50
+ "author": null,
51
+ "merged_by": null,
52
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426979038"
53
+ },
54
+ {
55
+ "test": "tests/models/bark/test_modeling_bark.py::BarkCoarseModelTest::test_flash_attention_2_continue_generate_with_position_ids",
56
+ "commit": null,
57
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
58
+ "pr_number": null,
59
+ "author": null,
60
+ "merged_by": null,
61
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426979038"
62
+ }
63
+ ]
64
+ },
65
+ "blt": {
66
+ "single-gpu": [
67
+ {
68
+ "test": "tests/models/blt/test_modeling_blt.py::BltModelTest::test_flash_attn_2_can_compile_with_attention_mask_None_without_graph_break",
69
+ "commit": null,
70
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
71
+ "pr_number": null,
72
+ "author": null,
73
+ "merged_by": null,
74
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426979529"
75
+ }
76
+ ]
77
+ },
78
+ "chameleon": {
79
+ "single-gpu": [
80
+ {
81
+ "test": "tests/models/chameleon/test_modeling_chameleon.py::ChameleonModelTest::test_flash_attention_2_padding_matches_padding_free_with_position_ids",
82
+ "commit": null,
83
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
84
+ "pr_number": null,
85
+ "author": null,
86
+ "merged_by": null,
87
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426979630"
88
+ },
89
+ {
90
+ "test": "tests/models/chameleon/test_modeling_chameleon.py::ChameleonVision2SeqModelTest::test_flash_attention_2_padding_matches_padding_free_with_position_ids_and_fa_kwargs",
91
+ "commit": null,
92
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
93
+ "pr_number": null,
94
+ "author": null,
95
+ "merged_by": null,
96
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426979630"
97
+ }
98
+ ]
99
+ },
100
+ "clap": {
101
+ "single-gpu": [
102
+ {
103
+ "test": "tests/models/clap/test_modeling_clap.py::ClapAudioModelTest::test_flash_attn_2_inference_equivalence",
104
+ "commit": null,
105
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
106
+ "pr_number": null,
107
+ "author": null,
108
+ "merged_by": null,
109
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426979690"
110
+ },
111
+ {
112
+ "test": "tests/models/clap/test_modeling_clap.py::ClapAudioModelTest::test_flash_attn_2_inference_equivalence_right_padding",
113
+ "commit": null,
114
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
115
+ "pr_number": null,
116
+ "author": null,
117
+ "merged_by": null,
118
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426979690"
119
+ },
120
+ {
121
+ "test": "tests/models/clap/test_modeling_clap.py::ClapModelTest::test_flash_attn_2_inference_equivalence",
122
+ "commit": null,
123
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
124
+ "pr_number": null,
125
+ "author": null,
126
+ "merged_by": null,
127
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426979690"
128
+ },
129
+ {
130
+ "test": "tests/models/clap/test_modeling_clap.py::ClapModelTest::test_flash_attn_2_inference_equivalence_right_padding",
131
+ "commit": null,
132
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
133
+ "pr_number": null,
134
+ "author": null,
135
+ "merged_by": null,
136
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426979690"
137
+ }
138
+ ]
139
+ },
140
+ "colpali": {
141
+ "single-gpu": [
142
+ {
143
+ "test": "tests/models/colpali/test_modeling_colpali.py::ColPaliForRetrievalModelTest::test_flash_attn_2_inference_equivalence",
144
+ "commit": null,
145
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
146
+ "pr_number": null,
147
+ "author": null,
148
+ "merged_by": null,
149
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426979791"
150
+ },
151
+ {
152
+ "test": "tests/models/colpali/test_modeling_colpali.py::ColPaliForRetrievalModelTest::test_flash_attn_2_inference_equivalence_right_padding",
153
+ "commit": null,
154
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
155
+ "pr_number": null,
156
+ "author": null,
157
+ "merged_by": null,
158
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426979791"
159
+ }
160
+ ]
161
+ },
162
+ "deepseek_v2": {
163
+ "single-gpu": [
164
+ {
165
+ "test": "tests/models/deepseek_v2/test_modeling_deepseek_v2.py::DeepseekV2ModelTest::test_flash_attention_2_padding_matches_padding_free_with_position_ids",
166
+ "commit": null,
167
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
168
+ "pr_number": null,
169
+ "author": null,
170
+ "merged_by": null,
171
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426980167"
172
+ },
173
+ {
174
+ "test": "tests/models/deepseek_v2/test_modeling_deepseek_v2.py::DeepseekV2ModelTest::test_flash_attn_2_can_compile_with_attention_mask_None_without_graph_break",
175
+ "commit": null,
176
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
177
+ "pr_number": null,
178
+ "author": null,
179
+ "merged_by": null,
180
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426980167"
181
+ },
182
+ {
183
+ "test": "tests/models/deepseek_v2/test_modeling_deepseek_v2.py::DeepseekV2ModelTest::test_flash_attn_2_fp32_ln",
184
+ "commit": null,
185
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
186
+ "pr_number": null,
187
+ "author": null,
188
+ "merged_by": null,
189
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426980167"
190
+ }
191
+ ]
192
+ },
193
+ "deepseek_vl_hybrid": {
194
+ "single-gpu": [
195
+ {
196
+ "test": "tests/models/deepseek_vl_hybrid/test_modeling_deepseek_vl_hybrid.py::DeepseekVLHybridModelTest::test_flash_attention_2_continue_generate_with_position_ids",
197
+ "commit": null,
198
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
199
+ "pr_number": null,
200
+ "author": null,
201
+ "merged_by": null,
202
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426980245"
203
+ },
204
+ {
205
+ "test": "tests/models/deepseek_vl_hybrid/test_modeling_deepseek_vl_hybrid.py::DeepseekVLHybridModelTest::test_flash_attn_2_fp32_ln",
206
+ "commit": null,
207
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
208
+ "pr_number": null,
209
+ "author": null,
210
+ "merged_by": null,
211
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426980245"
212
+ },
213
+ {
214
+ "test": "tests/models/deepseek_vl_hybrid/test_modeling_deepseek_vl_hybrid.py::DeepseekVLHybridModelTest::test_flash_attn_2_from_config",
215
+ "commit": null,
216
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
217
+ "pr_number": null,
218
+ "author": null,
219
+ "merged_by": null,
220
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426980245"
221
+ }
222
+ ]
223
+ },
224
+ "diffllama": {
225
+ "single-gpu": [
226
+ {
227
+ "test": "tests/models/diffllama/test_modeling_diffllama.py::DiffLlamaModelTest::test_flash_attn_2_generate_padding_right",
228
+ "commit": null,
229
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
230
+ "pr_number": null,
231
+ "author": null,
232
+ "merged_by": null,
233
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426980397"
234
+ },
235
+ {
236
+ "test": "tests/models/diffllama/test_modeling_diffllama.py::DiffLlamaModelTest::test_flash_attn_2_inference_equivalence",
237
+ "commit": null,
238
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
239
+ "pr_number": null,
240
+ "author": null,
241
+ "merged_by": null,
242
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426980397"
243
+ },
244
+ {
245
+ "test": "tests/models/diffllama/test_modeling_diffllama.py::DiffLlamaModelTest::test_flash_attn_2_inference_equivalence_right_padding",
246
+ "commit": null,
247
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
248
+ "pr_number": null,
249
+ "author": null,
250
+ "merged_by": null,
251
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426980397"
252
+ }
253
+ ]
254
+ },
255
+ "donut": {
256
+ "single-gpu": [
257
+ {
258
+ "test": "tests/models/donut/test_modeling_donut_swin.py::DonutSwinModelTest::test_flash_attn_2_inference_equivalence",
259
+ "commit": null,
260
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
261
+ "pr_number": null,
262
+ "author": null,
263
+ "merged_by": null,
264
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426980556"
265
+ },
266
+ {
267
+ "test": "tests/models/donut/test_modeling_donut_swin.py::DonutSwinModelTest::test_flash_attn_2_inference_equivalence_right_padding",
268
+ "commit": null,
269
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
270
+ "pr_number": null,
271
+ "author": null,
272
+ "merged_by": null,
273
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426980556"
274
+ }
275
+ ]
276
+ },
277
+ "efficientloftr": {
278
+ "single-gpu": [
279
+ {
280
+ "test": "tests/models/efficientloftr/test_modeling_efficientloftr.py::EfficientLoFTRModelTest::test_flash_attn_2_inference_equivalence",
281
+ "commit": null,
282
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
283
+ "pr_number": null,
284
+ "author": null,
285
+ "merged_by": null,
286
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426980690"
287
+ },
288
+ {
289
+ "test": "tests/models/efficientloftr/test_modeling_efficientloftr.py::EfficientLoFTRModelTest::test_flash_attn_2_inference_equivalence_right_padding",
290
+ "commit": null,
291
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
292
+ "pr_number": null,
293
+ "author": null,
294
+ "merged_by": null,
295
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426980690"
296
+ }
297
+ ]
298
+ },
299
+ "electra": {
300
+ "single-gpu": [
301
+ {
302
+ "test": "tests/models/electra/test_modeling_electra.py::ElectraModelTest::test_flash_attn_2_inference_equivalence_right_padding",
303
+ "commit": null,
304
+ "status": "flaky: test fails on the current CI run (commit: 64397a830188a35b5e9056036e13bb6ccfe434d8) but passes during the check.",
305
+ "pr_number": null,
306
+ "author": null,
307
+ "merged_by": null,
308
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426980701"
309
+ }
310
+ ]
311
+ },
312
+ "emu3": {
313
+ "single-gpu": [
314
+ {
315
+ "test": "tests/models/emu3/test_modeling_emu3.py::Emu3Vision2TextModelTest::test_flash_attn_2_inference_equivalence",
316
+ "commit": null,
317
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
318
+ "pr_number": null,
319
+ "author": null,
320
+ "merged_by": null,
321
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426980751"
322
+ },
323
+ {
324
+ "test": "tests/models/emu3/test_modeling_emu3.py::Emu3Vision2TextModelTest::test_flash_attn_2_inference_equivalence_right_padding",
325
+ "commit": null,
326
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
327
+ "pr_number": null,
328
+ "author": null,
329
+ "merged_by": null,
330
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426980751"
331
+ }
332
+ ]
333
+ },
334
+ "exaone4": {
335
+ "single-gpu": [
336
+ {
337
+ "test": "tests/models/exaone4/test_modeling_exaone4.py::Exaone4ModelTest::test_flash_attention_2_padding_matches_padding_free_with_position_ids",
338
+ "commit": null,
339
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
340
+ "pr_number": null,
341
+ "author": null,
342
+ "merged_by": null,
343
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426980911"
344
+ }
345
+ ]
346
+ },
347
+ "falcon": {
348
+ "single-gpu": [
349
+ {
350
+ "test": "tests/models/falcon/test_modeling_falcon.py::FalconModelTest::test_flash_attn_2_inference_equivalence",
351
+ "commit": null,
352
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
353
+ "pr_number": null,
354
+ "author": null,
355
+ "merged_by": null,
356
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426980957"
357
+ },
358
+ {
359
+ "test": "tests/models/falcon/test_modeling_falcon.py::FalconModelTest::test_flash_attn_2_inference_equivalence_right_padding",
360
+ "commit": null,
361
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
362
+ "pr_number": null,
363
+ "author": null,
364
+ "merged_by": null,
365
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426980957"
366
+ }
367
+ ]
368
+ },
369
+ "flex_olmo": {
370
+ "single-gpu": [
371
+ {
372
+ "test": "tests/models/flex_olmo/test_modeling_flex_olmo.py::FlexOlmoModelTest::test_flash_attention_2_padding_matches_padding_free_with_position_ids",
373
+ "commit": null,
374
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
375
+ "pr_number": null,
376
+ "author": null,
377
+ "merged_by": null,
378
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426981028"
379
+ },
380
+ {
381
+ "test": "tests/models/flex_olmo/test_modeling_flex_olmo.py::FlexOlmoModelTest::test_flash_attn_2_can_compile_with_attention_mask_None_without_graph_break",
382
+ "commit": null,
383
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
384
+ "pr_number": null,
385
+ "author": null,
386
+ "merged_by": null,
387
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426981028"
388
+ }
389
+ ]
390
+ },
391
+ "gemma3n": {
392
+ "single-gpu": [
393
+ {
394
+ "test": "tests/models/gemma3n/test_modeling_gemma3n.py::Gemma3nTextModelTest::test_flash_attn_2_equivalence",
395
+ "commit": null,
396
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
397
+ "pr_number": null,
398
+ "author": null,
399
+ "merged_by": null,
400
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426981191"
401
+ },
402
+ {
403
+ "test": "tests/models/gemma3n/test_modeling_gemma3n.py::Gemma3nTextModelTest::test_flash_attn_2_inference_equivalence",
404
+ "commit": null,
405
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
406
+ "pr_number": null,
407
+ "author": null,
408
+ "merged_by": null,
409
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426981191"
410
+ },
411
+ {
412
+ "test": "tests/models/gemma3n/test_modeling_gemma3n.py::Gemma3nTextModelTest::test_flash_attn_2_inference_equivalence_right_padding",
413
+ "commit": null,
414
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
415
+ "pr_number": null,
416
+ "author": null,
417
+ "merged_by": null,
418
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426981191"
419
+ }
420
+ ]
421
+ },
422
+ "glm4_moe": {
423
+ "single-gpu": [
424
+ {
425
+ "test": "tests/models/glm4_moe/test_modeling_glm4_moe.py::Glm4MoeModelTest::test_flash_attn_2_can_compile_with_attention_mask_None_without_graph_break",
426
+ "commit": null,
427
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
428
+ "pr_number": null,
429
+ "author": null,
430
+ "merged_by": null,
431
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426981304"
432
+ }
433
+ ]
434
+ },
435
+ "gpt2": {
436
+ "single-gpu": [
437
+ {
438
+ "test": "tests/models/gpt2/test_modeling_gpt2.py::GPT2ModelLanguageGenerationTest::test_flash_attn_2_generate_padding_left",
439
+ "commit": null,
440
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
441
+ "pr_number": null,
442
+ "author": null,
443
+ "merged_by": null,
444
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426981364"
445
+ }
446
+ ]
447
+ },
448
+ "gpt_oss": {
449
+ "single-gpu": [
450
+ {
451
+ "test": "tests/models/gpt_oss/test_modeling_gpt_oss.py::GptOssModelTest::test_flash_attn_2_inference_equivalence",
452
+ "commit": null,
453
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
454
+ "pr_number": null,
455
+ "author": null,
456
+ "merged_by": null,
457
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426981452"
458
+ },
459
+ {
460
+ "test": "tests/models/gpt_oss/test_modeling_gpt_oss.py::GptOssModelTest::test_flash_attn_2_inference_equivalence_right_padding",
461
+ "commit": null,
462
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
463
+ "pr_number": null,
464
+ "author": null,
465
+ "merged_by": null,
466
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426981452"
467
+ }
468
+ ]
469
+ },
470
+ "granitemoe": {
471
+ "single-gpu": [
472
+ {
473
+ "test": "tests/models/granitemoe/test_modeling_granitemoe.py::GraniteMoeModelTest::test_flash_attention_2_padding_matches_padding_free_with_position_ids_and_fa_kwargs",
474
+ "commit": null,
475
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
476
+ "pr_number": null,
477
+ "author": null,
478
+ "merged_by": null,
479
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426981564"
480
+ }
481
+ ]
482
+ },
483
+ "granitemoehybrid": {
484
+ "single-gpu": [
485
+ {
486
+ "test": "tests/models/granitemoehybrid/test_modeling_granitemoehybrid.py::BambaModelTest::test_flash_attn_2_inference_equivalence",
487
+ "commit": null,
488
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
489
+ "pr_number": null,
490
+ "author": null,
491
+ "merged_by": null,
492
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426981575"
493
+ },
494
+ {
495
+ "test": "tests/models/granitemoehybrid/test_modeling_granitemoehybrid.py::BambaModelTest::test_flash_attn_2_inference_equivalence_right_padding",
496
+ "commit": null,
497
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
498
+ "pr_number": null,
499
+ "author": null,
500
+ "merged_by": null,
501
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426981575"
502
+ },
503
+ {
504
+ "test": "tests/models/granitemoehybrid/test_modeling_granitemoehybrid.py::GraniteMoeHybridModelTest::test_flash_attn_2_inference_equivalence",
505
+ "commit": null,
506
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
507
+ "pr_number": null,
508
+ "author": null,
509
+ "merged_by": null,
510
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426981575"
511
+ },
512
+ {
513
+ "test": "tests/models/granitemoehybrid/test_modeling_granitemoehybrid.py::GraniteMoeHybridModelTest::test_flash_attn_2_inference_equivalence_right_padding",
514
+ "commit": null,
515
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
516
+ "pr_number": null,
517
+ "author": null,
518
+ "merged_by": null,
519
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426981575"
520
+ }
521
+ ]
522
+ },
523
+ "granitemoeshared": {
524
+ "single-gpu": [
525
+ {
526
+ "test": "tests/models/granitemoeshared/test_modeling_granitemoeshared.py::GraniteMoeSharedModelTest::test_flash_attention_2_padding_matches_padding_free_with_position_ids_and_fa_kwargs",
527
+ "commit": null,
528
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
529
+ "pr_number": null,
530
+ "author": null,
531
+ "merged_by": null,
532
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426981628"
533
+ }
534
+ ]
535
+ },
536
+ "grounding_dino": {
537
+ "single-gpu": [
538
+ {
539
+ "test": "tests/models/grounding_dino/test_modeling_grounding_dino.py::GroundingDinoModelTest::test_flash_attn_2_inference_equivalence",
540
+ "commit": null,
541
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
542
+ "pr_number": null,
543
+ "author": null,
544
+ "merged_by": null,
545
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426981661"
546
+ },
547
+ {
548
+ "test": "tests/models/grounding_dino/test_modeling_grounding_dino.py::GroundingDinoModelTest::test_flash_attn_2_inference_equivalence_right_padding",
549
+ "commit": null,
550
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
551
+ "pr_number": null,
552
+ "author": null,
553
+ "merged_by": null,
554
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426981661"
555
+ }
556
+ ]
557
+ },
558
+ "instructblip": {
559
+ "single-gpu": [
560
+ {
561
+ "test": "tests/models/instructblip/test_modeling_instructblip.py::InstructBlipForConditionalGenerationDecoderOnlyTest::test_flash_attn_2_fp32_ln",
562
+ "commit": null,
563
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
564
+ "pr_number": null,
565
+ "author": null,
566
+ "merged_by": null,
567
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426982031"
568
+ },
569
+ {
570
+ "test": "tests/models/instructblip/test_modeling_instructblip.py::InstructBlipForConditionalGenerationDecoderOnlyTest::test_flash_attn_2_from_config",
571
+ "commit": null,
572
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
573
+ "pr_number": null,
574
+ "author": null,
575
+ "merged_by": null,
576
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426982031"
577
+ }
578
+ ]
579
+ },
580
+ "instructblipvideo": {
581
+ "single-gpu": [
582
+ {
583
+ "test": "tests/models/instructblipvideo/test_modeling_instructblipvideo.py::InstructBlipVideoForConditionalGenerationDecoderOnlyTest::test_flash_attn_2_fp32_ln",
584
+ "commit": null,
585
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
586
+ "pr_number": null,
587
+ "author": null,
588
+ "merged_by": null,
589
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426982101"
590
+ },
591
+ {
592
+ "test": "tests/models/instructblipvideo/test_modeling_instructblipvideo.py::InstructBlipVideoForConditionalGenerationDecoderOnlyTest::test_flash_attn_2_from_config",
593
+ "commit": null,
594
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
595
+ "pr_number": null,
596
+ "author": null,
597
+ "merged_by": null,
598
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426982101"
599
+ }
600
+ ]
601
+ },
602
+ "janus": {
603
+ "single-gpu": [
604
+ {
605
+ "test": "tests/models/janus/test_modeling_janus.py::JanusVisionText2TextModelTest::test_flash_attn_2_inference_equivalence",
606
+ "commit": null,
607
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
608
+ "pr_number": null,
609
+ "author": null,
610
+ "merged_by": null,
611
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426982159"
612
+ },
613
+ {
614
+ "test": "tests/models/janus/test_modeling_janus.py::JanusVisionText2TextModelTest::test_flash_attn_2_inference_equivalence_right_padding",
615
+ "commit": null,
616
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
617
+ "pr_number": null,
618
+ "author": null,
619
+ "merged_by": null,
620
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426982159"
621
+ }
622
+ ]
623
+ },
624
+ "jetmoe": {
625
+ "single-gpu": [
626
+ {
627
+ "test": "tests/models/jetmoe/test_modeling_jetmoe.py::JetMoeModelTest::test_flash_attn_2_equivalence",
628
+ "commit": null,
629
+ "status": "flaky: test fails on the current CI run (commit: 64397a830188a35b5e9056036e13bb6ccfe434d8) but passes during the check.",
630
+ "pr_number": null,
631
+ "author": null,
632
+ "merged_by": null,
633
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426982182"
634
+ },
635
+ {
636
+ "test": "tests/models/jetmoe/test_modeling_jetmoe.py::JetMoeModelTest::test_flash_attn_2_fp32_ln",
637
+ "commit": null,
638
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
639
+ "pr_number": null,
640
+ "author": null,
641
+ "merged_by": null,
642
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426982182"
643
+ }
644
+ ]
645
+ },
646
+ "kosmos2": {
647
+ "single-gpu": [
648
+ {
649
+ "test": "tests/models/kosmos2/test_modeling_kosmos2.py::Kosmos2ModelTest::test_eager_matches_fa2_generate",
650
+ "commit": null,
651
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
652
+ "pr_number": null,
653
+ "author": null,
654
+ "merged_by": null,
655
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426982209"
656
+ },
657
+ {
658
+ "test": "tests/models/kosmos2/test_modeling_kosmos2.py::Kosmos2ModelTest::test_flash_attention_2_continue_generate_with_position_ids",
659
+ "commit": null,
660
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
661
+ "pr_number": null,
662
+ "author": null,
663
+ "merged_by": null,
664
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426982209"
665
+ },
666
+ {
667
+ "test": "tests/models/kosmos2/test_modeling_kosmos2.py::Kosmos2ModelTest::test_flash_attn_2_can_dispatch_composite_models",
668
+ "commit": null,
669
+ "status": "flaky: test fails on the current CI run (commit: 64397a830188a35b5e9056036e13bb6ccfe434d8) but passes during the check.",
670
+ "pr_number": null,
671
+ "author": null,
672
+ "merged_by": null,
673
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426982209"
674
+ },
675
+ {
676
+ "test": "tests/models/kosmos2/test_modeling_kosmos2.py::Kosmos2ModelTest::test_flash_attn_2_fp32_ln",
677
+ "commit": null,
678
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
679
+ "pr_number": null,
680
+ "author": null,
681
+ "merged_by": null,
682
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426982209"
683
+ },
684
+ {
685
+ "test": "tests/models/kosmos2/test_modeling_kosmos2.py::Kosmos2ModelTest::test_flash_attn_2_from_config",
686
+ "commit": null,
687
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
688
+ "pr_number": null,
689
+ "author": null,
690
+ "merged_by": null,
691
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426982209"
692
+ },
693
+ {
694
+ "test": "tests/models/kosmos2/test_modeling_kosmos2.py::Kosmos2ModelTest::test_flash_attn_2_inference_equivalence",
695
+ "commit": null,
696
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
697
+ "pr_number": null,
698
+ "author": null,
699
+ "merged_by": null,
700
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426982209"
701
+ },
702
+ {
703
+ "test": "tests/models/kosmos2/test_modeling_kosmos2.py::Kosmos2ModelTest::test_flash_attn_2_inference_equivalence_right_padding",
704
+ "commit": null,
705
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
706
+ "pr_number": null,
707
+ "author": null,
708
+ "merged_by": null,
709
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426982209"
710
+ }
711
+ ]
712
+ },
713
+ "kosmos2_5": {
714
+ "single-gpu": [
715
+ {
716
+ "test": "tests/models/kosmos2_5/test_modeling_kosmos2_5.py::Kosmos2_5ModelTest::test_flash_attn_2_can_dispatch_composite_models",
717
+ "commit": null,
718
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
719
+ "pr_number": null,
720
+ "author": null,
721
+ "merged_by": null,
722
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426982199"
723
+ }
724
+ ]
725
+ },
726
+ "kyutai_speech_to_text": {
727
+ "single-gpu": [
728
+ {
729
+ "test": "tests/models/kyutai_speech_to_text/test_modeling_kyutai_speech_to_text.py::KyutaiSpeechToTextModelTest::test_eager_matches_fa2_generate",
730
+ "commit": null,
731
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
732
+ "pr_number": null,
733
+ "author": null,
734
+ "merged_by": null,
735
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426982201"
736
+ },
737
+ {
738
+ "test": "tests/models/kyutai_speech_to_text/test_modeling_kyutai_speech_to_text.py::KyutaiSpeechToTextModelTest::test_flash_attn_2_inference_equivalence",
739
+ "commit": null,
740
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
741
+ "pr_number": null,
742
+ "author": null,
743
+ "merged_by": null,
744
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426982201"
745
+ },
746
+ {
747
+ "test": "tests/models/kyutai_speech_to_text/test_modeling_kyutai_speech_to_text.py::KyutaiSpeechToTextModelTest::test_flash_attn_2_inference_equivalence_right_padding",
748
+ "commit": null,
749
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
750
+ "pr_number": null,
751
+ "author": null,
752
+ "merged_by": null,
753
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426982201"
754
+ }
755
+ ]
756
+ },
757
+ "lfm2": {
758
+ "single-gpu": [
759
+ {
760
+ "test": "tests/models/lfm2/test_modeling_lfm2.py::Lfm2ModelTest::test_flash_attention_2_padding_matches_padding_free_with_position_ids",
761
+ "commit": null,
762
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
763
+ "pr_number": null,
764
+ "author": null,
765
+ "merged_by": null,
766
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426982376"
767
+ },
768
+ {
769
+ "test": "tests/models/lfm2/test_modeling_lfm2.py::Lfm2ModelTest::test_flash_attn_2_can_compile_with_attention_mask_None_without_graph_break",
770
+ "commit": null,
771
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
772
+ "pr_number": null,
773
+ "author": null,
774
+ "merged_by": null,
775
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426982376"
776
+ }
777
+ ]
778
+ },
779
+ "lfm2_moe": {
780
+ "single-gpu": [
781
+ {
782
+ "test": "tests/models/lfm2_moe/test_modeling_lfm2_moe.py::Lfm2MoeModelTest::test_flash_attention_2_padding_matches_padding_free_with_position_ids",
783
+ "commit": null,
784
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
785
+ "pr_number": null,
786
+ "author": null,
787
+ "merged_by": null,
788
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426982369"
789
+ },
790
+ {
791
+ "test": "tests/models/lfm2_moe/test_modeling_lfm2_moe.py::Lfm2MoeModelTest::test_flash_attn_2_can_compile_with_attention_mask_None_without_graph_break",
792
+ "commit": null,
793
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
794
+ "pr_number": null,
795
+ "author": null,
796
+ "merged_by": null,
797
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426982369"
798
+ }
799
+ ]
800
+ },
801
+ "lfm2_vl": {
802
+ "single-gpu": [
803
+ {
804
+ "test": "tests/models/lfm2_vl/test_modeling_lfm2_vl.py::Lfm2VlModelTest::test_flash_attn_2_inference_equivalence",
805
+ "commit": null,
806
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
807
+ "pr_number": null,
808
+ "author": null,
809
+ "merged_by": null,
810
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426982367"
811
+ },
812
+ {
813
+ "test": "tests/models/lfm2_vl/test_modeling_lfm2_vl.py::Lfm2VlModelTest::test_flash_attn_2_inference_equivalence_right_padding",
814
+ "commit": null,
815
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
816
+ "pr_number": null,
817
+ "author": null,
818
+ "merged_by": null,
819
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426982367"
820
+ }
821
+ ]
822
+ },
823
+ "llava_next": {
824
+ "single-gpu": [
825
+ {
826
+ "test": "tests/models/llava_next/test_modeling_llava_next.py::LlavaNextForConditionalGenerationModelTest::test_flash_attn_2_inference_equivalence",
827
+ "commit": null,
828
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
829
+ "pr_number": null,
830
+ "author": null,
831
+ "merged_by": null,
832
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426982498"
833
+ },
834
+ {
835
+ "test": "tests/models/llava_next/test_modeling_llava_next.py::LlavaNextForConditionalGenerationModelTest::test_flash_attn_2_inference_equivalence_right_padding",
836
+ "commit": null,
837
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
838
+ "pr_number": null,
839
+ "author": null,
840
+ "merged_by": null,
841
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426982498"
842
+ }
843
+ ]
844
+ },
845
+ "llava_next_video": {
846
+ "single-gpu": [
847
+ {
848
+ "test": "tests/models/llava_next_video/test_modeling_llava_next_video.py::LlavaNextVideoForConditionalGenerationModelTest::test_flash_attn_2_inference_equivalence",
849
+ "commit": null,
850
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
851
+ "pr_number": null,
852
+ "author": null,
853
+ "merged_by": null,
854
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426982511"
855
+ },
856
+ {
857
+ "test": "tests/models/llava_next_video/test_modeling_llava_next_video.py::LlavaNextVideoForConditionalGenerationModelTest::test_flash_attn_2_inference_equivalence_right_padding",
858
+ "commit": null,
859
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
860
+ "pr_number": null,
861
+ "author": null,
862
+ "merged_by": null,
863
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426982511"
864
+ }
865
+ ]
866
+ },
867
+ "llava_onevision": {
868
+ "single-gpu": [
869
+ {
870
+ "test": "tests/models/llava_onevision/test_modeling_llava_onevision.py::LlavaOnevisionForConditionalGenerationModelTest::test_flash_attn_2_inference_equivalence",
871
+ "commit": null,
872
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
873
+ "pr_number": null,
874
+ "author": null,
875
+ "merged_by": null,
876
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426982547"
877
+ },
878
+ {
879
+ "test": "tests/models/llava_onevision/test_modeling_llava_onevision.py::LlavaOnevisionForConditionalGenerationModelTest::test_flash_attn_2_inference_equivalence_right_padding",
880
+ "commit": null,
881
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
882
+ "pr_number": null,
883
+ "author": null,
884
+ "merged_by": null,
885
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426982547"
886
+ }
887
+ ]
888
+ },
889
+ "mask2former": {
890
+ "single-gpu": [
891
+ {
892
+ "test": "tests/models/mask2former/test_modeling_mask2former.py::Mask2FormerModelTest::test_flash_attn_2_inference_equivalence",
893
+ "commit": null,
894
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
895
+ "pr_number": null,
896
+ "author": null,
897
+ "merged_by": null,
898
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426982762"
899
+ },
900
+ {
901
+ "test": "tests/models/mask2former/test_modeling_mask2former.py::Mask2FormerModelTest::test_flash_attn_2_inference_equivalence_right_padding",
902
+ "commit": null,
903
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
904
+ "pr_number": null,
905
+ "author": null,
906
+ "merged_by": null,
907
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426982762"
908
+ }
909
+ ]
910
+ },
911
+ "maskformer": {
912
+ "single-gpu": [
913
+ {
914
+ "test": "tests/models/maskformer/test_modeling_maskformer.py::MaskFormerModelTest::test_flash_attn_2_inference_equivalence",
915
+ "commit": null,
916
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
917
+ "pr_number": null,
918
+ "author": null,
919
+ "merged_by": null,
920
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426982746"
921
+ },
922
+ {
923
+ "test": "tests/models/maskformer/test_modeling_maskformer.py::MaskFormerModelTest::test_flash_attn_2_inference_equivalence_right_padding",
924
+ "commit": null,
925
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
926
+ "pr_number": null,
927
+ "author": null,
928
+ "merged_by": null,
929
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426982746"
930
+ },
931
+ {
932
+ "test": "tests/models/maskformer/test_modeling_maskformer_swin.py::MaskFormerSwinModelTest::test_flash_attn_2_inference_equivalence",
933
+ "commit": null,
934
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
935
+ "pr_number": null,
936
+ "author": null,
937
+ "merged_by": null,
938
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426982746"
939
+ },
940
+ {
941
+ "test": "tests/models/maskformer/test_modeling_maskformer_swin.py::MaskFormerSwinModelTest::test_flash_attn_2_inference_equivalence_right_padding",
942
+ "commit": null,
943
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
944
+ "pr_number": null,
945
+ "author": null,
946
+ "merged_by": null,
947
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426982746"
948
+ }
949
+ ]
950
+ },
951
+ "mixtral": {
952
+ "single-gpu": [
953
+ {
954
+ "test": "tests/models/mixtral/test_modeling_mixtral.py::MixtralModelTest::test_flash_attn_2_equivalence",
955
+ "commit": null,
956
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
957
+ "pr_number": null,
958
+ "author": null,
959
+ "merged_by": null,
960
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426978950"
961
+ }
962
+ ]
963
+ },
964
+ "mllama": {
965
+ "single-gpu": [
966
+ {
967
+ "test": "tests/models/mllama/test_modeling_mllama.py::MllamaForCausalLMModelTest::test_flash_attention_2_padding_matches_padding_free_with_position_ids",
968
+ "commit": null,
969
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
970
+ "pr_number": null,
971
+ "author": null,
972
+ "merged_by": null,
973
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426978963"
974
+ },
975
+ {
976
+ "test": "tests/models/mllama/test_modeling_mllama.py::MllamaForConditionalGenerationModelTest::test_eager_matches_fa2_generate",
977
+ "commit": null,
978
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
979
+ "pr_number": null,
980
+ "author": null,
981
+ "merged_by": null,
982
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426978963"
983
+ },
984
+ {
985
+ "test": "tests/models/mllama/test_modeling_mllama.py::MllamaForConditionalGenerationModelTest::test_flash_attn_2_inference_equivalence",
986
+ "commit": null,
987
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
988
+ "pr_number": null,
989
+ "author": null,
990
+ "merged_by": null,
991
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426978963"
992
+ },
993
+ {
994
+ "test": "tests/models/mllama/test_modeling_mllama.py::MllamaForConditionalGenerationModelTest::test_flash_attn_2_inference_equivalence_right_padding",
995
+ "commit": null,
996
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
997
+ "pr_number": null,
998
+ "author": null,
999
+ "merged_by": null,
1000
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426978963"
1001
+ }
1002
+ ]
1003
+ },
1004
+ "mm_grounding_dino": {
1005
+ "single-gpu": [
1006
+ {
1007
+ "test": "tests/models/mm_grounding_dino/test_modeling_mm_grounding_dino.py::MMGroundingDinoModelTest::test_flash_attn_2_inference_equivalence",
1008
+ "commit": null,
1009
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
1010
+ "pr_number": null,
1011
+ "author": null,
1012
+ "merged_by": null,
1013
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426978976"
1014
+ },
1015
+ {
1016
+ "test": "tests/models/mm_grounding_dino/test_modeling_mm_grounding_dino.py::MMGroundingDinoModelTest::test_flash_attn_2_inference_equivalence_right_padding",
1017
+ "commit": null,
1018
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
1019
+ "pr_number": null,
1020
+ "author": null,
1021
+ "merged_by": null,
1022
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426978976"
1023
+ }
1024
+ ]
1025
+ },
1026
+ "moshi": {
1027
+ "single-gpu": [
1028
+ {
1029
+ "test": "tests/models/moshi/test_modeling_moshi.py::MoshiDecoderTest::test_flash_attn_2_inference_equivalence",
1030
+ "commit": null,
1031
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
1032
+ "pr_number": null,
1033
+ "author": null,
1034
+ "merged_by": null,
1035
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426979243"
1036
+ },
1037
+ {
1038
+ "test": "tests/models/moshi/test_modeling_moshi.py::MoshiDecoderTest::test_flash_attn_2_inference_equivalence_right_padding",
1039
+ "commit": null,
1040
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
1041
+ "pr_number": null,
1042
+ "author": null,
1043
+ "merged_by": null,
1044
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426979243"
1045
+ },
1046
+ {
1047
+ "test": "tests/models/moshi/test_modeling_moshi.py::MoshiTest::test_eager_matches_fa2_generate",
1048
+ "commit": null,
1049
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
1050
+ "pr_number": null,
1051
+ "author": null,
1052
+ "merged_by": null,
1053
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426979243"
1054
+ },
1055
+ {
1056
+ "test": "tests/models/moshi/test_modeling_moshi.py::MoshiTest::test_flash_attn_2_fp32_ln",
1057
+ "commit": null,
1058
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
1059
+ "pr_number": null,
1060
+ "author": null,
1061
+ "merged_by": null,
1062
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426979243"
1063
+ },
1064
+ {
1065
+ "test": "tests/models/moshi/test_modeling_moshi.py::MoshiTest::test_flash_attn_2_from_config",
1066
+ "commit": null,
1067
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
1068
+ "pr_number": null,
1069
+ "author": null,
1070
+ "merged_by": null,
1071
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426979243"
1072
+ },
1073
+ {
1074
+ "test": "tests/models/moshi/test_modeling_moshi.py::MoshiTest::test_flash_attn_2_inference_equivalence",
1075
+ "commit": null,
1076
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
1077
+ "pr_number": null,
1078
+ "author": null,
1079
+ "merged_by": null,
1080
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426979243"
1081
+ },
1082
+ {
1083
+ "test": "tests/models/moshi/test_modeling_moshi.py::MoshiTest::test_flash_attn_2_inference_equivalence_right_padding",
1084
+ "commit": null,
1085
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
1086
+ "pr_number": null,
1087
+ "author": null,
1088
+ "merged_by": null,
1089
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426979243"
1090
+ }
1091
+ ]
1092
+ },
1093
+ "nemotron": {
1094
+ "single-gpu": [
1095
+ {
1096
+ "test": "tests/models/nemotron/test_modeling_nemotron.py::NemotronModelTest::test_flash_attn_2_equivalence",
1097
+ "commit": null,
1098
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
1099
+ "pr_number": null,
1100
+ "author": null,
1101
+ "merged_by": null,
1102
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426979489"
1103
+ }
1104
+ ]
1105
+ },
1106
+ "olmo": {
1107
+ "single-gpu": [
1108
+ {
1109
+ "test": "tests/models/olmo/test_modeling_olmo.py::OlmoModelTest::test_flash_attention_2_padding_matches_padding_free_with_position_ids",
1110
+ "commit": null,
1111
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
1112
+ "pr_number": null,
1113
+ "author": null,
1114
+ "merged_by": null,
1115
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426979493"
1116
+ }
1117
+ ]
1118
+ },
1119
+ "olmo2": {
1120
+ "single-gpu": [
1121
+ {
1122
+ "test": "tests/models/olmo2/test_modeling_olmo2.py::Olmo2ModelTest::test_flash_attention_2_padding_matches_padding_free_with_position_ids",
1123
+ "commit": null,
1124
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
1125
+ "pr_number": null,
1126
+ "author": null,
1127
+ "merged_by": null,
1128
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426979496"
1129
+ }
1130
+ ]
1131
+ },
1132
+ "omdet_turbo": {
1133
+ "single-gpu": [
1134
+ {
1135
+ "test": "tests/models/omdet_turbo/test_modeling_omdet_turbo.py::OmDetTurboModelTest::test_flash_attn_2_inference_equivalence",
1136
+ "commit": null,
1137
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
1138
+ "pr_number": null,
1139
+ "author": null,
1140
+ "merged_by": null,
1141
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426979661"
1142
+ },
1143
+ {
1144
+ "test": "tests/models/omdet_turbo/test_modeling_omdet_turbo.py::OmDetTurboModelTest::test_flash_attn_2_inference_equivalence_right_padding",
1145
+ "commit": null,
1146
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
1147
+ "pr_number": null,
1148
+ "author": null,
1149
+ "merged_by": null,
1150
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426979661"
1151
+ }
1152
+ ]
1153
+ },
1154
+ "oneformer": {
1155
+ "single-gpu": [
1156
+ {
1157
+ "test": "tests/models/oneformer/test_modeling_oneformer.py::OneFormerModelTest::test_flash_attn_2_inference_equivalence",
1158
+ "commit": null,
1159
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
1160
+ "pr_number": null,
1161
+ "author": null,
1162
+ "merged_by": null,
1163
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426979671"
1164
+ },
1165
+ {
1166
+ "test": "tests/models/oneformer/test_modeling_oneformer.py::OneFormerModelTest::test_flash_attn_2_inference_equivalence_right_padding",
1167
+ "commit": null,
1168
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
1169
+ "pr_number": null,
1170
+ "author": null,
1171
+ "merged_by": null,
1172
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426979671"
1173
+ }
1174
+ ]
1175
+ },
1176
+ "paligemma": {
1177
+ "single-gpu": [
1178
+ {
1179
+ "test": "tests/models/paligemma/test_modeling_paligemma.py::PaliGemmaForConditionalGenerationModelTest::test_flash_attn_2_from_config",
1180
+ "commit": null,
1181
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
1182
+ "pr_number": null,
1183
+ "author": null,
1184
+ "merged_by": null,
1185
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426979663"
1186
+ },
1187
+ {
1188
+ "test": "tests/models/paligemma/test_modeling_paligemma.py::PaliGemmaForConditionalGenerationModelTest::test_flash_attn_2_inference_equivalence",
1189
+ "commit": null,
1190
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
1191
+ "pr_number": null,
1192
+ "author": null,
1193
+ "merged_by": null,
1194
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426979663"
1195
+ },
1196
+ {
1197
+ "test": "tests/models/paligemma/test_modeling_paligemma.py::PaliGemmaForConditionalGenerationModelTest::test_flash_attn_2_inference_equivalence_right_padding",
1198
+ "commit": null,
1199
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
1200
+ "pr_number": null,
1201
+ "author": null,
1202
+ "merged_by": null,
1203
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426979663"
1204
+ }
1205
+ ]
1206
+ },
1207
+ "paligemma2": {
1208
+ "single-gpu": [
1209
+ {
1210
+ "test": "tests/models/paligemma2/test_modeling_paligemma2.py::PaliGemma2ForConditionalGenerationModelTest::test_flash_attn_2_from_config",
1211
+ "commit": null,
1212
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
1213
+ "pr_number": null,
1214
+ "author": null,
1215
+ "merged_by": null,
1216
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426979705"
1217
+ },
1218
+ {
1219
+ "test": "tests/models/paligemma2/test_modeling_paligemma2.py::PaliGemma2ForConditionalGenerationModelTest::test_flash_attn_2_inference_equivalence",
1220
+ "commit": null,
1221
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
1222
+ "pr_number": null,
1223
+ "author": null,
1224
+ "merged_by": null,
1225
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426979705"
1226
+ },
1227
+ {
1228
+ "test": "tests/models/paligemma2/test_modeling_paligemma2.py::PaliGemma2ForConditionalGenerationModelTest::test_flash_attn_2_inference_equivalence_right_padding",
1229
+ "commit": null,
1230
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
1231
+ "pr_number": null,
1232
+ "author": null,
1233
+ "merged_by": null,
1234
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426979705"
1235
+ }
1236
+ ]
1237
+ },
1238
+ "pegasus_x": {
1239
+ "single-gpu": [
1240
+ {
1241
+ "test": "tests/models/pegasus_x/test_modeling_pegasus_x.py::PegasusXModelTest::test_flash_attn_2_inference_equivalence",
1242
+ "commit": null,
1243
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
1244
+ "pr_number": null,
1245
+ "author": null,
1246
+ "merged_by": null,
1247
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426979825"
1248
+ },
1249
+ {
1250
+ "test": "tests/models/pegasus_x/test_modeling_pegasus_x.py::PegasusXModelTest::test_flash_attn_2_inference_equivalence_right_padding",
1251
+ "commit": null,
1252
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
1253
+ "pr_number": null,
1254
+ "author": null,
1255
+ "merged_by": null,
1256
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426979825"
1257
+ },
1258
+ {
1259
+ "test": "tests/models/pegasus_x/test_modeling_pegasus_x.py::PegasusXStandaloneDecoderModelTest::test_flash_attn_2_inference_equivalence",
1260
+ "commit": null,
1261
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
1262
+ "pr_number": null,
1263
+ "author": null,
1264
+ "merged_by": null,
1265
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426979825"
1266
+ },
1267
+ {
1268
+ "test": "tests/models/pegasus_x/test_modeling_pegasus_x.py::PegasusXStandaloneDecoderModelTest::test_flash_attn_2_inference_equivalence_right_padding",
1269
+ "commit": null,
1270
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
1271
+ "pr_number": null,
1272
+ "author": null,
1273
+ "merged_by": null,
1274
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426979825"
1275
+ }
1276
+ ]
1277
+ },
1278
+ "perception_lm": {
1279
+ "single-gpu": [
1280
+ {
1281
+ "test": "tests/models/perception_lm/test_modeling_perception_lm.py::PerceptionLMForConditionalGenerationModelTest::test_flash_attention_2_continue_generate_with_position_ids",
1282
+ "commit": null,
1283
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
1284
+ "pr_number": null,
1285
+ "author": null,
1286
+ "merged_by": null,
1287
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426979850"
1288
+ }
1289
+ ]
1290
+ },
1291
+ "phi": {
1292
+ "single-gpu": [
1293
+ {
1294
+ "test": "tests/models/phi/test_modeling_phi.py::PhiModelTest::test_flash_attention_2_padding_matches_padding_free_with_position_ids",
1295
+ "commit": null,
1296
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
1297
+ "pr_number": null,
1298
+ "author": null,
1299
+ "merged_by": null,
1300
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426979858"
1301
+ }
1302
+ ]
1303
+ },
1304
+ "phimoe": {
1305
+ "single-gpu": [
1306
+ {
1307
+ "test": "tests/models/phimoe/test_modeling_phimoe.py::PhimoeModelTest::test_flash_attn_2_equivalence",
1308
+ "commit": null,
1309
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
1310
+ "pr_number": null,
1311
+ "author": null,
1312
+ "merged_by": null,
1313
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426979979"
1314
+ }
1315
+ ]
1316
+ },
1317
+ "pixtral": {
1318
+ "single-gpu": [
1319
+ {
1320
+ "test": "tests/models/pixtral/test_modeling_pixtral.py::PixtralVisionModelModelTest::test_flash_attn_2_inference_equivalence",
1321
+ "commit": null,
1322
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
1323
+ "pr_number": null,
1324
+ "author": null,
1325
+ "merged_by": null,
1326
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426980007"
1327
+ },
1328
+ {
1329
+ "test": "tests/models/pixtral/test_modeling_pixtral.py::PixtralVisionModelModelTest::test_flash_attn_2_inference_equivalence_right_padding",
1330
+ "commit": null,
1331
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
1332
+ "pr_number": null,
1333
+ "author": null,
1334
+ "merged_by": null,
1335
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426980007"
1336
+ }
1337
+ ]
1338
+ },
1339
+ "qwen2_5_vl": {
1340
+ "single-gpu": [
1341
+ {
1342
+ "test": "tests/models/qwen2_5_vl/test_modeling_qwen2_5_vl.py::Qwen2_5_VLIntegrationTest::test_small_model_integration_test_batch_wo_image_flashatt2",
1343
+ "commit": null,
1344
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
1345
+ "pr_number": null,
1346
+ "author": null,
1347
+ "merged_by": null,
1348
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426980179"
1349
+ }
1350
+ ]
1351
+ },
1352
+ "qwen3_omni_moe": {
1353
+ "single-gpu": [
1354
+ {
1355
+ "test": "tests/models/qwen3_omni_moe/test_modeling_qwen3_omni_moe.py::Qwen2_5OmniThinkerForConditionalGenerationModelTest::test_flash_attention_2_padding_matches_padding_free_with_position_ids",
1356
+ "commit": null,
1357
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
1358
+ "pr_number": null,
1359
+ "author": null,
1360
+ "merged_by": null,
1361
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426980332"
1362
+ },
1363
+ {
1364
+ "test": "tests/models/qwen3_omni_moe/test_modeling_qwen3_omni_moe.py::Qwen2_5OmniThinkerForConditionalGenerationModelTest::test_flash_attention_2_padding_matches_padding_free_with_position_ids_and_fa_kwargs",
1365
+ "commit": null,
1366
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
1367
+ "pr_number": null,
1368
+ "author": null,
1369
+ "merged_by": null,
1370
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426980332"
1371
+ },
1372
+ {
1373
+ "test": "tests/models/qwen3_omni_moe/test_modeling_qwen3_omni_moe.py::Qwen2_5OmniModelIntegrationTest::test_small_model_integration_test_batch_flashatt2",
1374
+ "commit": null,
1375
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
1376
+ "pr_number": null,
1377
+ "author": null,
1378
+ "merged_by": null,
1379
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426980332"
1380
+ }
1381
+ ]
1382
+ },
1383
+ "roberta_prelayernorm": {
1384
+ "single-gpu": [
1385
+ {
1386
+ "test": "tests/models/roberta_prelayernorm/test_modeling_roberta_prelayernorm.py::RobertaPreLayerNormModelTest::test_flash_attn_2_inference_equivalence_right_padding",
1387
+ "commit": null,
1388
+ "status": "flaky: test fails on the current CI run (commit: 64397a830188a35b5e9056036e13bb6ccfe434d8) but passes during the check.",
1389
+ "pr_number": null,
1390
+ "author": null,
1391
+ "merged_by": null,
1392
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426980538"
1393
+ }
1394
+ ]
1395
+ },
1396
+ "sam2": {
1397
+ "single-gpu": [
1398
+ {
1399
+ "test": "tests/models/sam2/test_modeling_sam2.py::Sam2ModelTest::test_flash_attn_2_can_dispatch_composite_models",
1400
+ "commit": null,
1401
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
1402
+ "pr_number": null,
1403
+ "author": null,
1404
+ "merged_by": null,
1405
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426980709"
1406
+ }
1407
+ ]
1408
+ },
1409
+ "smollm3": {
1410
+ "single-gpu": [
1411
+ {
1412
+ "test": "tests/models/smollm3/test_modeling_smollm3.py::SmolLM3IntegrationTest::test_model_3b_long_prompt",
1413
+ "commit": null,
1414
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
1415
+ "pr_number": null,
1416
+ "author": null,
1417
+ "merged_by": null,
1418
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426980932"
1419
+ }
1420
+ ]
1421
+ },
1422
+ "squeezebert": {
1423
+ "single-gpu": [
1424
+ {
1425
+ "test": "tests/models/squeezebert/test_modeling_squeezebert.py::SqueezeBertModelTest::test_flash_attn_2_inference_equivalence",
1426
+ "commit": null,
1427
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
1428
+ "pr_number": null,
1429
+ "author": null,
1430
+ "merged_by": null,
1431
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426981010"
1432
+ },
1433
+ {
1434
+ "test": "tests/models/squeezebert/test_modeling_squeezebert.py::SqueezeBertModelTest::test_flash_attn_2_inference_equivalence_right_padding",
1435
+ "commit": null,
1436
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
1437
+ "pr_number": null,
1438
+ "author": null,
1439
+ "merged_by": null,
1440
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426981010"
1441
+ }
1442
+ ]
1443
+ },
1444
+ "starcoder2": {
1445
+ "single-gpu": [
1446
+ {
1447
+ "test": "tests/models/starcoder2/test_modeling_starcoder2.py::Starcoder2IntegrationTest::test_starcoder2_batched_generation_fa2",
1448
+ "commit": null,
1449
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
1450
+ "pr_number": null,
1451
+ "author": null,
1452
+ "merged_by": null,
1453
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426981080"
1454
+ }
1455
+ ]
1456
+ },
1457
+ "swin": {
1458
+ "single-gpu": [
1459
+ {
1460
+ "test": "tests/models/swin/test_modeling_swin.py::SwinModelTest::test_flash_attn_2_inference_equivalence",
1461
+ "commit": null,
1462
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
1463
+ "pr_number": null,
1464
+ "author": null,
1465
+ "merged_by": null,
1466
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426981143"
1467
+ },
1468
+ {
1469
+ "test": "tests/models/swin/test_modeling_swin.py::SwinModelTest::test_flash_attn_2_inference_equivalence_right_padding",
1470
+ "commit": null,
1471
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
1472
+ "pr_number": null,
1473
+ "author": null,
1474
+ "merged_by": null,
1475
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426981143"
1476
+ }
1477
+ ]
1478
+ },
1479
+ "swin2sr": {
1480
+ "single-gpu": [
1481
+ {
1482
+ "test": "tests/models/swin2sr/test_modeling_swin2sr.py::Swin2SRModelTest::test_flash_attn_2_inference_equivalence",
1483
+ "commit": null,
1484
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
1485
+ "pr_number": null,
1486
+ "author": null,
1487
+ "merged_by": null,
1488
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426981134"
1489
+ },
1490
+ {
1491
+ "test": "tests/models/swin2sr/test_modeling_swin2sr.py::Swin2SRModelTest::test_flash_attn_2_inference_equivalence_right_padding",
1492
+ "commit": null,
1493
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
1494
+ "pr_number": null,
1495
+ "author": null,
1496
+ "merged_by": null,
1497
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426981134"
1498
+ }
1499
+ ]
1500
+ },
1501
+ "swinv2": {
1502
+ "single-gpu": [
1503
+ {
1504
+ "test": "tests/models/swinv2/test_modeling_swinv2.py::Swinv2ModelTest::test_flash_attn_2_inference_equivalence",
1505
+ "commit": null,
1506
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
1507
+ "pr_number": null,
1508
+ "author": null,
1509
+ "merged_by": null,
1510
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426981141"
1511
+ },
1512
+ {
1513
+ "test": "tests/models/swinv2/test_modeling_swinv2.py::Swinv2ModelTest::test_flash_attn_2_inference_equivalence_right_padding",
1514
+ "commit": null,
1515
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
1516
+ "pr_number": null,
1517
+ "author": null,
1518
+ "merged_by": null,
1519
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426981141"
1520
+ }
1521
+ ]
1522
+ },
1523
+ "t5gemma": {
1524
+ "single-gpu": [
1525
+ {
1526
+ "test": "tests/models/t5gemma/test_modeling_t5gemma.py::T5GemmaModelTest::test_flash_attn_2_can_compile_with_attention_mask_None_without_graph_break",
1527
+ "commit": null,
1528
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
1529
+ "pr_number": null,
1530
+ "author": null,
1531
+ "merged_by": null,
1532
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426981212"
1533
+ }
1534
+ ]
1535
+ },
1536
+ "zamba": {
1537
+ "single-gpu": [
1538
+ {
1539
+ "test": "tests/models/zamba/test_modeling_zamba.py::ZambaModelTest::test_flash_attn_2_fp32_ln",
1540
+ "commit": null,
1541
+ "status": "flaky: test passed in the previous run (commit: b9f90dc388fd415a2ba2a6a31a372f451d4a4eed) but failed (on the same commit) during the check of the current run.",
1542
+ "pr_number": null,
1543
+ "author": null,
1544
+ "merged_by": null,
1545
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19056156009/job/54426982378"
1546
+ }
1547
+ ]
1548
+ }
1549
+ }
1550
+ }