hf-transformers-bot commited on
Commit
b304b22
·
verified ·
1 Parent(s): 57abe81

Upload 2025-11-13/ci_results_run_models_gpu/new_failures_with_bad_commit.json with huggingface_hub

Browse files
2025-11-13/ci_results_run_models_gpu/new_failures_with_bad_commit.json ADDED
@@ -0,0 +1,1536 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "audioflamingo3": {
3
+ "single-gpu": [
4
+ {
5
+ "test": "tests/models/audioflamingo3/test_modeling_audioflamingo3.py::AudioFlamingo3ForConditionalGenerationModelTest::test_eager_matches_fa2_generate",
6
+ "commit": "1709ed96e47f36fe926e2cd2556fa839b41c2afd",
7
+ "status": "git bisect found the bad commit.",
8
+ "pr_number": 40290,
9
+ "author": "lashahub",
10
+ "merged_by": "ydshieh",
11
+ "parent": "fd36275be2f3e56bc20da01f1f320b623b413957",
12
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256262527"
13
+ }
14
+ ]
15
+ },
16
+ "bamba": {
17
+ "single-gpu": [
18
+ {
19
+ "test": "tests/models/bamba/test_modeling_bamba.py::BambaModelTest::test_flash_attn_2_inference_equivalence",
20
+ "commit": null,
21
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
22
+ "pr_number": null,
23
+ "author": null,
24
+ "merged_by": null,
25
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256262537"
26
+ },
27
+ {
28
+ "test": "tests/models/bamba/test_modeling_bamba.py::BambaModelTest::test_flash_attn_2_inference_equivalence_right_padding",
29
+ "commit": null,
30
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
31
+ "pr_number": null,
32
+ "author": null,
33
+ "merged_by": null,
34
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256262537"
35
+ }
36
+ ]
37
+ },
38
+ "bark": {
39
+ "single-gpu": [
40
+ {
41
+ "test": "tests/models/bark/test_modeling_bark.py::BarkSemanticModelTest::test_eager_matches_fa2_generate",
42
+ "commit": null,
43
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
44
+ "pr_number": null,
45
+ "author": null,
46
+ "merged_by": null,
47
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256262543"
48
+ },
49
+ {
50
+ "test": "tests/models/bark/test_modeling_bark.py::BarkSemanticModelTest::test_flash_attention_2_continue_generate_with_position_ids",
51
+ "commit": null,
52
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
53
+ "pr_number": null,
54
+ "author": null,
55
+ "merged_by": null,
56
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256262543"
57
+ },
58
+ {
59
+ "test": "tests/models/bark/test_modeling_bark.py::BarkCoarseModelTest::test_eager_matches_fa2_generate",
60
+ "commit": null,
61
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
62
+ "pr_number": null,
63
+ "author": null,
64
+ "merged_by": null,
65
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256262543"
66
+ },
67
+ {
68
+ "test": "tests/models/bark/test_modeling_bark.py::BarkCoarseModelTest::test_flash_attention_2_continue_generate_with_position_ids",
69
+ "commit": null,
70
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
71
+ "pr_number": null,
72
+ "author": null,
73
+ "merged_by": null,
74
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256262543"
75
+ }
76
+ ]
77
+ },
78
+ "blt": {
79
+ "single-gpu": [
80
+ {
81
+ "test": "tests/models/blt/test_modeling_blt.py::BltModelTest::test_flash_attn_2_can_compile_with_attention_mask_None_without_graph_break",
82
+ "commit": null,
83
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
84
+ "pr_number": null,
85
+ "author": null,
86
+ "merged_by": null,
87
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256263058"
88
+ }
89
+ ]
90
+ },
91
+ "chameleon": {
92
+ "single-gpu": [
93
+ {
94
+ "test": "tests/models/chameleon/test_modeling_chameleon.py::ChameleonModelTest::test_flash_attention_2_padding_matches_padding_free_with_position_ids",
95
+ "commit": null,
96
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
97
+ "pr_number": null,
98
+ "author": null,
99
+ "merged_by": null,
100
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256263264"
101
+ },
102
+ {
103
+ "test": "tests/models/chameleon/test_modeling_chameleon.py::ChameleonVision2SeqModelTest::test_flash_attention_2_padding_matches_padding_free_with_position_ids_and_fa_kwargs",
104
+ "commit": null,
105
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
106
+ "pr_number": null,
107
+ "author": null,
108
+ "merged_by": null,
109
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256263264"
110
+ }
111
+ ]
112
+ },
113
+ "clap": {
114
+ "single-gpu": [
115
+ {
116
+ "test": "tests/models/clap/test_modeling_clap.py::ClapAudioModelTest::test_flash_attn_2_inference_equivalence",
117
+ "commit": null,
118
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
119
+ "pr_number": null,
120
+ "author": null,
121
+ "merged_by": null,
122
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256263271"
123
+ },
124
+ {
125
+ "test": "tests/models/clap/test_modeling_clap.py::ClapAudioModelTest::test_flash_attn_2_inference_equivalence_right_padding",
126
+ "commit": null,
127
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
128
+ "pr_number": null,
129
+ "author": null,
130
+ "merged_by": null,
131
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256263271"
132
+ },
133
+ {
134
+ "test": "tests/models/clap/test_modeling_clap.py::ClapModelTest::test_flash_attn_2_inference_equivalence",
135
+ "commit": null,
136
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
137
+ "pr_number": null,
138
+ "author": null,
139
+ "merged_by": null,
140
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256263271"
141
+ },
142
+ {
143
+ "test": "tests/models/clap/test_modeling_clap.py::ClapModelTest::test_flash_attn_2_inference_equivalence_right_padding",
144
+ "commit": null,
145
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
146
+ "pr_number": null,
147
+ "author": null,
148
+ "merged_by": null,
149
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256263271"
150
+ }
151
+ ]
152
+ },
153
+ "colpali": {
154
+ "single-gpu": [
155
+ {
156
+ "test": "tests/models/colpali/test_modeling_colpali.py::ColPaliForRetrievalModelTest::test_flash_attn_2_inference_equivalence",
157
+ "commit": null,
158
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
159
+ "pr_number": null,
160
+ "author": null,
161
+ "merged_by": null,
162
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256263606"
163
+ },
164
+ {
165
+ "test": "tests/models/colpali/test_modeling_colpali.py::ColPaliForRetrievalModelTest::test_flash_attn_2_inference_equivalence_right_padding",
166
+ "commit": null,
167
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
168
+ "pr_number": null,
169
+ "author": null,
170
+ "merged_by": null,
171
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256263606"
172
+ }
173
+ ]
174
+ },
175
+ "deepseek_v2": {
176
+ "single-gpu": [
177
+ {
178
+ "test": "tests/models/deepseek_v2/test_modeling_deepseek_v2.py::DeepseekV2ModelTest::test_flash_attention_2_padding_matches_padding_free_with_position_ids",
179
+ "commit": null,
180
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
181
+ "pr_number": null,
182
+ "author": null,
183
+ "merged_by": null,
184
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256263996"
185
+ },
186
+ {
187
+ "test": "tests/models/deepseek_v2/test_modeling_deepseek_v2.py::DeepseekV2ModelTest::test_flash_attn_2_can_compile_with_attention_mask_None_without_graph_break",
188
+ "commit": null,
189
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
190
+ "pr_number": null,
191
+ "author": null,
192
+ "merged_by": null,
193
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256263996"
194
+ },
195
+ {
196
+ "test": "tests/models/deepseek_v2/test_modeling_deepseek_v2.py::DeepseekV2ModelTest::test_flash_attn_2_fp32_ln",
197
+ "commit": null,
198
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
199
+ "pr_number": null,
200
+ "author": null,
201
+ "merged_by": null,
202
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256263996"
203
+ }
204
+ ]
205
+ },
206
+ "deepseek_vl_hybrid": {
207
+ "single-gpu": [
208
+ {
209
+ "test": "tests/models/deepseek_vl_hybrid/test_modeling_deepseek_vl_hybrid.py::DeepseekVLHybridModelTest::test_flash_attention_2_continue_generate_with_position_ids",
210
+ "commit": null,
211
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
212
+ "pr_number": null,
213
+ "author": null,
214
+ "merged_by": null,
215
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256264025"
216
+ },
217
+ {
218
+ "test": "tests/models/deepseek_vl_hybrid/test_modeling_deepseek_vl_hybrid.py::DeepseekVLHybridModelTest::test_flash_attn_2_fp32_ln",
219
+ "commit": null,
220
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
221
+ "pr_number": null,
222
+ "author": null,
223
+ "merged_by": null,
224
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256264025"
225
+ },
226
+ {
227
+ "test": "tests/models/deepseek_vl_hybrid/test_modeling_deepseek_vl_hybrid.py::DeepseekVLHybridModelTest::test_flash_attn_2_from_config",
228
+ "commit": null,
229
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
230
+ "pr_number": null,
231
+ "author": null,
232
+ "merged_by": null,
233
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256264025"
234
+ }
235
+ ]
236
+ },
237
+ "diffllama": {
238
+ "single-gpu": [
239
+ {
240
+ "test": "tests/models/diffllama/test_modeling_diffllama.py::DiffLlamaModelTest::test_flash_attn_2_generate_padding_right",
241
+ "commit": null,
242
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
243
+ "pr_number": null,
244
+ "author": null,
245
+ "merged_by": null,
246
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256264149"
247
+ },
248
+ {
249
+ "test": "tests/models/diffllama/test_modeling_diffllama.py::DiffLlamaModelTest::test_flash_attn_2_inference_equivalence",
250
+ "commit": null,
251
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
252
+ "pr_number": null,
253
+ "author": null,
254
+ "merged_by": null,
255
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256264149"
256
+ },
257
+ {
258
+ "test": "tests/models/diffllama/test_modeling_diffllama.py::DiffLlamaModelTest::test_flash_attn_2_inference_equivalence_right_padding",
259
+ "commit": null,
260
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
261
+ "pr_number": null,
262
+ "author": null,
263
+ "merged_by": null,
264
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256264149"
265
+ }
266
+ ]
267
+ },
268
+ "donut": {
269
+ "single-gpu": [
270
+ {
271
+ "test": "tests/models/donut/test_modeling_donut_swin.py::DonutSwinModelTest::test_flash_attn_2_inference_equivalence",
272
+ "commit": null,
273
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
274
+ "pr_number": null,
275
+ "author": null,
276
+ "merged_by": null,
277
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256264372"
278
+ },
279
+ {
280
+ "test": "tests/models/donut/test_modeling_donut_swin.py::DonutSwinModelTest::test_flash_attn_2_inference_equivalence_right_padding",
281
+ "commit": null,
282
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
283
+ "pr_number": null,
284
+ "author": null,
285
+ "merged_by": null,
286
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256264372"
287
+ }
288
+ ]
289
+ },
290
+ "efficientloftr": {
291
+ "single-gpu": [
292
+ {
293
+ "test": "tests/models/efficientloftr/test_modeling_efficientloftr.py::EfficientLoFTRModelTest::test_flash_attn_2_inference_equivalence",
294
+ "commit": null,
295
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
296
+ "pr_number": null,
297
+ "author": null,
298
+ "merged_by": null,
299
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256264426"
300
+ },
301
+ {
302
+ "test": "tests/models/efficientloftr/test_modeling_efficientloftr.py::EfficientLoFTRModelTest::test_flash_attn_2_inference_equivalence_right_padding",
303
+ "commit": null,
304
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
305
+ "pr_number": null,
306
+ "author": null,
307
+ "merged_by": null,
308
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256264426"
309
+ }
310
+ ]
311
+ },
312
+ "emu3": {
313
+ "single-gpu": [
314
+ {
315
+ "test": "tests/models/emu3/test_modeling_emu3.py::Emu3Vision2TextModelTest::test_flash_attn_2_inference_equivalence",
316
+ "commit": null,
317
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
318
+ "pr_number": null,
319
+ "author": null,
320
+ "merged_by": null,
321
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256264594"
322
+ },
323
+ {
324
+ "test": "tests/models/emu3/test_modeling_emu3.py::Emu3Vision2TextModelTest::test_flash_attn_2_inference_equivalence_right_padding",
325
+ "commit": null,
326
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
327
+ "pr_number": null,
328
+ "author": null,
329
+ "merged_by": null,
330
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256264594"
331
+ }
332
+ ]
333
+ },
334
+ "exaone4": {
335
+ "single-gpu": [
336
+ {
337
+ "test": "tests/models/exaone4/test_modeling_exaone4.py::Exaone4ModelTest::test_flash_attention_2_padding_matches_padding_free_with_position_ids",
338
+ "commit": null,
339
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
340
+ "pr_number": null,
341
+ "author": null,
342
+ "merged_by": null,
343
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256264724"
344
+ }
345
+ ]
346
+ },
347
+ "falcon": {
348
+ "single-gpu": [
349
+ {
350
+ "test": "tests/models/falcon/test_modeling_falcon.py::FalconModelTest::test_flash_attn_2_inference_equivalence",
351
+ "commit": null,
352
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
353
+ "pr_number": null,
354
+ "author": null,
355
+ "merged_by": null,
356
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256264803"
357
+ },
358
+ {
359
+ "test": "tests/models/falcon/test_modeling_falcon.py::FalconModelTest::test_flash_attn_2_inference_equivalence_right_padding",
360
+ "commit": null,
361
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
362
+ "pr_number": null,
363
+ "author": null,
364
+ "merged_by": null,
365
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256264803"
366
+ }
367
+ ]
368
+ },
369
+ "flex_olmo": {
370
+ "single-gpu": [
371
+ {
372
+ "test": "tests/models/flex_olmo/test_modeling_flex_olmo.py::FlexOlmoModelTest::test_flash_attention_2_padding_matches_padding_free_with_position_ids",
373
+ "commit": null,
374
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
375
+ "pr_number": null,
376
+ "author": null,
377
+ "merged_by": null,
378
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256264910"
379
+ },
380
+ {
381
+ "test": "tests/models/flex_olmo/test_modeling_flex_olmo.py::FlexOlmoModelTest::test_flash_attention_2_padding_matches_padding_free_with_position_ids_and_fa_kwargs",
382
+ "commit": null,
383
+ "status": "flaky: test fails on the current CI run (commit: ffb35fe14283d61cd5434a32d04d07993e66477a) but passes during the check.",
384
+ "pr_number": null,
385
+ "author": null,
386
+ "merged_by": null,
387
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256264910"
388
+ },
389
+ {
390
+ "test": "tests/models/flex_olmo/test_modeling_flex_olmo.py::FlexOlmoModelTest::test_flash_attn_2_can_compile_with_attention_mask_None_without_graph_break",
391
+ "commit": null,
392
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
393
+ "pr_number": null,
394
+ "author": null,
395
+ "merged_by": null,
396
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256264910"
397
+ }
398
+ ]
399
+ },
400
+ "gemma3n": {
401
+ "single-gpu": [
402
+ {
403
+ "test": "tests/models/gemma3n/test_modeling_gemma3n.py::Gemma3nTextModelTest::test_flash_attn_2_equivalence",
404
+ "commit": null,
405
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
406
+ "pr_number": null,
407
+ "author": null,
408
+ "merged_by": null,
409
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256265090"
410
+ },
411
+ {
412
+ "test": "tests/models/gemma3n/test_modeling_gemma3n.py::Gemma3nTextModelTest::test_flash_attn_2_inference_equivalence",
413
+ "commit": null,
414
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
415
+ "pr_number": null,
416
+ "author": null,
417
+ "merged_by": null,
418
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256265090"
419
+ },
420
+ {
421
+ "test": "tests/models/gemma3n/test_modeling_gemma3n.py::Gemma3nTextModelTest::test_flash_attn_2_inference_equivalence_right_padding",
422
+ "commit": null,
423
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
424
+ "pr_number": null,
425
+ "author": null,
426
+ "merged_by": null,
427
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256265090"
428
+ }
429
+ ]
430
+ },
431
+ "glm4_moe": {
432
+ "single-gpu": [
433
+ {
434
+ "test": "tests/models/glm4_moe/test_modeling_glm4_moe.py::Glm4MoeModelTest::test_flash_attn_2_can_compile_with_attention_mask_None_without_graph_break",
435
+ "commit": null,
436
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
437
+ "pr_number": null,
438
+ "author": null,
439
+ "merged_by": null,
440
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256265226"
441
+ }
442
+ ]
443
+ },
444
+ "gpt_neo": {
445
+ "single-gpu": [
446
+ {
447
+ "test": "tests/models/gpt_neo/test_modeling_gpt_neo.py::GPTNeoModelTest::test_eager_matches_fa2_generate",
448
+ "commit": null,
449
+ "status": "flaky: test fails on the current CI run (commit: ffb35fe14283d61cd5434a32d04d07993e66477a) but passes during the check.",
450
+ "pr_number": null,
451
+ "author": null,
452
+ "merged_by": null,
453
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256265344"
454
+ }
455
+ ]
456
+ },
457
+ "gpt_oss": {
458
+ "single-gpu": [
459
+ {
460
+ "test": "tests/models/gpt_oss/test_modeling_gpt_oss.py::GptOssModelTest::test_flash_attn_2_inference_equivalence",
461
+ "commit": null,
462
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
463
+ "pr_number": null,
464
+ "author": null,
465
+ "merged_by": null,
466
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256265453"
467
+ },
468
+ {
469
+ "test": "tests/models/gpt_oss/test_modeling_gpt_oss.py::GptOssModelTest::test_flash_attn_2_inference_equivalence_right_padding",
470
+ "commit": null,
471
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
472
+ "pr_number": null,
473
+ "author": null,
474
+ "merged_by": null,
475
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256265453"
476
+ }
477
+ ]
478
+ },
479
+ "granitemoe": {
480
+ "single-gpu": [
481
+ {
482
+ "test": "tests/models/granitemoe/test_modeling_granitemoe.py::GraniteMoeModelTest::test_flash_attention_2_padding_matches_padding_free_with_position_ids_and_fa_kwargs",
483
+ "commit": null,
484
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
485
+ "pr_number": null,
486
+ "author": null,
487
+ "merged_by": null,
488
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256265505"
489
+ }
490
+ ]
491
+ },
492
+ "granitemoehybrid": {
493
+ "single-gpu": [
494
+ {
495
+ "test": "tests/models/granitemoehybrid/test_modeling_granitemoehybrid.py::BambaModelTest::test_flash_attn_2_inference_equivalence",
496
+ "commit": null,
497
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
498
+ "pr_number": null,
499
+ "author": null,
500
+ "merged_by": null,
501
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256265535"
502
+ },
503
+ {
504
+ "test": "tests/models/granitemoehybrid/test_modeling_granitemoehybrid.py::BambaModelTest::test_flash_attn_2_inference_equivalence_right_padding",
505
+ "commit": null,
506
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
507
+ "pr_number": null,
508
+ "author": null,
509
+ "merged_by": null,
510
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256265535"
511
+ },
512
+ {
513
+ "test": "tests/models/granitemoehybrid/test_modeling_granitemoehybrid.py::GraniteMoeHybridModelTest::test_flash_attn_2_inference_equivalence",
514
+ "commit": null,
515
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
516
+ "pr_number": null,
517
+ "author": null,
518
+ "merged_by": null,
519
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256265535"
520
+ },
521
+ {
522
+ "test": "tests/models/granitemoehybrid/test_modeling_granitemoehybrid.py::GraniteMoeHybridModelTest::test_flash_attn_2_inference_equivalence_right_padding",
523
+ "commit": null,
524
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
525
+ "pr_number": null,
526
+ "author": null,
527
+ "merged_by": null,
528
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256265535"
529
+ }
530
+ ]
531
+ },
532
+ "granitemoeshared": {
533
+ "single-gpu": [
534
+ {
535
+ "test": "tests/models/granitemoeshared/test_modeling_granitemoeshared.py::GraniteMoeSharedModelTest::test_flash_attention_2_padding_matches_padding_free_with_position_ids_and_fa_kwargs",
536
+ "commit": null,
537
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
538
+ "pr_number": null,
539
+ "author": null,
540
+ "merged_by": null,
541
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256265588"
542
+ }
543
+ ]
544
+ },
545
+ "grounding_dino": {
546
+ "single-gpu": [
547
+ {
548
+ "test": "tests/models/grounding_dino/test_modeling_grounding_dino.py::GroundingDinoModelTest::test_flash_attn_2_inference_equivalence",
549
+ "commit": null,
550
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
551
+ "pr_number": null,
552
+ "author": null,
553
+ "merged_by": null,
554
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256265608"
555
+ },
556
+ {
557
+ "test": "tests/models/grounding_dino/test_modeling_grounding_dino.py::GroundingDinoModelTest::test_flash_attn_2_inference_equivalence_right_padding",
558
+ "commit": null,
559
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
560
+ "pr_number": null,
561
+ "author": null,
562
+ "merged_by": null,
563
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256265608"
564
+ }
565
+ ]
566
+ },
567
+ "instructblip": {
568
+ "single-gpu": [
569
+ {
570
+ "test": "tests/models/instructblip/test_modeling_instructblip.py::InstructBlipForConditionalGenerationDecoderOnlyTest::test_flash_attn_2_fp32_ln",
571
+ "commit": null,
572
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
573
+ "pr_number": null,
574
+ "author": null,
575
+ "merged_by": null,
576
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256265897"
577
+ },
578
+ {
579
+ "test": "tests/models/instructblip/test_modeling_instructblip.py::InstructBlipForConditionalGenerationDecoderOnlyTest::test_flash_attn_2_from_config",
580
+ "commit": null,
581
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
582
+ "pr_number": null,
583
+ "author": null,
584
+ "merged_by": null,
585
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256265897"
586
+ }
587
+ ]
588
+ },
589
+ "instructblipvideo": {
590
+ "single-gpu": [
591
+ {
592
+ "test": "tests/models/instructblipvideo/test_modeling_instructblipvideo.py::InstructBlipVideoForConditionalGenerationDecoderOnlyTest::test_flash_attn_2_fp32_ln",
593
+ "commit": null,
594
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
595
+ "pr_number": null,
596
+ "author": null,
597
+ "merged_by": null,
598
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256265919"
599
+ },
600
+ {
601
+ "test": "tests/models/instructblipvideo/test_modeling_instructblipvideo.py::InstructBlipVideoForConditionalGenerationDecoderOnlyTest::test_flash_attn_2_from_config",
602
+ "commit": null,
603
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
604
+ "pr_number": null,
605
+ "author": null,
606
+ "merged_by": null,
607
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256265919"
608
+ }
609
+ ]
610
+ },
611
+ "janus": {
612
+ "single-gpu": [
613
+ {
614
+ "test": "tests/models/janus/test_modeling_janus.py::JanusVisionText2TextModelTest::test_flash_attn_2_inference_equivalence",
615
+ "commit": null,
616
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
617
+ "pr_number": null,
618
+ "author": null,
619
+ "merged_by": null,
620
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256265979"
621
+ },
622
+ {
623
+ "test": "tests/models/janus/test_modeling_janus.py::JanusVisionText2TextModelTest::test_flash_attn_2_inference_equivalence_right_padding",
624
+ "commit": null,
625
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
626
+ "pr_number": null,
627
+ "author": null,
628
+ "merged_by": null,
629
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256265979"
630
+ }
631
+ ]
632
+ },
633
+ "jetmoe": {
634
+ "single-gpu": [
635
+ {
636
+ "test": "tests/models/jetmoe/test_modeling_jetmoe.py::JetMoeModelTest::test_flash_attn_2_fp32_ln",
637
+ "commit": null,
638
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
639
+ "pr_number": null,
640
+ "author": null,
641
+ "merged_by": null,
642
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256266023"
643
+ }
644
+ ]
645
+ },
646
+ "kosmos2": {
647
+ "single-gpu": [
648
+ {
649
+ "test": "tests/models/kosmos2/test_modeling_kosmos2.py::Kosmos2ModelTest::test_eager_matches_fa2_generate",
650
+ "commit": null,
651
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
652
+ "pr_number": null,
653
+ "author": null,
654
+ "merged_by": null,
655
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256266038"
656
+ },
657
+ {
658
+ "test": "tests/models/kosmos2/test_modeling_kosmos2.py::Kosmos2ModelTest::test_flash_attention_2_continue_generate_with_position_ids",
659
+ "commit": null,
660
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
661
+ "pr_number": null,
662
+ "author": null,
663
+ "merged_by": null,
664
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256266038"
665
+ },
666
+ {
667
+ "test": "tests/models/kosmos2/test_modeling_kosmos2.py::Kosmos2ModelTest::test_flash_attn_2_can_dispatch_composite_models",
668
+ "commit": null,
669
+ "status": "flaky: test fails on the current CI run (commit: ffb35fe14283d61cd5434a32d04d07993e66477a) but passes during the check.",
670
+ "pr_number": null,
671
+ "author": null,
672
+ "merged_by": null,
673
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256266038"
674
+ },
675
+ {
676
+ "test": "tests/models/kosmos2/test_modeling_kosmos2.py::Kosmos2ModelTest::test_flash_attn_2_fp32_ln",
677
+ "commit": null,
678
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
679
+ "pr_number": null,
680
+ "author": null,
681
+ "merged_by": null,
682
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256266038"
683
+ },
684
+ {
685
+ "test": "tests/models/kosmos2/test_modeling_kosmos2.py::Kosmos2ModelTest::test_flash_attn_2_from_config",
686
+ "commit": null,
687
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
688
+ "pr_number": null,
689
+ "author": null,
690
+ "merged_by": null,
691
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256266038"
692
+ },
693
+ {
694
+ "test": "tests/models/kosmos2/test_modeling_kosmos2.py::Kosmos2ModelTest::test_flash_attn_2_inference_equivalence",
695
+ "commit": null,
696
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
697
+ "pr_number": null,
698
+ "author": null,
699
+ "merged_by": null,
700
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256266038"
701
+ },
702
+ {
703
+ "test": "tests/models/kosmos2/test_modeling_kosmos2.py::Kosmos2ModelTest::test_flash_attn_2_inference_equivalence_right_padding",
704
+ "commit": null,
705
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
706
+ "pr_number": null,
707
+ "author": null,
708
+ "merged_by": null,
709
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256266038"
710
+ }
711
+ ]
712
+ },
713
+ "kosmos2_5": {
714
+ "single-gpu": [
715
+ {
716
+ "test": "tests/models/kosmos2_5/test_modeling_kosmos2_5.py::Kosmos2_5ModelTest::test_flash_attn_2_can_dispatch_composite_models",
717
+ "commit": null,
718
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
719
+ "pr_number": null,
720
+ "author": null,
721
+ "merged_by": null,
722
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256266067"
723
+ }
724
+ ]
725
+ },
726
+ "kyutai_speech_to_text": {
727
+ "single-gpu": [
728
+ {
729
+ "test": "tests/models/kyutai_speech_to_text/test_modeling_kyutai_speech_to_text.py::KyutaiSpeechToTextModelTest::test_eager_matches_fa2_generate",
730
+ "commit": null,
731
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
732
+ "pr_number": null,
733
+ "author": null,
734
+ "merged_by": null,
735
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256266060"
736
+ },
737
+ {
738
+ "test": "tests/models/kyutai_speech_to_text/test_modeling_kyutai_speech_to_text.py::KyutaiSpeechToTextModelTest::test_flash_attn_2_inference_equivalence",
739
+ "commit": null,
740
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
741
+ "pr_number": null,
742
+ "author": null,
743
+ "merged_by": null,
744
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256266060"
745
+ },
746
+ {
747
+ "test": "tests/models/kyutai_speech_to_text/test_modeling_kyutai_speech_to_text.py::KyutaiSpeechToTextModelTest::test_flash_attn_2_inference_equivalence_right_padding",
748
+ "commit": null,
749
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
750
+ "pr_number": null,
751
+ "author": null,
752
+ "merged_by": null,
753
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256266060"
754
+ }
755
+ ]
756
+ },
757
+ "lfm2": {
758
+ "single-gpu": [
759
+ {
760
+ "test": "tests/models/lfm2/test_modeling_lfm2.py::Lfm2ModelTest::test_flash_attention_2_padding_matches_padding_free_with_position_ids",
761
+ "commit": null,
762
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
763
+ "pr_number": null,
764
+ "author": null,
765
+ "merged_by": null,
766
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256266197"
767
+ },
768
+ {
769
+ "test": "tests/models/lfm2/test_modeling_lfm2.py::Lfm2ModelTest::test_flash_attn_2_can_compile_with_attention_mask_None_without_graph_break",
770
+ "commit": null,
771
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
772
+ "pr_number": null,
773
+ "author": null,
774
+ "merged_by": null,
775
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256266197"
776
+ }
777
+ ]
778
+ },
779
+ "lfm2_moe": {
780
+ "single-gpu": [
781
+ {
782
+ "test": "tests/models/lfm2_moe/test_modeling_lfm2_moe.py::Lfm2MoeModelTest::test_flash_attention_2_padding_matches_padding_free_with_position_ids",
783
+ "commit": null,
784
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
785
+ "pr_number": null,
786
+ "author": null,
787
+ "merged_by": null,
788
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256266241"
789
+ },
790
+ {
791
+ "test": "tests/models/lfm2_moe/test_modeling_lfm2_moe.py::Lfm2MoeModelTest::test_flash_attn_2_can_compile_with_attention_mask_None_without_graph_break",
792
+ "commit": null,
793
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
794
+ "pr_number": null,
795
+ "author": null,
796
+ "merged_by": null,
797
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256266241"
798
+ }
799
+ ]
800
+ },
801
+ "lfm2_vl": {
802
+ "single-gpu": [
803
+ {
804
+ "test": "tests/models/lfm2_vl/test_modeling_lfm2_vl.py::Lfm2VlModelTest::test_flash_attn_2_inference_equivalence",
805
+ "commit": null,
806
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
807
+ "pr_number": null,
808
+ "author": null,
809
+ "merged_by": null,
810
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256266254"
811
+ },
812
+ {
813
+ "test": "tests/models/lfm2_vl/test_modeling_lfm2_vl.py::Lfm2VlModelTest::test_flash_attn_2_inference_equivalence_right_padding",
814
+ "commit": null,
815
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
816
+ "pr_number": null,
817
+ "author": null,
818
+ "merged_by": null,
819
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256266254"
820
+ }
821
+ ]
822
+ },
823
+ "llava_next": {
824
+ "single-gpu": [
825
+ {
826
+ "test": "tests/models/llava_next/test_modeling_llava_next.py::LlavaNextForConditionalGenerationModelTest::test_flash_attn_2_inference_equivalence",
827
+ "commit": null,
828
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
829
+ "pr_number": null,
830
+ "author": null,
831
+ "merged_by": null,
832
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256266347"
833
+ },
834
+ {
835
+ "test": "tests/models/llava_next/test_modeling_llava_next.py::LlavaNextForConditionalGenerationModelTest::test_flash_attn_2_inference_equivalence_right_padding",
836
+ "commit": null,
837
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
838
+ "pr_number": null,
839
+ "author": null,
840
+ "merged_by": null,
841
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256266347"
842
+ }
843
+ ]
844
+ },
845
+ "llava_next_video": {
846
+ "single-gpu": [
847
+ {
848
+ "test": "tests/models/llava_next_video/test_modeling_llava_next_video.py::LlavaNextVideoForConditionalGenerationModelTest::test_flash_attn_2_inference_equivalence",
849
+ "commit": null,
850
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
851
+ "pr_number": null,
852
+ "author": null,
853
+ "merged_by": null,
854
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256266334"
855
+ },
856
+ {
857
+ "test": "tests/models/llava_next_video/test_modeling_llava_next_video.py::LlavaNextVideoForConditionalGenerationModelTest::test_flash_attn_2_inference_equivalence_right_padding",
858
+ "commit": null,
859
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
860
+ "pr_number": null,
861
+ "author": null,
862
+ "merged_by": null,
863
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256266334"
864
+ }
865
+ ]
866
+ },
867
+ "llava_onevision": {
868
+ "single-gpu": [
869
+ {
870
+ "test": "tests/models/llava_onevision/test_modeling_llava_onevision.py::LlavaOnevisionForConditionalGenerationModelTest::test_flash_attn_2_inference_equivalence",
871
+ "commit": null,
872
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
873
+ "pr_number": null,
874
+ "author": null,
875
+ "merged_by": null,
876
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256266412"
877
+ },
878
+ {
879
+ "test": "tests/models/llava_onevision/test_modeling_llava_onevision.py::LlavaOnevisionForConditionalGenerationModelTest::test_flash_attn_2_inference_equivalence_right_padding",
880
+ "commit": null,
881
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
882
+ "pr_number": null,
883
+ "author": null,
884
+ "merged_by": null,
885
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256266412"
886
+ }
887
+ ]
888
+ },
889
+ "mask2former": {
890
+ "single-gpu": [
891
+ {
892
+ "test": "tests/models/mask2former/test_modeling_mask2former.py::Mask2FormerModelTest::test_flash_attn_2_inference_equivalence",
893
+ "commit": null,
894
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
895
+ "pr_number": null,
896
+ "author": null,
897
+ "merged_by": null,
898
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256266649"
899
+ },
900
+ {
901
+ "test": "tests/models/mask2former/test_modeling_mask2former.py::Mask2FormerModelTest::test_flash_attn_2_inference_equivalence_right_padding",
902
+ "commit": null,
903
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
904
+ "pr_number": null,
905
+ "author": null,
906
+ "merged_by": null,
907
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256266649"
908
+ }
909
+ ]
910
+ },
911
+ "maskformer": {
912
+ "single-gpu": [
913
+ {
914
+ "test": "tests/models/maskformer/test_modeling_maskformer.py::MaskFormerModelTest::test_flash_attn_2_inference_equivalence",
915
+ "commit": null,
916
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
917
+ "pr_number": null,
918
+ "author": null,
919
+ "merged_by": null,
920
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256266666"
921
+ },
922
+ {
923
+ "test": "tests/models/maskformer/test_modeling_maskformer.py::MaskFormerModelTest::test_flash_attn_2_inference_equivalence_right_padding",
924
+ "commit": null,
925
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
926
+ "pr_number": null,
927
+ "author": null,
928
+ "merged_by": null,
929
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256266666"
930
+ },
931
+ {
932
+ "test": "tests/models/maskformer/test_modeling_maskformer_swin.py::MaskFormerSwinModelTest::test_flash_attn_2_inference_equivalence",
933
+ "commit": null,
934
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
935
+ "pr_number": null,
936
+ "author": null,
937
+ "merged_by": null,
938
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256266666"
939
+ },
940
+ {
941
+ "test": "tests/models/maskformer/test_modeling_maskformer_swin.py::MaskFormerSwinModelTest::test_flash_attn_2_inference_equivalence_right_padding",
942
+ "commit": null,
943
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
944
+ "pr_number": null,
945
+ "author": null,
946
+ "merged_by": null,
947
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256266666"
948
+ }
949
+ ]
950
+ },
951
+ "mllama": {
952
+ "single-gpu": [
953
+ {
954
+ "test": "tests/models/mllama/test_modeling_mllama.py::MllamaForConditionalGenerationModelTest::test_eager_matches_fa2_generate",
955
+ "commit": null,
956
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
957
+ "pr_number": null,
958
+ "author": null,
959
+ "merged_by": null,
960
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256262497"
961
+ },
962
+ {
963
+ "test": "tests/models/mllama/test_modeling_mllama.py::MllamaForConditionalGenerationModelTest::test_flash_attn_2_inference_equivalence",
964
+ "commit": null,
965
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
966
+ "pr_number": null,
967
+ "author": null,
968
+ "merged_by": null,
969
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256262497"
970
+ },
971
+ {
972
+ "test": "tests/models/mllama/test_modeling_mllama.py::MllamaForConditionalGenerationModelTest::test_flash_attn_2_inference_equivalence_right_padding",
973
+ "commit": null,
974
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
975
+ "pr_number": null,
976
+ "author": null,
977
+ "merged_by": null,
978
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256262497"
979
+ }
980
+ ]
981
+ },
982
+ "mm_grounding_dino": {
983
+ "single-gpu": [
984
+ {
985
+ "test": "tests/models/mm_grounding_dino/test_modeling_mm_grounding_dino.py::MMGroundingDinoModelTest::test_flash_attn_2_inference_equivalence",
986
+ "commit": null,
987
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
988
+ "pr_number": null,
989
+ "author": null,
990
+ "merged_by": null,
991
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256262500"
992
+ },
993
+ {
994
+ "test": "tests/models/mm_grounding_dino/test_modeling_mm_grounding_dino.py::MMGroundingDinoModelTest::test_flash_attn_2_inference_equivalence_right_padding",
995
+ "commit": null,
996
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
997
+ "pr_number": null,
998
+ "author": null,
999
+ "merged_by": null,
1000
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256262500"
1001
+ }
1002
+ ]
1003
+ },
1004
+ "moshi": {
1005
+ "single-gpu": [
1006
+ {
1007
+ "test": "tests/models/moshi/test_modeling_moshi.py::MoshiDecoderTest::test_flash_attn_2_inference_equivalence",
1008
+ "commit": null,
1009
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1010
+ "pr_number": null,
1011
+ "author": null,
1012
+ "merged_by": null,
1013
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256262704"
1014
+ },
1015
+ {
1016
+ "test": "tests/models/moshi/test_modeling_moshi.py::MoshiDecoderTest::test_flash_attn_2_inference_equivalence_right_padding",
1017
+ "commit": null,
1018
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1019
+ "pr_number": null,
1020
+ "author": null,
1021
+ "merged_by": null,
1022
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256262704"
1023
+ },
1024
+ {
1025
+ "test": "tests/models/moshi/test_modeling_moshi.py::MoshiTest::test_eager_matches_fa2_generate",
1026
+ "commit": null,
1027
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1028
+ "pr_number": null,
1029
+ "author": null,
1030
+ "merged_by": null,
1031
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256262704"
1032
+ },
1033
+ {
1034
+ "test": "tests/models/moshi/test_modeling_moshi.py::MoshiTest::test_flash_attn_2_fp32_ln",
1035
+ "commit": null,
1036
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1037
+ "pr_number": null,
1038
+ "author": null,
1039
+ "merged_by": null,
1040
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256262704"
1041
+ },
1042
+ {
1043
+ "test": "tests/models/moshi/test_modeling_moshi.py::MoshiTest::test_flash_attn_2_from_config",
1044
+ "commit": null,
1045
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1046
+ "pr_number": null,
1047
+ "author": null,
1048
+ "merged_by": null,
1049
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256262704"
1050
+ },
1051
+ {
1052
+ "test": "tests/models/moshi/test_modeling_moshi.py::MoshiTest::test_flash_attn_2_inference_equivalence",
1053
+ "commit": null,
1054
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1055
+ "pr_number": null,
1056
+ "author": null,
1057
+ "merged_by": null,
1058
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256262704"
1059
+ },
1060
+ {
1061
+ "test": "tests/models/moshi/test_modeling_moshi.py::MoshiTest::test_flash_attn_2_inference_equivalence_right_padding",
1062
+ "commit": null,
1063
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1064
+ "pr_number": null,
1065
+ "author": null,
1066
+ "merged_by": null,
1067
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256262704"
1068
+ }
1069
+ ]
1070
+ },
1071
+ "nemotron": {
1072
+ "single-gpu": [
1073
+ {
1074
+ "test": "tests/models/nemotron/test_modeling_nemotron.py::NemotronModelTest::test_flash_attn_2_equivalence",
1075
+ "commit": null,
1076
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1077
+ "pr_number": null,
1078
+ "author": null,
1079
+ "merged_by": null,
1080
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256262902"
1081
+ }
1082
+ ]
1083
+ },
1084
+ "olmo": {
1085
+ "single-gpu": [
1086
+ {
1087
+ "test": "tests/models/olmo/test_modeling_olmo.py::OlmoModelTest::test_flash_attention_2_padding_matches_padding_free_with_position_ids",
1088
+ "commit": null,
1089
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1090
+ "pr_number": null,
1091
+ "author": null,
1092
+ "merged_by": null,
1093
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256263145"
1094
+ }
1095
+ ]
1096
+ },
1097
+ "olmo2": {
1098
+ "single-gpu": [
1099
+ {
1100
+ "test": "tests/models/olmo2/test_modeling_olmo2.py::Olmo2ModelTest::test_flash_attention_2_padding_matches_padding_free_with_position_ids",
1101
+ "commit": null,
1102
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1103
+ "pr_number": null,
1104
+ "author": null,
1105
+ "merged_by": null,
1106
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256263087"
1107
+ }
1108
+ ]
1109
+ },
1110
+ "omdet_turbo": {
1111
+ "single-gpu": [
1112
+ {
1113
+ "test": "tests/models/omdet_turbo/test_modeling_omdet_turbo.py::OmDetTurboModelTest::test_flash_attn_2_inference_equivalence",
1114
+ "commit": null,
1115
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1116
+ "pr_number": null,
1117
+ "author": null,
1118
+ "merged_by": null,
1119
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256263331"
1120
+ },
1121
+ {
1122
+ "test": "tests/models/omdet_turbo/test_modeling_omdet_turbo.py::OmDetTurboModelTest::test_flash_attn_2_inference_equivalence_right_padding",
1123
+ "commit": null,
1124
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1125
+ "pr_number": null,
1126
+ "author": null,
1127
+ "merged_by": null,
1128
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256263331"
1129
+ }
1130
+ ]
1131
+ },
1132
+ "oneformer": {
1133
+ "single-gpu": [
1134
+ {
1135
+ "test": "tests/models/oneformer/test_modeling_oneformer.py::OneFormerModelTest::test_flash_attn_2_inference_equivalence",
1136
+ "commit": null,
1137
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1138
+ "pr_number": null,
1139
+ "author": null,
1140
+ "merged_by": null,
1141
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256263301"
1142
+ },
1143
+ {
1144
+ "test": "tests/models/oneformer/test_modeling_oneformer.py::OneFormerModelTest::test_flash_attn_2_inference_equivalence_right_padding",
1145
+ "commit": null,
1146
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1147
+ "pr_number": null,
1148
+ "author": null,
1149
+ "merged_by": null,
1150
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256263301"
1151
+ }
1152
+ ]
1153
+ },
1154
+ "paligemma": {
1155
+ "single-gpu": [
1156
+ {
1157
+ "test": "tests/models/paligemma/test_modeling_paligemma.py::PaliGemmaForConditionalGenerationModelTest::test_flash_attn_2_from_config",
1158
+ "commit": null,
1159
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1160
+ "pr_number": null,
1161
+ "author": null,
1162
+ "merged_by": null,
1163
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256263449"
1164
+ },
1165
+ {
1166
+ "test": "tests/models/paligemma/test_modeling_paligemma.py::PaliGemmaForConditionalGenerationModelTest::test_flash_attn_2_inference_equivalence",
1167
+ "commit": null,
1168
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1169
+ "pr_number": null,
1170
+ "author": null,
1171
+ "merged_by": null,
1172
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256263449"
1173
+ },
1174
+ {
1175
+ "test": "tests/models/paligemma/test_modeling_paligemma.py::PaliGemmaForConditionalGenerationModelTest::test_flash_attn_2_inference_equivalence_right_padding",
1176
+ "commit": null,
1177
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1178
+ "pr_number": null,
1179
+ "author": null,
1180
+ "merged_by": null,
1181
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256263449"
1182
+ }
1183
+ ]
1184
+ },
1185
+ "paligemma2": {
1186
+ "single-gpu": [
1187
+ {
1188
+ "test": "tests/models/paligemma2/test_modeling_paligemma2.py::PaliGemma2ForConditionalGenerationModelTest::test_flash_attn_2_from_config",
1189
+ "commit": null,
1190
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1191
+ "pr_number": null,
1192
+ "author": null,
1193
+ "merged_by": null,
1194
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256263448"
1195
+ },
1196
+ {
1197
+ "test": "tests/models/paligemma2/test_modeling_paligemma2.py::PaliGemma2ForConditionalGenerationModelTest::test_flash_attn_2_inference_equivalence",
1198
+ "commit": null,
1199
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1200
+ "pr_number": null,
1201
+ "author": null,
1202
+ "merged_by": null,
1203
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256263448"
1204
+ },
1205
+ {
1206
+ "test": "tests/models/paligemma2/test_modeling_paligemma2.py::PaliGemma2ForConditionalGenerationModelTest::test_flash_attn_2_inference_equivalence_right_padding",
1207
+ "commit": null,
1208
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1209
+ "pr_number": null,
1210
+ "author": null,
1211
+ "merged_by": null,
1212
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256263448"
1213
+ }
1214
+ ]
1215
+ },
1216
+ "pegasus_x": {
1217
+ "single-gpu": [
1218
+ {
1219
+ "test": "tests/models/pegasus_x/test_modeling_pegasus_x.py::PegasusXModelTest::test_flash_attn_2_inference_equivalence",
1220
+ "commit": null,
1221
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1222
+ "pr_number": null,
1223
+ "author": null,
1224
+ "merged_by": null,
1225
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256263633"
1226
+ },
1227
+ {
1228
+ "test": "tests/models/pegasus_x/test_modeling_pegasus_x.py::PegasusXModelTest::test_flash_attn_2_inference_equivalence_right_padding",
1229
+ "commit": null,
1230
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1231
+ "pr_number": null,
1232
+ "author": null,
1233
+ "merged_by": null,
1234
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256263633"
1235
+ },
1236
+ {
1237
+ "test": "tests/models/pegasus_x/test_modeling_pegasus_x.py::PegasusXStandaloneDecoderModelTest::test_flash_attn_2_inference_equivalence",
1238
+ "commit": null,
1239
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1240
+ "pr_number": null,
1241
+ "author": null,
1242
+ "merged_by": null,
1243
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256263633"
1244
+ },
1245
+ {
1246
+ "test": "tests/models/pegasus_x/test_modeling_pegasus_x.py::PegasusXStandaloneDecoderModelTest::test_flash_attn_2_inference_equivalence_right_padding",
1247
+ "commit": null,
1248
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1249
+ "pr_number": null,
1250
+ "author": null,
1251
+ "merged_by": null,
1252
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256263633"
1253
+ }
1254
+ ]
1255
+ },
1256
+ "perception_lm": {
1257
+ "single-gpu": [
1258
+ {
1259
+ "test": "tests/models/perception_lm/test_modeling_perception_lm.py::PerceptionLMForConditionalGenerationModelTest::test_flash_attention_2_continue_generate_with_position_ids",
1260
+ "commit": null,
1261
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1262
+ "pr_number": null,
1263
+ "author": null,
1264
+ "merged_by": null,
1265
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256263665"
1266
+ }
1267
+ ]
1268
+ },
1269
+ "phi": {
1270
+ "single-gpu": [
1271
+ {
1272
+ "test": "tests/models/phi/test_modeling_phi.py::PhiModelTest::test_flash_attention_2_padding_matches_padding_free_with_position_ids",
1273
+ "commit": null,
1274
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1275
+ "pr_number": null,
1276
+ "author": null,
1277
+ "merged_by": null,
1278
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256263676"
1279
+ }
1280
+ ]
1281
+ },
1282
+ "phimoe": {
1283
+ "single-gpu": [
1284
+ {
1285
+ "test": "tests/models/phimoe/test_modeling_phimoe.py::PhimoeModelTest::test_flash_attn_2_equivalence",
1286
+ "commit": null,
1287
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1288
+ "pr_number": null,
1289
+ "author": null,
1290
+ "merged_by": null,
1291
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256263814"
1292
+ }
1293
+ ]
1294
+ },
1295
+ "pixtral": {
1296
+ "single-gpu": [
1297
+ {
1298
+ "test": "tests/models/pixtral/test_modeling_pixtral.py::PixtralVisionModelModelTest::test_flash_attn_2_inference_equivalence",
1299
+ "commit": null,
1300
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1301
+ "pr_number": null,
1302
+ "author": null,
1303
+ "merged_by": null,
1304
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256263835"
1305
+ },
1306
+ {
1307
+ "test": "tests/models/pixtral/test_modeling_pixtral.py::PixtralVisionModelModelTest::test_flash_attn_2_inference_equivalence_right_padding",
1308
+ "commit": null,
1309
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1310
+ "pr_number": null,
1311
+ "author": null,
1312
+ "merged_by": null,
1313
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256263835"
1314
+ }
1315
+ ]
1316
+ },
1317
+ "qwen2_5_vl": {
1318
+ "single-gpu": [
1319
+ {
1320
+ "test": "tests/models/qwen2_5_vl/test_modeling_qwen2_5_vl.py::Qwen2_5_VLIntegrationTest::test_small_model_integration_test_batch_wo_image_flashatt2",
1321
+ "commit": null,
1322
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1323
+ "pr_number": null,
1324
+ "author": null,
1325
+ "merged_by": null,
1326
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256264041"
1327
+ }
1328
+ ]
1329
+ },
1330
+ "qwen3_omni_moe": {
1331
+ "single-gpu": [
1332
+ {
1333
+ "test": "tests/models/qwen3_omni_moe/test_modeling_qwen3_omni_moe.py::Qwen2_5OmniThinkerForConditionalGenerationModelTest::test_flash_attention_2_padding_matches_padding_free_with_position_ids",
1334
+ "commit": null,
1335
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1336
+ "pr_number": null,
1337
+ "author": null,
1338
+ "merged_by": null,
1339
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256264215"
1340
+ },
1341
+ {
1342
+ "test": "tests/models/qwen3_omni_moe/test_modeling_qwen3_omni_moe.py::Qwen2_5OmniThinkerForConditionalGenerationModelTest::test_flash_attention_2_padding_matches_padding_free_with_position_ids_and_fa_kwargs",
1343
+ "commit": null,
1344
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1345
+ "pr_number": null,
1346
+ "author": null,
1347
+ "merged_by": null,
1348
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256264215"
1349
+ },
1350
+ {
1351
+ "test": "tests/models/qwen3_omni_moe/test_modeling_qwen3_omni_moe.py::Qwen2_5OmniModelIntegrationTest::test_small_model_integration_test_batch_flashatt2",
1352
+ "commit": null,
1353
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1354
+ "pr_number": null,
1355
+ "author": null,
1356
+ "merged_by": null,
1357
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256264215"
1358
+ }
1359
+ ]
1360
+ },
1361
+ "qwen3_vl_moe": {
1362
+ "single-gpu": [
1363
+ {
1364
+ "test": "tests/models/qwen3_vl_moe/test_modeling_qwen3_vl_moe.py::Qwen3VLMoeIntegrationTest::test_small_model_integration_test_batch_flashatt2",
1365
+ "commit": null,
1366
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1367
+ "pr_number": null,
1368
+ "author": null,
1369
+ "merged_by": null,
1370
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256264218"
1371
+ },
1372
+ {
1373
+ "test": "tests/models/qwen3_vl_moe/test_modeling_qwen3_vl_moe.py::Qwen3VLMoeIntegrationTest::test_small_model_integration_test_batch_wo_image_flashatt2",
1374
+ "commit": null,
1375
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1376
+ "pr_number": null,
1377
+ "author": null,
1378
+ "merged_by": null,
1379
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256264218"
1380
+ }
1381
+ ]
1382
+ },
1383
+ "sam2": {
1384
+ "single-gpu": [
1385
+ {
1386
+ "test": "tests/models/sam2/test_modeling_sam2.py::Sam2ModelTest::test_flash_attn_2_can_dispatch_composite_models",
1387
+ "commit": null,
1388
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1389
+ "pr_number": null,
1390
+ "author": null,
1391
+ "merged_by": null,
1392
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256264556"
1393
+ }
1394
+ ]
1395
+ },
1396
+ "smollm3": {
1397
+ "single-gpu": [
1398
+ {
1399
+ "test": "tests/models/smollm3/test_modeling_smollm3.py::SmolLM3IntegrationTest::test_model_3b_long_prompt",
1400
+ "commit": null,
1401
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1402
+ "pr_number": null,
1403
+ "author": null,
1404
+ "merged_by": null,
1405
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256264885"
1406
+ }
1407
+ ]
1408
+ },
1409
+ "squeezebert": {
1410
+ "single-gpu": [
1411
+ {
1412
+ "test": "tests/models/squeezebert/test_modeling_squeezebert.py::SqueezeBertModelTest::test_flash_attn_2_inference_equivalence",
1413
+ "commit": null,
1414
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1415
+ "pr_number": null,
1416
+ "author": null,
1417
+ "merged_by": null,
1418
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256264948"
1419
+ },
1420
+ {
1421
+ "test": "tests/models/squeezebert/test_modeling_squeezebert.py::SqueezeBertModelTest::test_flash_attn_2_inference_equivalence_right_padding",
1422
+ "commit": null,
1423
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1424
+ "pr_number": null,
1425
+ "author": null,
1426
+ "merged_by": null,
1427
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256264948"
1428
+ }
1429
+ ]
1430
+ },
1431
+ "starcoder2": {
1432
+ "single-gpu": [
1433
+ {
1434
+ "test": "tests/models/starcoder2/test_modeling_starcoder2.py::Starcoder2IntegrationTest::test_starcoder2_batched_generation_fa2",
1435
+ "commit": null,
1436
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1437
+ "pr_number": null,
1438
+ "author": null,
1439
+ "merged_by": null,
1440
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256265091"
1441
+ }
1442
+ ]
1443
+ },
1444
+ "swin": {
1445
+ "single-gpu": [
1446
+ {
1447
+ "test": "tests/models/swin/test_modeling_swin.py::SwinModelTest::test_flash_attn_2_inference_equivalence",
1448
+ "commit": null,
1449
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1450
+ "pr_number": null,
1451
+ "author": null,
1452
+ "merged_by": null,
1453
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256265137"
1454
+ },
1455
+ {
1456
+ "test": "tests/models/swin/test_modeling_swin.py::SwinModelTest::test_flash_attn_2_inference_equivalence_right_padding",
1457
+ "commit": null,
1458
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1459
+ "pr_number": null,
1460
+ "author": null,
1461
+ "merged_by": null,
1462
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256265137"
1463
+ }
1464
+ ]
1465
+ },
1466
+ "swin2sr": {
1467
+ "single-gpu": [
1468
+ {
1469
+ "test": "tests/models/swin2sr/test_modeling_swin2sr.py::Swin2SRModelTest::test_flash_attn_2_inference_equivalence",
1470
+ "commit": null,
1471
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1472
+ "pr_number": null,
1473
+ "author": null,
1474
+ "merged_by": null,
1475
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256265136"
1476
+ },
1477
+ {
1478
+ "test": "tests/models/swin2sr/test_modeling_swin2sr.py::Swin2SRModelTest::test_flash_attn_2_inference_equivalence_right_padding",
1479
+ "commit": null,
1480
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1481
+ "pr_number": null,
1482
+ "author": null,
1483
+ "merged_by": null,
1484
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256265136"
1485
+ }
1486
+ ]
1487
+ },
1488
+ "swinv2": {
1489
+ "single-gpu": [
1490
+ {
1491
+ "test": "tests/models/swinv2/test_modeling_swinv2.py::Swinv2ModelTest::test_flash_attn_2_inference_equivalence",
1492
+ "commit": null,
1493
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1494
+ "pr_number": null,
1495
+ "author": null,
1496
+ "merged_by": null,
1497
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256265145"
1498
+ },
1499
+ {
1500
+ "test": "tests/models/swinv2/test_modeling_swinv2.py::Swinv2ModelTest::test_flash_attn_2_inference_equivalence_right_padding",
1501
+ "commit": null,
1502
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1503
+ "pr_number": null,
1504
+ "author": null,
1505
+ "merged_by": null,
1506
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256265145"
1507
+ }
1508
+ ]
1509
+ },
1510
+ "t5gemma": {
1511
+ "single-gpu": [
1512
+ {
1513
+ "test": "tests/models/t5gemma/test_modeling_t5gemma.py::T5GemmaModelTest::test_flash_attn_2_can_compile_with_attention_mask_None_without_graph_break",
1514
+ "commit": null,
1515
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1516
+ "pr_number": null,
1517
+ "author": null,
1518
+ "merged_by": null,
1519
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256265203"
1520
+ }
1521
+ ]
1522
+ },
1523
+ "zamba": {
1524
+ "single-gpu": [
1525
+ {
1526
+ "test": "tests/models/zamba/test_modeling_zamba.py::ZambaModelTest::test_flash_attn_2_fp32_ln",
1527
+ "commit": null,
1528
+ "status": "flaky: test passed in the previous run (commit: 2b8068c306a4c79350e1af5fea5f7e3d93d82d95) but failed (on the same commit) during the check of the current run.",
1529
+ "pr_number": null,
1530
+ "author": null,
1531
+ "merged_by": null,
1532
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19318904934/job/55256266375"
1533
+ }
1534
+ ]
1535
+ }
1536
+ }