hf-transformers-bot commited on
Commit
04d2de2
·
verified ·
1 Parent(s): 5fa01d6

Upload 2025-11-18/ci_results_run_models_gpu/new_failures_with_bad_commit.json with huggingface_hub

Browse files
2025-11-18/ci_results_run_models_gpu/new_failures_with_bad_commit.json ADDED
@@ -0,0 +1,243 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bart": {
3
+ "single-gpu": [
4
+ {
5
+ "test": "tests/models/bart/test_modeling_bart.py::BartModelTest::test_flash_attn_2_fp32_ln",
6
+ "commit": null,
7
+ "status": "flaky: test passed in the previous run (commit: 66d57110f089789ae285cc9d54d3bf051123246b) but failed (on the same commit) during the check of the current run.",
8
+ "pr_number": null,
9
+ "author": null,
10
+ "merged_by": null,
11
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19452252123/job/55659493451"
12
+ },
13
+ {
14
+ "test": "tests/models/bart/test_modeling_bart.py::BartStandaloneDecoderModelTest::test_flash_attn_2_fp32_ln",
15
+ "commit": null,
16
+ "status": "flaky: test passed in the previous run (commit: 66d57110f089789ae285cc9d54d3bf051123246b) but failed (on the same commit) during the check of the current run.",
17
+ "pr_number": null,
18
+ "author": null,
19
+ "merged_by": null,
20
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19452252123/job/55659493451"
21
+ }
22
+ ]
23
+ },
24
+ "bert": {
25
+ "single-gpu": [
26
+ {
27
+ "test": "tests/models/bert/test_modeling_bert.py::BertModelTest::test_flash_attn_2_fp32_ln",
28
+ "commit": null,
29
+ "status": "flaky: test passed in the previous run (commit: 66d57110f089789ae285cc9d54d3bf051123246b) but failed (on the same commit) during the check of the current run.",
30
+ "pr_number": null,
31
+ "author": null,
32
+ "merged_by": null,
33
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19452252123/job/55659493643"
34
+ },
35
+ {
36
+ "test": "tests/models/bert/test_modeling_bert.py::BertModelTest::test_flash_attn_2_inference_equivalence",
37
+ "commit": null,
38
+ "status": "flaky: test passed in the previous run (commit: 66d57110f089789ae285cc9d54d3bf051123246b) but failed (on the same commit) during the check of the current run.",
39
+ "pr_number": null,
40
+ "author": null,
41
+ "merged_by": null,
42
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19452252123/job/55659493643"
43
+ }
44
+ ]
45
+ },
46
+ "bert_generation": {
47
+ "single-gpu": [
48
+ {
49
+ "test": "tests/models/bert_generation/test_modeling_bert_generation.py::BertGenerationEncoderTest::test_flash_attn_2_fp32_ln",
50
+ "commit": null,
51
+ "status": "flaky: test passed in the previous run (commit: 66d57110f089789ae285cc9d54d3bf051123246b) but failed (on the same commit) during the check of the current run.",
52
+ "pr_number": null,
53
+ "author": null,
54
+ "merged_by": null,
55
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19452252123/job/55659493642"
56
+ }
57
+ ]
58
+ },
59
+ "data2vec": {
60
+ "single-gpu": [
61
+ {
62
+ "test": "tests/models/data2vec/test_modeling_data2vec_text.py::Data2VecTextModelTest::test_flash_attn_2_inference_equivalence_right_padding",
63
+ "commit": null,
64
+ "status": "flaky: test passed in the previous run (commit: 66d57110f089789ae285cc9d54d3bf051123246b) but failed (on the same commit) during the check of the current run.",
65
+ "pr_number": null,
66
+ "author": null,
67
+ "merged_by": null,
68
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19452252123/job/55659495569"
69
+ }
70
+ ]
71
+ },
72
+ "electra": {
73
+ "single-gpu": [
74
+ {
75
+ "test": "tests/models/electra/test_modeling_electra.py::ElectraModelTest::test_flash_attn_2_inference_equivalence",
76
+ "commit": null,
77
+ "status": "flaky: test passed in the previous run (commit: 66d57110f089789ae285cc9d54d3bf051123246b) but failed (on the same commit) during the check of the current run.",
78
+ "pr_number": null,
79
+ "author": null,
80
+ "merged_by": null,
81
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19452252123/job/55659496814"
82
+ }
83
+ ]
84
+ },
85
+ "ernie": {
86
+ "single-gpu": [
87
+ {
88
+ "test": "tests/models/ernie/test_modeling_ernie.py::ErnieModelTest::test_flash_attn_2_inference_equivalence_right_padding",
89
+ "commit": null,
90
+ "status": "flaky: test passed in the previous run (commit: 66d57110f089789ae285cc9d54d3bf051123246b) but failed (on the same commit) during the check of the current run.",
91
+ "pr_number": null,
92
+ "author": null,
93
+ "merged_by": null,
94
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19452252123/job/55659497000"
95
+ }
96
+ ]
97
+ },
98
+ "ernie4_5_moe": {
99
+ "single-gpu": [
100
+ {
101
+ "test": "tests/models/ernie4_5_moe/test_modeling_ernie4_5_moe.py::Ernie4_5_MoeModelTest::test_flash_attn_2_equivalence",
102
+ "commit": null,
103
+ "status": "flaky: test fails on the current CI run (commit: 266d3b05684ed5968c613f00b208909f03c72c0e) but passes during the check.",
104
+ "pr_number": null,
105
+ "author": null,
106
+ "merged_by": null,
107
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19452252123/job/55659497008"
108
+ }
109
+ ]
110
+ },
111
+ "flex_olmo": {
112
+ "single-gpu": [
113
+ {
114
+ "test": "tests/models/flex_olmo/test_modeling_flex_olmo.py::FlexOlmoModelTest::test_flash_attn_2_equivalence",
115
+ "commit": null,
116
+ "status": "flaky: test passed in the previous run (commit: 66d57110f089789ae285cc9d54d3bf051123246b) but failed (on the same commit) during the check of the current run.",
117
+ "pr_number": null,
118
+ "author": null,
119
+ "merged_by": null,
120
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19452252123/job/55659497336"
121
+ }
122
+ ]
123
+ },
124
+ "mixtral": {
125
+ "single-gpu": [
126
+ {
127
+ "test": "tests/models/mixtral/test_modeling_mixtral.py::MixtralModelTest::test_flash_attn_2_equivalence",
128
+ "commit": null,
129
+ "status": "flaky: test passed in the previous run (commit: 66d57110f089789ae285cc9d54d3bf051123246b) but failed (on the same commit) during the check of the current run.",
130
+ "pr_number": null,
131
+ "author": null,
132
+ "merged_by": null,
133
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19452252123/job/55659493385"
134
+ },
135
+ {
136
+ "test": "tests/models/mixtral/test_modeling_mixtral.py::MixtralModelTest::test_flash_attn_2_fp32_ln",
137
+ "commit": null,
138
+ "status": "flaky: test passed in the previous run (commit: 66d57110f089789ae285cc9d54d3bf051123246b) but failed (on the same commit) during the check of the current run.",
139
+ "pr_number": null,
140
+ "author": null,
141
+ "merged_by": null,
142
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19452252123/job/55659493385"
143
+ }
144
+ ]
145
+ },
146
+ "mllama": {
147
+ "single-gpu": [
148
+ {
149
+ "test": "tests/models/mllama/test_modeling_mllama.py::MllamaForCausalLMModelTest::test_flash_attn_2_fp32_ln",
150
+ "commit": null,
151
+ "status": "flaky: test passed in the previous run (commit: 66d57110f089789ae285cc9d54d3bf051123246b) but failed (on the same commit) during the check of the current run.",
152
+ "pr_number": null,
153
+ "author": null,
154
+ "merged_by": null,
155
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19452252123/job/55659493362"
156
+ },
157
+ {
158
+ "test": "tests/models/mllama/test_modeling_mllama.py::MllamaForConditionalGenerationModelTest::test_eager_matches_fa2_generate",
159
+ "commit": null,
160
+ "status": "flaky: test passed in the previous run (commit: 66d57110f089789ae285cc9d54d3bf051123246b) but failed (on the same commit) during the check of the current run.",
161
+ "pr_number": null,
162
+ "author": null,
163
+ "merged_by": null,
164
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19452252123/job/55659493362"
165
+ },
166
+ {
167
+ "test": "tests/models/mllama/test_modeling_mllama.py::MllamaForConditionalGenerationModelTest::test_flash_attn_2_fp32_ln",
168
+ "commit": null,
169
+ "status": "flaky: test passed in the previous run (commit: 66d57110f089789ae285cc9d54d3bf051123246b) but failed (on the same commit) during the check of the current run.",
170
+ "pr_number": null,
171
+ "author": null,
172
+ "merged_by": null,
173
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19452252123/job/55659493362"
174
+ },
175
+ {
176
+ "test": "tests/models/mllama/test_modeling_mllama.py::MllamaForConditionalGenerationModelTest::test_flash_attn_2_inference_equivalence",
177
+ "commit": null,
178
+ "status": "flaky: test passed in the previous run (commit: 66d57110f089789ae285cc9d54d3bf051123246b) but failed (on the same commit) during the check of the current run.",
179
+ "pr_number": null,
180
+ "author": null,
181
+ "merged_by": null,
182
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19452252123/job/55659493362"
183
+ },
184
+ {
185
+ "test": "tests/models/mllama/test_modeling_mllama.py::MllamaForConditionalGenerationModelTest::test_flash_attn_2_inference_equivalence_right_padding",
186
+ "commit": null,
187
+ "status": "flaky: test passed in the previous run (commit: 66d57110f089789ae285cc9d54d3bf051123246b) but failed (on the same commit) during the check of the current run.",
188
+ "pr_number": null,
189
+ "author": null,
190
+ "merged_by": null,
191
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19452252123/job/55659493362"
192
+ }
193
+ ]
194
+ },
195
+ "mm_grounding_dino": {
196
+ "single-gpu": [
197
+ {
198
+ "test": "tests/models/mm_grounding_dino/test_modeling_mm_grounding_dino.py::MMGroundingDinoModelTest::test_flash_attn_2_inference_equivalence",
199
+ "commit": null,
200
+ "status": "flaky: test passed in the previous run (commit: 66d57110f089789ae285cc9d54d3bf051123246b) but failed (on the same commit) during the check of the current run.",
201
+ "pr_number": null,
202
+ "author": null,
203
+ "merged_by": null,
204
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19452252123/job/55659493373"
205
+ },
206
+ {
207
+ "test": "tests/models/mm_grounding_dino/test_modeling_mm_grounding_dino.py::MMGroundingDinoModelTest::test_flash_attn_2_inference_equivalence_right_padding",
208
+ "commit": null,
209
+ "status": "flaky: test passed in the previous run (commit: 66d57110f089789ae285cc9d54d3bf051123246b) but failed (on the same commit) during the check of the current run.",
210
+ "pr_number": null,
211
+ "author": null,
212
+ "merged_by": null,
213
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19452252123/job/55659493373"
214
+ }
215
+ ]
216
+ },
217
+ "roberta_prelayernorm": {
218
+ "single-gpu": [
219
+ {
220
+ "test": "tests/models/roberta_prelayernorm/test_modeling_roberta_prelayernorm.py::RobertaPreLayerNormModelTest::test_flash_attn_2_inference_equivalence",
221
+ "commit": null,
222
+ "status": "flaky: test fails on the current CI run (commit: 266d3b05684ed5968c613f00b208909f03c72c0e) but passes during the check.",
223
+ "pr_number": null,
224
+ "author": null,
225
+ "merged_by": null,
226
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19452252123/job/55659496802"
227
+ }
228
+ ]
229
+ },
230
+ "xmod": {
231
+ "single-gpu": [
232
+ {
233
+ "test": "tests/models/xmod/test_modeling_xmod.py::XmodModelTest::test_flash_attn_2_inference_equivalence_right_padding",
234
+ "commit": null,
235
+ "status": "flaky: test fails on the current CI run (commit: 266d3b05684ed5968c613f00b208909f03c72c0e) but passes during the check.",
236
+ "pr_number": null,
237
+ "author": null,
238
+ "merged_by": null,
239
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/19452252123/job/55659499172"
240
+ }
241
+ ]
242
+ }
243
+ }