hf-transformers-bot commited on
Commit
280be1a
·
verified ·
1 Parent(s): f211ec1

Upload 2025-12-08/ci_results_run_models_gpu/new_failures_with_bad_commit.json with huggingface_hub

Browse files
2025-12-08/ci_results_run_models_gpu/new_failures_with_bad_commit.json ADDED
@@ -0,0 +1,67 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "electra": {
3
+ "single-gpu": [
4
+ {
5
+ "test": "tests/models/electra/test_modeling_electra.py::ElectraModelTest::test_flash_attn_2_inference_equivalence",
6
+ "commit": null,
7
+ "status": "flaky: test fails on the current CI run but passed in the previous run which is running on the same commit ff13eb668aa03f151ded71636d723f2e490ad967.",
8
+ "pr_number": null,
9
+ "author": null,
10
+ "merged_by": null,
11
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/20015154789/job/57391465612"
12
+ }
13
+ ]
14
+ },
15
+ "glm4": {
16
+ "single-gpu": [
17
+ {
18
+ "test": "tests/models/glm4/test_modeling_glm4.py::Glm4ModelTest::test_flash_attn_2_equivalence",
19
+ "commit": null,
20
+ "status": "flaky: test passed in the previous run (commit: ff13eb668aa03f151ded71636d723f2e490ad967) but failed (on the same commit) during the check of the current run.",
21
+ "pr_number": null,
22
+ "author": null,
23
+ "merged_by": null,
24
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/20015154789/job/57391466318"
25
+ }
26
+ ]
27
+ },
28
+ "lfm2_vl": {
29
+ "single-gpu": [
30
+ {
31
+ "test": "tests/models/lfm2_vl/test_modeling_lfm2_vl.py::Lfm2VlModelTest::test_flash_attention_2_padding_matches_padding_free_with_position_ids_and_fa_kwargs",
32
+ "commit": null,
33
+ "status": "flaky: test fails on the current CI run but passed in the previous run which is running on the same commit ff13eb668aa03f151ded71636d723f2e490ad967.",
34
+ "pr_number": null,
35
+ "author": null,
36
+ "merged_by": null,
37
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/20015154789/job/57391467981"
38
+ }
39
+ ]
40
+ },
41
+ "roberta": {
42
+ "single-gpu": [
43
+ {
44
+ "test": "tests/models/roberta/test_modeling_roberta.py::RobertaModelTest::test_flash_attn_2_inference_equivalence",
45
+ "commit": null,
46
+ "status": "flaky: test passed in the previous run (commit: ff13eb668aa03f151ded71636d723f2e490ad967) but failed (on the same commit) during the check of the current run.",
47
+ "pr_number": null,
48
+ "author": null,
49
+ "merged_by": null,
50
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/20015154789/job/57391464940"
51
+ }
52
+ ]
53
+ },
54
+ "xmod": {
55
+ "single-gpu": [
56
+ {
57
+ "test": "tests/models/xmod/test_modeling_xmod.py::XmodModelTest::test_flash_attn_2_inference_equivalence",
58
+ "commit": null,
59
+ "status": "flaky: test fails on the current CI run but passed in the previous run which is running on the same commit ff13eb668aa03f151ded71636d723f2e490ad967.",
60
+ "pr_number": null,
61
+ "author": null,
62
+ "merged_by": null,
63
+ "job_link": "https://github.com/huggingface/transformers/actions/runs/20015154789/job/57391467238"
64
+ }
65
+ ]
66
+ }
67
+ }