Upload 2025-11-25/ci_results_run_models_gpu/new_failures.txt with huggingface_hub
Browse files
2025-11-25/ci_results_run_models_gpu/new_failures.txt
ADDED
|
@@ -0,0 +1,27 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
<https://github.com/huggingface/transformers/actions/runs/19656569354/job/56294609855|multi> gpu
|
| 2 |
+
tests/models/ernie/test_modeling_ernie.py::ErnieModelTest::test_flash_attn_2_inference_equivalence_right_padding
|
| 3 |
+
|
| 4 |
+
<https://github.com/huggingface/transformers/actions/runs/19656569354/job/56294610254|single> gpu
|
| 5 |
+
tests/models/gemma2/test_modeling_gemma2.py::Gemma2ModelTest::test_flash_attn_2_equivalence
|
| 6 |
+
|
| 7 |
+
<https://github.com/huggingface/transformers/actions/runs/19656569354/job/56294610275|single>|<https://github.com/huggingface/transformers/actions/runs/19656569354/job/56294610294|multi> gpu
|
| 8 |
+
tests/models/gemma3/test_modeling_gemma3.py::Gemma3IntegrationTest::test_model_4b_flash_attn
|
| 9 |
+
|
| 10 |
+
<https://github.com/huggingface/transformers/actions/runs/19656569354/job/56294610710|single> gpu
|
| 11 |
+
tests/models/gpt_oss/test_modeling_gpt_oss.py::GptOssModelTest::test_flash_attention_2_padding_matches_padding_free_with_position_ids
|
| 12 |
+
|
| 13 |
+
<https://github.com/huggingface/transformers/actions/runs/19656569354/job/56294607080|multi> gpu
|
| 14 |
+
tests/models/mixtral/test_modeling_mixtral.py::MixtralModelTest::test_flash_attn_2_equivalence
|
| 15 |
+
|
| 16 |
+
<https://github.com/huggingface/transformers/actions/runs/19656569354/job/56294609296|multi> gpu
|
| 17 |
+
tests/models/roberta/test_modeling_roberta.py::RobertaModelTest::test_flash_attn_2_inference_equivalence
|
| 18 |
+
|
| 19 |
+
<https://github.com/huggingface/transformers/actions/runs/19656569354/job/56294609455|single>|<https://github.com/huggingface/transformers/actions/runs/19656569354/job/56294609488|multi> gpu
|
| 20 |
+
tests/models/roberta_prelayernorm/test_modeling_roberta_prelayernorm.py::RobertaPreLayerNormModelTest::test_flash_attn_2_inference_equivalence
|
| 21 |
+
|
| 22 |
+
<https://github.com/huggingface/transformers/actions/runs/19656569354/job/56294611874|multi> gpu
|
| 23 |
+
tests/models/xmod/test_modeling_xmod.py::XmodModelTest::test_flash_attn_2_inference_equivalence
|
| 24 |
+
|
| 25 |
+
<https://github.com/huggingface/transformers/actions/runs/19656569354/job/56294611874|multi> gpu
|
| 26 |
+
tests/models/xmod/test_modeling_xmod.py::XmodModelTest::test_flash_attn_2_inference_equivalence_right_padding
|
| 27 |
+
|