Upload 2025-12-08/ci_results_run_models_gpu/new_failures.txt with huggingface_hub
Browse files
2025-12-08/ci_results_run_models_gpu/new_failures.txt
ADDED
|
@@ -0,0 +1,24 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
<https://github.com/huggingface/transformers/actions/runs/20015154789/job/57391463932|multi> gpu
|
| 2 |
+
tests/models/data2vec/test_modeling_data2vec_text.py::Data2VecTextModelTest::test_flash_attn_2_inference_equivalence_right_padding
|
| 3 |
+
|
| 4 |
+
<https://github.com/huggingface/transformers/actions/runs/20015154789/job/57391465612|single>|<https://github.com/huggingface/transformers/actions/runs/20015154789/job/57391465223|multi> gpu
|
| 5 |
+
tests/models/electra/test_modeling_electra.py::ElectraModelTest::test_flash_attn_2_inference_equivalence
|
| 6 |
+
|
| 7 |
+
<https://github.com/huggingface/transformers/actions/runs/20015154789/job/57391465656|multi> gpu
|
| 8 |
+
tests/models/flex_olmo/test_modeling_flex_olmo.py::FlexOlmoModelTest::test_flash_attention_2_padding_matches_padding_free_with_position_ids_and_fa_kwargs
|
| 9 |
+
|
| 10 |
+
<https://github.com/huggingface/transformers/actions/runs/20015154789/job/57391466318|single> gpu
|
| 11 |
+
tests/models/glm4/test_modeling_glm4.py::Glm4ModelTest::test_flash_attn_2_equivalence
|
| 12 |
+
|
| 13 |
+
<https://github.com/huggingface/transformers/actions/runs/20015154789/job/57391467981|single> gpu
|
| 14 |
+
tests/models/lfm2_vl/test_modeling_lfm2_vl.py::Lfm2VlModelTest::test_flash_attention_2_padding_matches_padding_free_with_position_ids_and_fa_kwargs
|
| 15 |
+
|
| 16 |
+
<https://github.com/huggingface/transformers/actions/runs/20015154789/job/57391461988|multi> gpu
|
| 17 |
+
tests/models/mixtral/test_modeling_mixtral.py::MixtralModelTest::test_flash_attn_2_equivalence
|
| 18 |
+
|
| 19 |
+
<https://github.com/huggingface/transformers/actions/runs/20015154789/job/57391464940|single> gpu
|
| 20 |
+
tests/models/roberta/test_modeling_roberta.py::RobertaModelTest::test_flash_attn_2_inference_equivalence
|
| 21 |
+
|
| 22 |
+
<https://github.com/huggingface/transformers/actions/runs/20015154789/job/57391467238|single> gpu
|
| 23 |
+
tests/models/xmod/test_modeling_xmod.py::XmodModelTest::test_flash_attn_2_inference_equivalence
|
| 24 |
+
|