Upload 2025-11-27/ci_results_run_models_gpu/collated_reports_multi-gpu_1ae4d91.json with huggingface_hub
Browse files
2025-11-27/ci_results_run_models_gpu/collated_reports_multi-gpu_1ae4d91.json
CHANGED
|
@@ -3,9 +3,9 @@
|
|
| 3 |
"machine_type": "multi-gpu",
|
| 4 |
"commit_hash": "1ae4d91",
|
| 5 |
"total_status_count": {
|
| 6 |
-
"passed":
|
| 7 |
"failed": 124,
|
| 8 |
-
"skipped":
|
| 9 |
"error": 0,
|
| 10 |
"null": 0
|
| 11 |
},
|
|
@@ -17611,160 +17611,6 @@
|
|
| 17611 |
}
|
| 17612 |
]
|
| 17613 |
},
|
| 17614 |
-
{
|
| 17615 |
-
"model": "megatron_gpt2",
|
| 17616 |
-
"results": []
|
| 17617 |
-
},
|
| 17618 |
-
{
|
| 17619 |
-
"model": "metaclip_2",
|
| 17620 |
-
"results": [
|
| 17621 |
-
{
|
| 17622 |
-
"status": "passed",
|
| 17623 |
-
"test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2VisionModelTest::test_flash_attn_2_fp32_ln",
|
| 17624 |
-
"count": 1
|
| 17625 |
-
},
|
| 17626 |
-
{
|
| 17627 |
-
"status": "passed",
|
| 17628 |
-
"test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2VisionModelTest::test_flash_attn_2_from_config",
|
| 17629 |
-
"count": 1
|
| 17630 |
-
},
|
| 17631 |
-
{
|
| 17632 |
-
"status": "passed",
|
| 17633 |
-
"test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2VisionModelTest::test_flash_attn_2_inference_equivalence",
|
| 17634 |
-
"count": 1
|
| 17635 |
-
},
|
| 17636 |
-
{
|
| 17637 |
-
"status": "passed",
|
| 17638 |
-
"test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2VisionModelTest::test_flash_attn_2_inference_equivalence_right_padding",
|
| 17639 |
-
"count": 1
|
| 17640 |
-
},
|
| 17641 |
-
{
|
| 17642 |
-
"status": "passed",
|
| 17643 |
-
"test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2TextModelTest::test_flash_attn_2_fp32_ln",
|
| 17644 |
-
"count": 1
|
| 17645 |
-
},
|
| 17646 |
-
{
|
| 17647 |
-
"status": "passed",
|
| 17648 |
-
"test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2TextModelTest::test_flash_attn_2_from_config",
|
| 17649 |
-
"count": 1
|
| 17650 |
-
},
|
| 17651 |
-
{
|
| 17652 |
-
"status": "passed",
|
| 17653 |
-
"test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2TextModelTest::test_flash_attn_2_inference_equivalence",
|
| 17654 |
-
"count": 1
|
| 17655 |
-
},
|
| 17656 |
-
{
|
| 17657 |
-
"status": "passed",
|
| 17658 |
-
"test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2TextModelTest::test_flash_attn_2_inference_equivalence_right_padding",
|
| 17659 |
-
"count": 1
|
| 17660 |
-
},
|
| 17661 |
-
{
|
| 17662 |
-
"status": "passed",
|
| 17663 |
-
"test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2ModelTest::test_flash_attn_2_can_dispatch_composite_models",
|
| 17664 |
-
"count": 1
|
| 17665 |
-
},
|
| 17666 |
-
{
|
| 17667 |
-
"status": "passed",
|
| 17668 |
-
"test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2ModelTest::test_flash_attn_2_fp32_ln",
|
| 17669 |
-
"count": 1
|
| 17670 |
-
},
|
| 17671 |
-
{
|
| 17672 |
-
"status": "passed",
|
| 17673 |
-
"test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2ModelTest::test_flash_attn_2_from_config",
|
| 17674 |
-
"count": 1
|
| 17675 |
-
},
|
| 17676 |
-
{
|
| 17677 |
-
"status": "passed",
|
| 17678 |
-
"test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2ModelTest::test_flash_attn_2_inference_equivalence",
|
| 17679 |
-
"count": 1
|
| 17680 |
-
},
|
| 17681 |
-
{
|
| 17682 |
-
"status": "passed",
|
| 17683 |
-
"test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2ModelTest::test_flash_attn_2_inference_equivalence_right_padding",
|
| 17684 |
-
"count": 1
|
| 17685 |
-
},
|
| 17686 |
-
{
|
| 17687 |
-
"status": "passed",
|
| 17688 |
-
"test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2ForImageClassificationModelTest::test_flash_attn_2_can_dispatch_composite_models",
|
| 17689 |
-
"count": 1
|
| 17690 |
-
},
|
| 17691 |
-
{
|
| 17692 |
-
"status": "passed",
|
| 17693 |
-
"test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2ForImageClassificationModelTest::test_flash_attn_2_fp32_ln",
|
| 17694 |
-
"count": 1
|
| 17695 |
-
},
|
| 17696 |
-
{
|
| 17697 |
-
"status": "passed",
|
| 17698 |
-
"test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2ForImageClassificationModelTest::test_flash_attn_2_from_config",
|
| 17699 |
-
"count": 1
|
| 17700 |
-
},
|
| 17701 |
-
{
|
| 17702 |
-
"status": "passed",
|
| 17703 |
-
"test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2ForImageClassificationModelTest::test_flash_attn_2_inference_equivalence",
|
| 17704 |
-
"count": 1
|
| 17705 |
-
},
|
| 17706 |
-
{
|
| 17707 |
-
"status": "passed",
|
| 17708 |
-
"test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2ForImageClassificationModelTest::test_flash_attn_2_inference_equivalence_right_padding",
|
| 17709 |
-
"count": 1
|
| 17710 |
-
},
|
| 17711 |
-
{
|
| 17712 |
-
"status": "skipped",
|
| 17713 |
-
"test": "[1] tests/test_modeling_common.py:3349: MetaClip2VisionModelTest doesn't have the attribute `_torch_compile_train_cls`.",
|
| 17714 |
-
"count": 1
|
| 17715 |
-
},
|
| 17716 |
-
{
|
| 17717 |
-
"status": "skipped",
|
| 17718 |
-
"test": "[2] tests/test_modeling_common.py:3281: This model is not a composite model!",
|
| 17719 |
-
"count": 2
|
| 17720 |
-
},
|
| 17721 |
-
{
|
| 17722 |
-
"status": "skipped",
|
| 17723 |
-
"test": "[4] tests/test_modeling_common.py:3287: test requires Flash Attention 3",
|
| 17724 |
-
"count": 4
|
| 17725 |
-
},
|
| 17726 |
-
{
|
| 17727 |
-
"status": "skipped",
|
| 17728 |
-
"test": "[4] tests/test_modeling_common.py:3443: test requires Flash Attention 3",
|
| 17729 |
-
"count": 4
|
| 17730 |
-
},
|
| 17731 |
-
{
|
| 17732 |
-
"status": "skipped",
|
| 17733 |
-
"test": "[4] tests/test_modeling_common.py:2897: test requires Flash Attention 3",
|
| 17734 |
-
"count": 4
|
| 17735 |
-
},
|
| 17736 |
-
{
|
| 17737 |
-
"status": "skipped",
|
| 17738 |
-
"test": "[4] tests/test_modeling_common.py:2905: test requires Flash Attention 3",
|
| 17739 |
-
"count": 4
|
| 17740 |
-
},
|
| 17741 |
-
{
|
| 17742 |
-
"status": "skipped",
|
| 17743 |
-
"test": "[4] tests/test_modeling_common.py:2863: test requires the kernels library",
|
| 17744 |
-
"count": 4
|
| 17745 |
-
},
|
| 17746 |
-
{
|
| 17747 |
-
"status": "skipped",
|
| 17748 |
-
"test": "[4] tests/test_modeling_common.py:2871: test requires MPS",
|
| 17749 |
-
"count": 4
|
| 17750 |
-
},
|
| 17751 |
-
{
|
| 17752 |
-
"status": "skipped",
|
| 17753 |
-
"test": "[1] tests/test_modeling_common.py:3349: MetaClip2TextModelTest doesn't have the attribute `_torch_compile_train_cls`.",
|
| 17754 |
-
"count": 1
|
| 17755 |
-
},
|
| 17756 |
-
{
|
| 17757 |
-
"status": "skipped",
|
| 17758 |
-
"test": "[1] tests/test_modeling_common.py:3349: MetaClip2ModelTest doesn't have the attribute `_torch_compile_train_cls`.",
|
| 17759 |
-
"count": 1
|
| 17760 |
-
},
|
| 17761 |
-
{
|
| 17762 |
-
"status": "skipped",
|
| 17763 |
-
"test": "[1] tests/test_modeling_common.py:3349: MetaClip2ForImageClassificationModelTest doesn't have the attribute `_torch_compile_train_cls`.",
|
| 17764 |
-
"count": 1
|
| 17765 |
-
}
|
| 17766 |
-
]
|
| 17767 |
-
},
|
| 17768 |
{
|
| 17769 |
"model": "mgp_str",
|
| 17770 |
"results": [
|
|
|
|
| 3 |
"machine_type": "multi-gpu",
|
| 4 |
"commit_hash": "1ae4d91",
|
| 5 |
"total_status_count": {
|
| 6 |
+
"passed": 1645,
|
| 7 |
"failed": 124,
|
| 8 |
+
"skipped": 5436,
|
| 9 |
"error": 0,
|
| 10 |
"null": 0
|
| 11 |
},
|
|
|
|
| 17611 |
}
|
| 17612 |
]
|
| 17613 |
},
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 17614 |
{
|
| 17615 |
"model": "mgp_str",
|
| 17616 |
"results": [
|