Upload 2025-12-04/ci_results_run_models_gpu/collated_reports_multi-gpu_a48d68c.json with huggingface_hub
Browse files
2025-12-04/ci_results_run_models_gpu/collated_reports_multi-gpu_a48d68c.json
CHANGED
|
@@ -3,9 +3,9 @@
|
|
| 3 |
"machine_type": "multi-gpu",
|
| 4 |
"commit_hash": "a48d68c",
|
| 5 |
"total_status_count": {
|
| 6 |
-
"passed":
|
| 7 |
"failed": 110,
|
| 8 |
-
"skipped":
|
| 9 |
"error": 0,
|
| 10 |
"null": 0
|
| 11 |
},
|
|
@@ -16942,6 +16942,1034 @@
|
|
| 16942 |
}
|
| 16943 |
]
|
| 16944 |
},
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 16945 |
{
|
| 16946 |
"model": "mgp_str",
|
| 16947 |
"results": [
|
|
|
|
| 3 |
"machine_type": "multi-gpu",
|
| 4 |
"commit_hash": "a48d68c",
|
| 5 |
"total_status_count": {
|
| 6 |
+
"passed": 1689,
|
| 7 |
"failed": 110,
|
| 8 |
+
"skipped": 5551,
|
| 9 |
"error": 0,
|
| 10 |
"null": 0
|
| 11 |
},
|
|
|
|
| 16942 |
}
|
| 16943 |
]
|
| 16944 |
},
|
| 16945 |
+
{
|
| 16946 |
+
"model": "m2m_100",
|
| 16947 |
+
"results": [
|
| 16948 |
+
{
|
| 16949 |
+
"status": "passed",
|
| 16950 |
+
"test": "tests/models/m2m_100/test_modeling_m2m_100.py::M2M100ModelTest::test_eager_matches_fa2_generate",
|
| 16951 |
+
"count": 1
|
| 16952 |
+
},
|
| 16953 |
+
{
|
| 16954 |
+
"status": "passed",
|
| 16955 |
+
"test": "tests/models/m2m_100/test_modeling_m2m_100.py::M2M100ModelTest::test_flash_attn_2_fp32_ln",
|
| 16956 |
+
"count": 1
|
| 16957 |
+
},
|
| 16958 |
+
{
|
| 16959 |
+
"status": "passed",
|
| 16960 |
+
"test": "tests/models/m2m_100/test_modeling_m2m_100.py::M2M100ModelTest::test_flash_attn_2_from_config",
|
| 16961 |
+
"count": 1
|
| 16962 |
+
},
|
| 16963 |
+
{
|
| 16964 |
+
"status": "passed",
|
| 16965 |
+
"test": "tests/models/m2m_100/test_modeling_m2m_100.py::M2M100ModelTest::test_flash_attn_2_inference_equivalence",
|
| 16966 |
+
"count": 1
|
| 16967 |
+
},
|
| 16968 |
+
{
|
| 16969 |
+
"status": "passed",
|
| 16970 |
+
"test": "tests/models/m2m_100/test_modeling_m2m_100.py::M2M100ModelTest::test_flash_attn_2_inference_equivalence_right_padding",
|
| 16971 |
+
"count": 1
|
| 16972 |
+
},
|
| 16973 |
+
{
|
| 16974 |
+
"status": "passed",
|
| 16975 |
+
"test": "tests/models/m2m_100/test_modeling_m2m_100.py::M2M100ModelIntegrationTests::test_flash_attn_2_seq_to_seq_generation",
|
| 16976 |
+
"count": 1
|
| 16977 |
+
},
|
| 16978 |
+
{
|
| 16979 |
+
"status": "skipped",
|
| 16980 |
+
"test": "[1] tests/generation/test_utils.py:1857: test requires Flash Attention 3",
|
| 16981 |
+
"count": 1
|
| 16982 |
+
},
|
| 16983 |
+
{
|
| 16984 |
+
"status": "skipped",
|
| 16985 |
+
"test": "[1] tests/generation/test_utils.py:1865: Model is an encoder-decoder",
|
| 16986 |
+
"count": 1
|
| 16987 |
+
},
|
| 16988 |
+
{
|
| 16989 |
+
"status": "skipped",
|
| 16990 |
+
"test": "[1] tests/generation/test_utils.py:2082: M2M100ForConditionalGeneration does not support new attention mask API",
|
| 16991 |
+
"count": 1
|
| 16992 |
+
},
|
| 16993 |
+
{
|
| 16994 |
+
"status": "skipped",
|
| 16995 |
+
"test": "[1] tests/generation/test_utils.py:2089: M2M100ForConditionalGeneration does not support new attention mask API",
|
| 16996 |
+
"count": 1
|
| 16997 |
+
},
|
| 16998 |
+
{
|
| 16999 |
+
"status": "skipped",
|
| 17000 |
+
"test": "[1] tests/generation/test_utils.py:2098: test requires Flash Attention 3",
|
| 17001 |
+
"count": 1
|
| 17002 |
+
},
|
| 17003 |
+
{
|
| 17004 |
+
"status": "skipped",
|
| 17005 |
+
"test": "[1] tests/generation/test_utils.py:2105: test requires Flash Attention 3",
|
| 17006 |
+
"count": 1
|
| 17007 |
+
},
|
| 17008 |
+
{
|
| 17009 |
+
"status": "skipped",
|
| 17010 |
+
"test": "[1] tests/test_modeling_common.py:3367: M2M100ModelTest doesn't have the attribute `_torch_compile_train_cls`.",
|
| 17011 |
+
"count": 1
|
| 17012 |
+
},
|
| 17013 |
+
{
|
| 17014 |
+
"status": "skipped",
|
| 17015 |
+
"test": "[1] tests/test_modeling_common.py:3294: This model is not a composite model!",
|
| 17016 |
+
"count": 1
|
| 17017 |
+
},
|
| 17018 |
+
{
|
| 17019 |
+
"status": "skipped",
|
| 17020 |
+
"test": "[1] tests/test_modeling_common.py:3300: test requires Flash Attention 3",
|
| 17021 |
+
"count": 1
|
| 17022 |
+
},
|
| 17023 |
+
{
|
| 17024 |
+
"status": "skipped",
|
| 17025 |
+
"test": "[1] tests/test_modeling_common.py:3469: test requires Flash Attention 3",
|
| 17026 |
+
"count": 1
|
| 17027 |
+
},
|
| 17028 |
+
{
|
| 17029 |
+
"status": "skipped",
|
| 17030 |
+
"test": "[1] tests/test_modeling_common.py:2910: test requires Flash Attention 3",
|
| 17031 |
+
"count": 1
|
| 17032 |
+
},
|
| 17033 |
+
{
|
| 17034 |
+
"status": "skipped",
|
| 17035 |
+
"test": "[1] tests/test_modeling_common.py:2918: test requires Flash Attention 3",
|
| 17036 |
+
"count": 1
|
| 17037 |
+
},
|
| 17038 |
+
{
|
| 17039 |
+
"status": "skipped",
|
| 17040 |
+
"test": "[1] tests/test_modeling_common.py:2876: test requires the kernels library",
|
| 17041 |
+
"count": 1
|
| 17042 |
+
},
|
| 17043 |
+
{
|
| 17044 |
+
"status": "skipped",
|
| 17045 |
+
"test": "[1] tests/test_modeling_common.py:2884: test requires MPS",
|
| 17046 |
+
"count": 1
|
| 17047 |
+
}
|
| 17048 |
+
]
|
| 17049 |
+
},
|
| 17050 |
+
{
|
| 17051 |
+
"model": "mamba2",
|
| 17052 |
+
"results": [
|
| 17053 |
+
{
|
| 17054 |
+
"status": "skipped",
|
| 17055 |
+
"test": "[1] tests/generation/test_utils.py:1849: Mamba2ForCausalLM does not support `attn_implementation=flash_attention_2`",
|
| 17056 |
+
"count": 1
|
| 17057 |
+
},
|
| 17058 |
+
{
|
| 17059 |
+
"status": "skipped",
|
| 17060 |
+
"test": "[1] tests/generation/test_utils.py:1857: test requires Flash Attention 3",
|
| 17061 |
+
"count": 1
|
| 17062 |
+
},
|
| 17063 |
+
{
|
| 17064 |
+
"status": "skipped",
|
| 17065 |
+
"test": "[1] tests/generation/test_utils.py:1865: Mamba2ForCausalLM does not support Flash Attention.",
|
| 17066 |
+
"count": 1
|
| 17067 |
+
},
|
| 17068 |
+
{
|
| 17069 |
+
"status": "skipped",
|
| 17070 |
+
"test": "[1] tests/generation/test_utils.py:2082: Model architecture does not support attentions",
|
| 17071 |
+
"count": 1
|
| 17072 |
+
},
|
| 17073 |
+
{
|
| 17074 |
+
"status": "skipped",
|
| 17075 |
+
"test": "[1] tests/generation/test_utils.py:2089: Model architecture does not support attentions",
|
| 17076 |
+
"count": 1
|
| 17077 |
+
},
|
| 17078 |
+
{
|
| 17079 |
+
"status": "skipped",
|
| 17080 |
+
"test": "[1] tests/generation/test_utils.py:2098: test requires Flash Attention 3",
|
| 17081 |
+
"count": 1
|
| 17082 |
+
},
|
| 17083 |
+
{
|
| 17084 |
+
"status": "skipped",
|
| 17085 |
+
"test": "[1] tests/generation/test_utils.py:2105: test requires Flash Attention 3",
|
| 17086 |
+
"count": 1
|
| 17087 |
+
},
|
| 17088 |
+
{
|
| 17089 |
+
"status": "skipped",
|
| 17090 |
+
"test": "[1] tests/test_modeling_common.py:3367: Mamba2ModelTest doesn't have the attribute `_torch_compile_train_cls`.",
|
| 17091 |
+
"count": 1
|
| 17092 |
+
},
|
| 17093 |
+
{
|
| 17094 |
+
"status": "skipped",
|
| 17095 |
+
"test": "[1] tests/test_modeling_common.py:3294: Model architecture does not support attentions",
|
| 17096 |
+
"count": 1
|
| 17097 |
+
},
|
| 17098 |
+
{
|
| 17099 |
+
"status": "skipped",
|
| 17100 |
+
"test": "[1] tests/test_modeling_common.py:3306: Model architecture does not support attentions",
|
| 17101 |
+
"count": 1
|
| 17102 |
+
},
|
| 17103 |
+
{
|
| 17104 |
+
"status": "skipped",
|
| 17105 |
+
"test": "[1] tests/test_modeling_common.py:3462: Model architecture does not support attentions",
|
| 17106 |
+
"count": 1
|
| 17107 |
+
},
|
| 17108 |
+
{
|
| 17109 |
+
"status": "skipped",
|
| 17110 |
+
"test": "[1] tests/test_modeling_common.py:2894: Model architecture does not support attentions",
|
| 17111 |
+
"count": 1
|
| 17112 |
+
},
|
| 17113 |
+
{
|
| 17114 |
+
"status": "skipped",
|
| 17115 |
+
"test": "[1] tests/test_modeling_common.py:2902: Model architecture does not support attentions",
|
| 17116 |
+
"count": 1
|
| 17117 |
+
},
|
| 17118 |
+
{
|
| 17119 |
+
"status": "skipped",
|
| 17120 |
+
"test": "[1] tests/test_modeling_common.py:3300: test requires Flash Attention 3",
|
| 17121 |
+
"count": 1
|
| 17122 |
+
},
|
| 17123 |
+
{
|
| 17124 |
+
"status": "skipped",
|
| 17125 |
+
"test": "[1] tests/test_modeling_common.py:3469: test requires Flash Attention 3",
|
| 17126 |
+
"count": 1
|
| 17127 |
+
},
|
| 17128 |
+
{
|
| 17129 |
+
"status": "skipped",
|
| 17130 |
+
"test": "[1] tests/test_modeling_common.py:2910: test requires Flash Attention 3",
|
| 17131 |
+
"count": 1
|
| 17132 |
+
},
|
| 17133 |
+
{
|
| 17134 |
+
"status": "skipped",
|
| 17135 |
+
"test": "[1] tests/test_modeling_common.py:2918: test requires Flash Attention 3",
|
| 17136 |
+
"count": 1
|
| 17137 |
+
},
|
| 17138 |
+
{
|
| 17139 |
+
"status": "skipped",
|
| 17140 |
+
"test": "[1] tests/test_modeling_common.py:2876: test requires the kernels library",
|
| 17141 |
+
"count": 1
|
| 17142 |
+
},
|
| 17143 |
+
{
|
| 17144 |
+
"status": "skipped",
|
| 17145 |
+
"test": "[1] tests/test_modeling_common.py:2884: test requires MPS",
|
| 17146 |
+
"count": 1
|
| 17147 |
+
}
|
| 17148 |
+
]
|
| 17149 |
+
},
|
| 17150 |
+
{
|
| 17151 |
+
"model": "mamba",
|
| 17152 |
+
"results": [
|
| 17153 |
+
{
|
| 17154 |
+
"status": "skipped",
|
| 17155 |
+
"test": "[1] tests/generation/test_utils.py:1849: MambaForCausalLM does not support `attn_implementation=flash_attention_2`",
|
| 17156 |
+
"count": 1
|
| 17157 |
+
},
|
| 17158 |
+
{
|
| 17159 |
+
"status": "skipped",
|
| 17160 |
+
"test": "[1] tests/generation/test_utils.py:1857: test requires Flash Attention 3",
|
| 17161 |
+
"count": 1
|
| 17162 |
+
},
|
| 17163 |
+
{
|
| 17164 |
+
"status": "skipped",
|
| 17165 |
+
"test": "[1] tests/generation/test_utils.py:1865: MambaForCausalLM does not support Flash Attention.",
|
| 17166 |
+
"count": 1
|
| 17167 |
+
},
|
| 17168 |
+
{
|
| 17169 |
+
"status": "skipped",
|
| 17170 |
+
"test": "[1] tests/generation/test_utils.py:2082: Model architecture does not support attentions",
|
| 17171 |
+
"count": 1
|
| 17172 |
+
},
|
| 17173 |
+
{
|
| 17174 |
+
"status": "skipped",
|
| 17175 |
+
"test": "[1] tests/generation/test_utils.py:2089: Model architecture does not support attentions",
|
| 17176 |
+
"count": 1
|
| 17177 |
+
},
|
| 17178 |
+
{
|
| 17179 |
+
"status": "skipped",
|
| 17180 |
+
"test": "[1] tests/generation/test_utils.py:2098: test requires Flash Attention 3",
|
| 17181 |
+
"count": 1
|
| 17182 |
+
},
|
| 17183 |
+
{
|
| 17184 |
+
"status": "skipped",
|
| 17185 |
+
"test": "[1] tests/generation/test_utils.py:2105: test requires Flash Attention 3",
|
| 17186 |
+
"count": 1
|
| 17187 |
+
},
|
| 17188 |
+
{
|
| 17189 |
+
"status": "skipped",
|
| 17190 |
+
"test": "[1] tests/test_modeling_common.py:3367: MambaModelTest doesn't have the attribute `_torch_compile_train_cls`.",
|
| 17191 |
+
"count": 1
|
| 17192 |
+
},
|
| 17193 |
+
{
|
| 17194 |
+
"status": "skipped",
|
| 17195 |
+
"test": "[1] tests/test_modeling_common.py:3294: Model architecture does not support attentions",
|
| 17196 |
+
"count": 1
|
| 17197 |
+
},
|
| 17198 |
+
{
|
| 17199 |
+
"status": "skipped",
|
| 17200 |
+
"test": "[1] tests/test_modeling_common.py:3306: Model architecture does not support attentions",
|
| 17201 |
+
"count": 1
|
| 17202 |
+
},
|
| 17203 |
+
{
|
| 17204 |
+
"status": "skipped",
|
| 17205 |
+
"test": "[1] tests/test_modeling_common.py:3462: Model architecture does not support attentions",
|
| 17206 |
+
"count": 1
|
| 17207 |
+
},
|
| 17208 |
+
{
|
| 17209 |
+
"status": "skipped",
|
| 17210 |
+
"test": "[1] tests/test_modeling_common.py:2894: Model architecture does not support attentions",
|
| 17211 |
+
"count": 1
|
| 17212 |
+
},
|
| 17213 |
+
{
|
| 17214 |
+
"status": "skipped",
|
| 17215 |
+
"test": "[1] tests/test_modeling_common.py:2902: Model architecture does not support attentions",
|
| 17216 |
+
"count": 1
|
| 17217 |
+
},
|
| 17218 |
+
{
|
| 17219 |
+
"status": "skipped",
|
| 17220 |
+
"test": "[1] tests/test_modeling_common.py:3300: test requires Flash Attention 3",
|
| 17221 |
+
"count": 1
|
| 17222 |
+
},
|
| 17223 |
+
{
|
| 17224 |
+
"status": "skipped",
|
| 17225 |
+
"test": "[1] tests/test_modeling_common.py:3469: test requires Flash Attention 3",
|
| 17226 |
+
"count": 1
|
| 17227 |
+
},
|
| 17228 |
+
{
|
| 17229 |
+
"status": "skipped",
|
| 17230 |
+
"test": "[1] tests/test_modeling_common.py:2910: test requires Flash Attention 3",
|
| 17231 |
+
"count": 1
|
| 17232 |
+
},
|
| 17233 |
+
{
|
| 17234 |
+
"status": "skipped",
|
| 17235 |
+
"test": "[1] tests/test_modeling_common.py:2918: test requires Flash Attention 3",
|
| 17236 |
+
"count": 1
|
| 17237 |
+
},
|
| 17238 |
+
{
|
| 17239 |
+
"status": "skipped",
|
| 17240 |
+
"test": "[1] tests/test_modeling_common.py:2876: test requires the kernels library",
|
| 17241 |
+
"count": 1
|
| 17242 |
+
},
|
| 17243 |
+
{
|
| 17244 |
+
"status": "skipped",
|
| 17245 |
+
"test": "[1] tests/test_modeling_common.py:2884: test requires MPS",
|
| 17246 |
+
"count": 1
|
| 17247 |
+
}
|
| 17248 |
+
]
|
| 17249 |
+
},
|
| 17250 |
+
{
|
| 17251 |
+
"model": "marian",
|
| 17252 |
+
"results": [
|
| 17253 |
+
{
|
| 17254 |
+
"status": "passed",
|
| 17255 |
+
"test": "tests/models/marian/test_modeling_marian.py::MarianModelTest::test_eager_matches_fa2_generate",
|
| 17256 |
+
"count": 1
|
| 17257 |
+
},
|
| 17258 |
+
{
|
| 17259 |
+
"status": "passed",
|
| 17260 |
+
"test": "tests/models/marian/test_modeling_marian.py::MarianModelTest::test_flash_attn_2_fp32_ln",
|
| 17261 |
+
"count": 1
|
| 17262 |
+
},
|
| 17263 |
+
{
|
| 17264 |
+
"status": "passed",
|
| 17265 |
+
"test": "tests/models/marian/test_modeling_marian.py::MarianModelTest::test_flash_attn_2_from_config",
|
| 17266 |
+
"count": 1
|
| 17267 |
+
},
|
| 17268 |
+
{
|
| 17269 |
+
"status": "passed",
|
| 17270 |
+
"test": "tests/models/marian/test_modeling_marian.py::MarianModelTest::test_flash_attn_2_inference_equivalence",
|
| 17271 |
+
"count": 1
|
| 17272 |
+
},
|
| 17273 |
+
{
|
| 17274 |
+
"status": "passed",
|
| 17275 |
+
"test": "tests/models/marian/test_modeling_marian.py::MarianModelTest::test_flash_attn_2_inference_equivalence_right_padding",
|
| 17276 |
+
"count": 1
|
| 17277 |
+
},
|
| 17278 |
+
{
|
| 17279 |
+
"status": "passed",
|
| 17280 |
+
"test": "tests/models/marian/test_modeling_marian.py::MarianStandaloneDecoderModelTest::test_eager_matches_fa2_generate",
|
| 17281 |
+
"count": 1
|
| 17282 |
+
},
|
| 17283 |
+
{
|
| 17284 |
+
"status": "passed",
|
| 17285 |
+
"test": "tests/models/marian/test_modeling_marian.py::MarianStandaloneDecoderModelTest::test_flash_attn_2_fp32_ln",
|
| 17286 |
+
"count": 1
|
| 17287 |
+
},
|
| 17288 |
+
{
|
| 17289 |
+
"status": "passed",
|
| 17290 |
+
"test": "tests/models/marian/test_modeling_marian.py::MarianStandaloneDecoderModelTest::test_flash_attn_2_from_config",
|
| 17291 |
+
"count": 1
|
| 17292 |
+
},
|
| 17293 |
+
{
|
| 17294 |
+
"status": "passed",
|
| 17295 |
+
"test": "tests/models/marian/test_modeling_marian.py::MarianStandaloneDecoderModelTest::test_flash_attn_2_inference_equivalence",
|
| 17296 |
+
"count": 1
|
| 17297 |
+
},
|
| 17298 |
+
{
|
| 17299 |
+
"status": "passed",
|
| 17300 |
+
"test": "tests/models/marian/test_modeling_marian.py::MarianStandaloneDecoderModelTest::test_flash_attn_2_inference_equivalence_right_padding",
|
| 17301 |
+
"count": 1
|
| 17302 |
+
},
|
| 17303 |
+
{
|
| 17304 |
+
"status": "skipped",
|
| 17305 |
+
"test": "[2] tests/generation/test_utils.py:1857: test requires Flash Attention 3",
|
| 17306 |
+
"count": 2
|
| 17307 |
+
},
|
| 17308 |
+
{
|
| 17309 |
+
"status": "skipped",
|
| 17310 |
+
"test": "[1] tests/generation/test_utils.py:1865: Model is an encoder-decoder",
|
| 17311 |
+
"count": 1
|
| 17312 |
+
},
|
| 17313 |
+
{
|
| 17314 |
+
"status": "skipped",
|
| 17315 |
+
"test": "[1] tests/generation/test_utils.py:2082: MarianMTModel does not support new attention mask API",
|
| 17316 |
+
"count": 1
|
| 17317 |
+
},
|
| 17318 |
+
{
|
| 17319 |
+
"status": "skipped",
|
| 17320 |
+
"test": "[1] tests/generation/test_utils.py:2089: MarianMTModel does not support new attention mask API",
|
| 17321 |
+
"count": 1
|
| 17322 |
+
},
|
| 17323 |
+
{
|
| 17324 |
+
"status": "skipped",
|
| 17325 |
+
"test": "[2] tests/generation/test_utils.py:2098: test requires Flash Attention 3",
|
| 17326 |
+
"count": 2
|
| 17327 |
+
},
|
| 17328 |
+
{
|
| 17329 |
+
"status": "skipped",
|
| 17330 |
+
"test": "[2] tests/generation/test_utils.py:2105: test requires Flash Attention 3",
|
| 17331 |
+
"count": 2
|
| 17332 |
+
},
|
| 17333 |
+
{
|
| 17334 |
+
"status": "skipped",
|
| 17335 |
+
"test": "[1] tests/test_modeling_common.py:3367: MarianModelTest doesn't have the attribute `_torch_compile_train_cls`.",
|
| 17336 |
+
"count": 1
|
| 17337 |
+
},
|
| 17338 |
+
{
|
| 17339 |
+
"status": "skipped",
|
| 17340 |
+
"test": "[2] tests/test_modeling_common.py:3294: This model is not a composite model!",
|
| 17341 |
+
"count": 2
|
| 17342 |
+
},
|
| 17343 |
+
{
|
| 17344 |
+
"status": "skipped",
|
| 17345 |
+
"test": "[2] tests/test_modeling_common.py:3300: test requires Flash Attention 3",
|
| 17346 |
+
"count": 2
|
| 17347 |
+
},
|
| 17348 |
+
{
|
| 17349 |
+
"status": "skipped",
|
| 17350 |
+
"test": "[2] tests/test_modeling_common.py:3469: test requires Flash Attention 3",
|
| 17351 |
+
"count": 2
|
| 17352 |
+
},
|
| 17353 |
+
{
|
| 17354 |
+
"status": "skipped",
|
| 17355 |
+
"test": "[2] tests/test_modeling_common.py:2910: test requires Flash Attention 3",
|
| 17356 |
+
"count": 2
|
| 17357 |
+
},
|
| 17358 |
+
{
|
| 17359 |
+
"status": "skipped",
|
| 17360 |
+
"test": "[2] tests/test_modeling_common.py:2918: test requires Flash Attention 3",
|
| 17361 |
+
"count": 2
|
| 17362 |
+
},
|
| 17363 |
+
{
|
| 17364 |
+
"status": "skipped",
|
| 17365 |
+
"test": "[2] tests/test_modeling_common.py:2876: test requires the kernels library",
|
| 17366 |
+
"count": 2
|
| 17367 |
+
},
|
| 17368 |
+
{
|
| 17369 |
+
"status": "skipped",
|
| 17370 |
+
"test": "[2] tests/test_modeling_common.py:2884: test requires MPS",
|
| 17371 |
+
"count": 2
|
| 17372 |
+
},
|
| 17373 |
+
{
|
| 17374 |
+
"status": "skipped",
|
| 17375 |
+
"test": "[1] tests/generation/test_utils.py:1865: Model does not support position_ids",
|
| 17376 |
+
"count": 1
|
| 17377 |
+
},
|
| 17378 |
+
{
|
| 17379 |
+
"status": "skipped",
|
| 17380 |
+
"test": "[1] tests/generation/test_utils.py:2082: MarianForCausalLM does not support new attention mask API",
|
| 17381 |
+
"count": 1
|
| 17382 |
+
},
|
| 17383 |
+
{
|
| 17384 |
+
"status": "skipped",
|
| 17385 |
+
"test": "[1] tests/generation/test_utils.py:2089: MarianForCausalLM does not support new attention mask API",
|
| 17386 |
+
"count": 1
|
| 17387 |
+
},
|
| 17388 |
+
{
|
| 17389 |
+
"status": "skipped",
|
| 17390 |
+
"test": "[1] tests/test_modeling_common.py:3367: MarianStandaloneDecoderModelTest doesn't have the attribute `_torch_compile_train_cls`.",
|
| 17391 |
+
"count": 1
|
| 17392 |
+
}
|
| 17393 |
+
]
|
| 17394 |
+
},
|
| 17395 |
+
{
|
| 17396 |
+
"model": "markuplm",
|
| 17397 |
+
"results": [
|
| 17398 |
+
{
|
| 17399 |
+
"status": "passed",
|
| 17400 |
+
"test": "tests/models/markuplm/test_modeling_markuplm.py::MarkupLMModelTest::test_flash_attn_2_fp32_ln",
|
| 17401 |
+
"count": 1
|
| 17402 |
+
},
|
| 17403 |
+
{
|
| 17404 |
+
"status": "passed",
|
| 17405 |
+
"test": "tests/models/markuplm/test_modeling_markuplm.py::MarkupLMModelTest::test_flash_attn_2_from_config",
|
| 17406 |
+
"count": 1
|
| 17407 |
+
},
|
| 17408 |
+
{
|
| 17409 |
+
"status": "skipped",
|
| 17410 |
+
"test": "[1] tests/test_modeling_common.py:3367: MarkupLMModelTest doesn't have the attribute `_torch_compile_train_cls`.",
|
| 17411 |
+
"count": 1
|
| 17412 |
+
},
|
| 17413 |
+
{
|
| 17414 |
+
"status": "skipped",
|
| 17415 |
+
"test": "[1] tests/test_modeling_common.py:3294: This model is not a composite model!",
|
| 17416 |
+
"count": 1
|
| 17417 |
+
},
|
| 17418 |
+
{
|
| 17419 |
+
"status": "skipped",
|
| 17420 |
+
"test": "[1] tests/test_modeling_common.py:2894: Model architecture does not support flash_attention_2, or setting its attention dynamically",
|
| 17421 |
+
"count": 1
|
| 17422 |
+
},
|
| 17423 |
+
{
|
| 17424 |
+
"status": "skipped",
|
| 17425 |
+
"test": "[1] tests/test_modeling_common.py:2902: Model architecture does not support flash_attention_2, or setting its attention dynamically",
|
| 17426 |
+
"count": 1
|
| 17427 |
+
},
|
| 17428 |
+
{
|
| 17429 |
+
"status": "skipped",
|
| 17430 |
+
"test": "[1] tests/test_modeling_common.py:3300: test requires Flash Attention 3",
|
| 17431 |
+
"count": 1
|
| 17432 |
+
},
|
| 17433 |
+
{
|
| 17434 |
+
"status": "skipped",
|
| 17435 |
+
"test": "[1] tests/test_modeling_common.py:3469: test requires Flash Attention 3",
|
| 17436 |
+
"count": 1
|
| 17437 |
+
},
|
| 17438 |
+
{
|
| 17439 |
+
"status": "skipped",
|
| 17440 |
+
"test": "[1] tests/test_modeling_common.py:2910: test requires Flash Attention 3",
|
| 17441 |
+
"count": 1
|
| 17442 |
+
},
|
| 17443 |
+
{
|
| 17444 |
+
"status": "skipped",
|
| 17445 |
+
"test": "[1] tests/test_modeling_common.py:2918: test requires Flash Attention 3",
|
| 17446 |
+
"count": 1
|
| 17447 |
+
},
|
| 17448 |
+
{
|
| 17449 |
+
"status": "skipped",
|
| 17450 |
+
"test": "[1] tests/test_modeling_common.py:2876: test requires the kernels library",
|
| 17451 |
+
"count": 1
|
| 17452 |
+
},
|
| 17453 |
+
{
|
| 17454 |
+
"status": "skipped",
|
| 17455 |
+
"test": "[1] tests/test_modeling_common.py:2884: test requires MPS",
|
| 17456 |
+
"count": 1
|
| 17457 |
+
}
|
| 17458 |
+
]
|
| 17459 |
+
},
|
| 17460 |
+
{
|
| 17461 |
+
"model": "mask2former",
|
| 17462 |
+
"results": [
|
| 17463 |
+
{
|
| 17464 |
+
"status": "passed",
|
| 17465 |
+
"test": "tests/models/mask2former/test_modeling_mask2former.py::Mask2FormerModelTest::test_flash_attn_2_fp32_ln",
|
| 17466 |
+
"count": 1
|
| 17467 |
+
},
|
| 17468 |
+
{
|
| 17469 |
+
"status": "passed",
|
| 17470 |
+
"test": "tests/models/mask2former/test_modeling_mask2former.py::Mask2FormerModelTest::test_flash_attn_2_from_config",
|
| 17471 |
+
"count": 1
|
| 17472 |
+
},
|
| 17473 |
+
{
|
| 17474 |
+
"status": "skipped",
|
| 17475 |
+
"test": "[1] tests/test_modeling_common.py:3367: Mask2FormerModelTest doesn't have the attribute `_torch_compile_train_cls`.",
|
| 17476 |
+
"count": 1
|
| 17477 |
+
},
|
| 17478 |
+
{
|
| 17479 |
+
"status": "skipped",
|
| 17480 |
+
"test": "[1] tests/test_modeling_common.py:3294: This model is not a composite model!",
|
| 17481 |
+
"count": 1
|
| 17482 |
+
},
|
| 17483 |
+
{
|
| 17484 |
+
"status": "skipped",
|
| 17485 |
+
"test": "[1] tests/test_modeling_common.py:2894: Model architecture does not support flash_attention_2, or setting its attention dynamically",
|
| 17486 |
+
"count": 1
|
| 17487 |
+
},
|
| 17488 |
+
{
|
| 17489 |
+
"status": "skipped",
|
| 17490 |
+
"test": "[1] tests/test_modeling_common.py:2902: Model architecture does not support flash_attention_2, or setting its attention dynamically",
|
| 17491 |
+
"count": 1
|
| 17492 |
+
},
|
| 17493 |
+
{
|
| 17494 |
+
"status": "skipped",
|
| 17495 |
+
"test": "[1] tests/test_modeling_common.py:3300: test requires Flash Attention 3",
|
| 17496 |
+
"count": 1
|
| 17497 |
+
},
|
| 17498 |
+
{
|
| 17499 |
+
"status": "skipped",
|
| 17500 |
+
"test": "[1] tests/test_modeling_common.py:3469: test requires Flash Attention 3",
|
| 17501 |
+
"count": 1
|
| 17502 |
+
},
|
| 17503 |
+
{
|
| 17504 |
+
"status": "skipped",
|
| 17505 |
+
"test": "[1] tests/test_modeling_common.py:2910: test requires Flash Attention 3",
|
| 17506 |
+
"count": 1
|
| 17507 |
+
},
|
| 17508 |
+
{
|
| 17509 |
+
"status": "skipped",
|
| 17510 |
+
"test": "[1] tests/test_modeling_common.py:2918: test requires Flash Attention 3",
|
| 17511 |
+
"count": 1
|
| 17512 |
+
},
|
| 17513 |
+
{
|
| 17514 |
+
"status": "skipped",
|
| 17515 |
+
"test": "[1] tests/test_modeling_common.py:2876: test requires the kernels library",
|
| 17516 |
+
"count": 1
|
| 17517 |
+
},
|
| 17518 |
+
{
|
| 17519 |
+
"status": "skipped",
|
| 17520 |
+
"test": "[1] tests/test_modeling_common.py:2884: test requires MPS",
|
| 17521 |
+
"count": 1
|
| 17522 |
+
}
|
| 17523 |
+
]
|
| 17524 |
+
},
|
| 17525 |
+
{
|
| 17526 |
+
"model": "maskformer",
|
| 17527 |
+
"results": [
|
| 17528 |
+
{
|
| 17529 |
+
"status": "passed",
|
| 17530 |
+
"test": "tests/models/maskformer/test_modeling_maskformer.py::MaskFormerModelTest::test_flash_attn_2_fp32_ln",
|
| 17531 |
+
"count": 1
|
| 17532 |
+
},
|
| 17533 |
+
{
|
| 17534 |
+
"status": "passed",
|
| 17535 |
+
"test": "tests/models/maskformer/test_modeling_maskformer.py::MaskFormerModelTest::test_flash_attn_2_from_config",
|
| 17536 |
+
"count": 1
|
| 17537 |
+
},
|
| 17538 |
+
{
|
| 17539 |
+
"status": "passed",
|
| 17540 |
+
"test": "tests/models/maskformer/test_modeling_maskformer_swin.py::MaskFormerSwinModelTest::test_flash_attn_2_fp32_ln",
|
| 17541 |
+
"count": 1
|
| 17542 |
+
},
|
| 17543 |
+
{
|
| 17544 |
+
"status": "passed",
|
| 17545 |
+
"test": "tests/models/maskformer/test_modeling_maskformer_swin.py::MaskFormerSwinModelTest::test_flash_attn_2_from_config",
|
| 17546 |
+
"count": 1
|
| 17547 |
+
},
|
| 17548 |
+
{
|
| 17549 |
+
"status": "skipped",
|
| 17550 |
+
"test": "[1] tests/test_modeling_common.py:3367: MaskFormerModelTest doesn't have the attribute `_torch_compile_train_cls`.",
|
| 17551 |
+
"count": 1
|
| 17552 |
+
},
|
| 17553 |
+
{
|
| 17554 |
+
"status": "skipped",
|
| 17555 |
+
"test": "[2] tests/test_modeling_common.py:3294: This model is not a composite model!",
|
| 17556 |
+
"count": 2
|
| 17557 |
+
},
|
| 17558 |
+
{
|
| 17559 |
+
"status": "skipped",
|
| 17560 |
+
"test": "[2] tests/test_modeling_common.py:2894: Model architecture does not support flash_attention_2, or setting its attention dynamically",
|
| 17561 |
+
"count": 2
|
| 17562 |
+
},
|
| 17563 |
+
{
|
| 17564 |
+
"status": "skipped",
|
| 17565 |
+
"test": "[2] tests/test_modeling_common.py:2902: Model architecture does not support flash_attention_2, or setting its attention dynamically",
|
| 17566 |
+
"count": 2
|
| 17567 |
+
},
|
| 17568 |
+
{
|
| 17569 |
+
"status": "skipped",
|
| 17570 |
+
"test": "[2] tests/test_modeling_common.py:3300: test requires Flash Attention 3",
|
| 17571 |
+
"count": 2
|
| 17572 |
+
},
|
| 17573 |
+
{
|
| 17574 |
+
"status": "skipped",
|
| 17575 |
+
"test": "[2] tests/test_modeling_common.py:3469: test requires Flash Attention 3",
|
| 17576 |
+
"count": 2
|
| 17577 |
+
},
|
| 17578 |
+
{
|
| 17579 |
+
"status": "skipped",
|
| 17580 |
+
"test": "[2] tests/test_modeling_common.py:2910: test requires Flash Attention 3",
|
| 17581 |
+
"count": 2
|
| 17582 |
+
},
|
| 17583 |
+
{
|
| 17584 |
+
"status": "skipped",
|
| 17585 |
+
"test": "[2] tests/test_modeling_common.py:2918: test requires Flash Attention 3",
|
| 17586 |
+
"count": 2
|
| 17587 |
+
},
|
| 17588 |
+
{
|
| 17589 |
+
"status": "skipped",
|
| 17590 |
+
"test": "[2] tests/test_modeling_common.py:2876: test requires the kernels library",
|
| 17591 |
+
"count": 2
|
| 17592 |
+
},
|
| 17593 |
+
{
|
| 17594 |
+
"status": "skipped",
|
| 17595 |
+
"test": "[2] tests/test_modeling_common.py:2884: test requires MPS",
|
| 17596 |
+
"count": 2
|
| 17597 |
+
},
|
| 17598 |
+
{
|
| 17599 |
+
"status": "skipped",
|
| 17600 |
+
"test": "[1] tests/test_modeling_common.py:3367: MaskFormerSwinModelTest doesn't have the attribute `_torch_compile_train_cls`.",
|
| 17601 |
+
"count": 1
|
| 17602 |
+
}
|
| 17603 |
+
]
|
| 17604 |
+
},
|
| 17605 |
+
{
|
| 17606 |
+
"model": "mbart50",
|
| 17607 |
+
"results": []
|
| 17608 |
+
},
|
| 17609 |
+
{
|
| 17610 |
+
"model": "mbart",
|
| 17611 |
+
"results": [
|
| 17612 |
+
{
|
| 17613 |
+
"status": "passed",
|
| 17614 |
+
"test": "tests/models/mbart/test_modeling_mbart.py::MBartModelTest::test_eager_matches_fa2_generate",
|
| 17615 |
+
"count": 1
|
| 17616 |
+
},
|
| 17617 |
+
{
|
| 17618 |
+
"status": "passed",
|
| 17619 |
+
"test": "tests/models/mbart/test_modeling_mbart.py::MBartModelTest::test_flash_attn_2_fp32_ln",
|
| 17620 |
+
"count": 1
|
| 17621 |
+
},
|
| 17622 |
+
{
|
| 17623 |
+
"status": "passed",
|
| 17624 |
+
"test": "tests/models/mbart/test_modeling_mbart.py::MBartModelTest::test_flash_attn_2_from_config",
|
| 17625 |
+
"count": 1
|
| 17626 |
+
},
|
| 17627 |
+
{
|
| 17628 |
+
"status": "passed",
|
| 17629 |
+
"test": "tests/models/mbart/test_modeling_mbart.py::MBartModelTest::test_flash_attn_2_inference_equivalence",
|
| 17630 |
+
"count": 1
|
| 17631 |
+
},
|
| 17632 |
+
{
|
| 17633 |
+
"status": "passed",
|
| 17634 |
+
"test": "tests/models/mbart/test_modeling_mbart.py::MBartModelTest::test_flash_attn_2_inference_equivalence_right_padding",
|
| 17635 |
+
"count": 1
|
| 17636 |
+
},
|
| 17637 |
+
{
|
| 17638 |
+
"status": "passed",
|
| 17639 |
+
"test": "tests/models/mbart/test_modeling_mbart.py::MBartStandaloneDecoderModelTest::test_eager_matches_fa2_generate",
|
| 17640 |
+
"count": 1
|
| 17641 |
+
},
|
| 17642 |
+
{
|
| 17643 |
+
"status": "passed",
|
| 17644 |
+
"test": "tests/models/mbart/test_modeling_mbart.py::MBartStandaloneDecoderModelTest::test_flash_attn_2_fp32_ln",
|
| 17645 |
+
"count": 1
|
| 17646 |
+
},
|
| 17647 |
+
{
|
| 17648 |
+
"status": "passed",
|
| 17649 |
+
"test": "tests/models/mbart/test_modeling_mbart.py::MBartStandaloneDecoderModelTest::test_flash_attn_2_from_config",
|
| 17650 |
+
"count": 1
|
| 17651 |
+
},
|
| 17652 |
+
{
|
| 17653 |
+
"status": "passed",
|
| 17654 |
+
"test": "tests/models/mbart/test_modeling_mbart.py::MBartStandaloneDecoderModelTest::test_flash_attn_2_inference_equivalence",
|
| 17655 |
+
"count": 1
|
| 17656 |
+
},
|
| 17657 |
+
{
|
| 17658 |
+
"status": "passed",
|
| 17659 |
+
"test": "tests/models/mbart/test_modeling_mbart.py::MBartStandaloneDecoderModelTest::test_flash_attn_2_inference_equivalence_right_padding",
|
| 17660 |
+
"count": 1
|
| 17661 |
+
},
|
| 17662 |
+
{
|
| 17663 |
+
"status": "skipped",
|
| 17664 |
+
"test": "[2] tests/generation/test_utils.py:1857: test requires Flash Attention 3",
|
| 17665 |
+
"count": 2
|
| 17666 |
+
},
|
| 17667 |
+
{
|
| 17668 |
+
"status": "skipped",
|
| 17669 |
+
"test": "[1] tests/generation/test_utils.py:1865: Model is an encoder-decoder",
|
| 17670 |
+
"count": 1
|
| 17671 |
+
},
|
| 17672 |
+
{
|
| 17673 |
+
"status": "skipped",
|
| 17674 |
+
"test": "[1] tests/generation/test_utils.py:2082: MBartForConditionalGeneration does not support new attention mask API",
|
| 17675 |
+
"count": 1
|
| 17676 |
+
},
|
| 17677 |
+
{
|
| 17678 |
+
"status": "skipped",
|
| 17679 |
+
"test": "[1] tests/generation/test_utils.py:2089: MBartForConditionalGeneration does not support new attention mask API",
|
| 17680 |
+
"count": 1
|
| 17681 |
+
},
|
| 17682 |
+
{
|
| 17683 |
+
"status": "skipped",
|
| 17684 |
+
"test": "[2] tests/generation/test_utils.py:2098: test requires Flash Attention 3",
|
| 17685 |
+
"count": 2
|
| 17686 |
+
},
|
| 17687 |
+
{
|
| 17688 |
+
"status": "skipped",
|
| 17689 |
+
"test": "[2] tests/generation/test_utils.py:2105: test requires Flash Attention 3",
|
| 17690 |
+
"count": 2
|
| 17691 |
+
},
|
| 17692 |
+
{
|
| 17693 |
+
"status": "skipped",
|
| 17694 |
+
"test": "[1] tests/test_modeling_common.py:3367: MBartModelTest doesn't have the attribute `_torch_compile_train_cls`.",
|
| 17695 |
+
"count": 1
|
| 17696 |
+
},
|
| 17697 |
+
{
|
| 17698 |
+
"status": "skipped",
|
| 17699 |
+
"test": "[2] tests/test_modeling_common.py:3294: This model is not a composite model!",
|
| 17700 |
+
"count": 2
|
| 17701 |
+
},
|
| 17702 |
+
{
|
| 17703 |
+
"status": "skipped",
|
| 17704 |
+
"test": "[2] tests/test_modeling_common.py:3300: test requires Flash Attention 3",
|
| 17705 |
+
"count": 2
|
| 17706 |
+
},
|
| 17707 |
+
{
|
| 17708 |
+
"status": "skipped",
|
| 17709 |
+
"test": "[2] tests/test_modeling_common.py:3469: test requires Flash Attention 3",
|
| 17710 |
+
"count": 2
|
| 17711 |
+
},
|
| 17712 |
+
{
|
| 17713 |
+
"status": "skipped",
|
| 17714 |
+
"test": "[2] tests/test_modeling_common.py:2910: test requires Flash Attention 3",
|
| 17715 |
+
"count": 2
|
| 17716 |
+
},
|
| 17717 |
+
{
|
| 17718 |
+
"status": "skipped",
|
| 17719 |
+
"test": "[2] tests/test_modeling_common.py:2918: test requires Flash Attention 3",
|
| 17720 |
+
"count": 2
|
| 17721 |
+
},
|
| 17722 |
+
{
|
| 17723 |
+
"status": "skipped",
|
| 17724 |
+
"test": "[2] tests/test_modeling_common.py:2876: test requires the kernels library",
|
| 17725 |
+
"count": 2
|
| 17726 |
+
},
|
| 17727 |
+
{
|
| 17728 |
+
"status": "skipped",
|
| 17729 |
+
"test": "[2] tests/test_modeling_common.py:2884: test requires MPS",
|
| 17730 |
+
"count": 2
|
| 17731 |
+
},
|
| 17732 |
+
{
|
| 17733 |
+
"status": "skipped",
|
| 17734 |
+
"test": "[1] tests/generation/test_utils.py:1865: Model does not support position_ids",
|
| 17735 |
+
"count": 1
|
| 17736 |
+
},
|
| 17737 |
+
{
|
| 17738 |
+
"status": "skipped",
|
| 17739 |
+
"test": "[1] tests/generation/test_utils.py:2082: MBartForCausalLM does not support new attention mask API",
|
| 17740 |
+
"count": 1
|
| 17741 |
+
},
|
| 17742 |
+
{
|
| 17743 |
+
"status": "skipped",
|
| 17744 |
+
"test": "[1] tests/generation/test_utils.py:2089: MBartForCausalLM does not support new attention mask API",
|
| 17745 |
+
"count": 1
|
| 17746 |
+
},
|
| 17747 |
+
{
|
| 17748 |
+
"status": "skipped",
|
| 17749 |
+
"test": "[1] tests/test_modeling_common.py:3367: MBartStandaloneDecoderModelTest doesn't have the attribute `_torch_compile_train_cls`.",
|
| 17750 |
+
"count": 1
|
| 17751 |
+
}
|
| 17752 |
+
]
|
| 17753 |
+
},
|
| 17754 |
+
{
|
| 17755 |
+
"model": "megatron_bert",
|
| 17756 |
+
"results": [
|
| 17757 |
+
{
|
| 17758 |
+
"status": "passed",
|
| 17759 |
+
"test": "tests/models/megatron_bert/test_modeling_megatron_bert.py::MegatronBertModelTest::test_flash_attn_2_fp32_ln",
|
| 17760 |
+
"count": 1
|
| 17761 |
+
},
|
| 17762 |
+
{
|
| 17763 |
+
"status": "passed",
|
| 17764 |
+
"test": "tests/models/megatron_bert/test_modeling_megatron_bert.py::MegatronBertModelTest::test_flash_attn_2_from_config",
|
| 17765 |
+
"count": 1
|
| 17766 |
+
},
|
| 17767 |
+
{
|
| 17768 |
+
"status": "skipped",
|
| 17769 |
+
"test": "[1] tests/test_modeling_common.py:3367: MegatronBertModelTest doesn't have the attribute `_torch_compile_train_cls`.",
|
| 17770 |
+
"count": 1
|
| 17771 |
+
},
|
| 17772 |
+
{
|
| 17773 |
+
"status": "skipped",
|
| 17774 |
+
"test": "[1] tests/test_modeling_common.py:3294: This model is not a composite model!",
|
| 17775 |
+
"count": 1
|
| 17776 |
+
},
|
| 17777 |
+
{
|
| 17778 |
+
"status": "skipped",
|
| 17779 |
+
"test": "[1] tests/test_modeling_common.py:2894: Model architecture does not support flash_attention_2, or setting its attention dynamically",
|
| 17780 |
+
"count": 1
|
| 17781 |
+
},
|
| 17782 |
+
{
|
| 17783 |
+
"status": "skipped",
|
| 17784 |
+
"test": "[1] tests/test_modeling_common.py:2902: Model architecture does not support flash_attention_2, or setting its attention dynamically",
|
| 17785 |
+
"count": 1
|
| 17786 |
+
},
|
| 17787 |
+
{
|
| 17788 |
+
"status": "skipped",
|
| 17789 |
+
"test": "[1] tests/test_modeling_common.py:3300: test requires Flash Attention 3",
|
| 17790 |
+
"count": 1
|
| 17791 |
+
},
|
| 17792 |
+
{
|
| 17793 |
+
"status": "skipped",
|
| 17794 |
+
"test": "[1] tests/test_modeling_common.py:3469: test requires Flash Attention 3",
|
| 17795 |
+
"count": 1
|
| 17796 |
+
},
|
| 17797 |
+
{
|
| 17798 |
+
"status": "skipped",
|
| 17799 |
+
"test": "[1] tests/test_modeling_common.py:2910: test requires Flash Attention 3",
|
| 17800 |
+
"count": 1
|
| 17801 |
+
},
|
| 17802 |
+
{
|
| 17803 |
+
"status": "skipped",
|
| 17804 |
+
"test": "[1] tests/test_modeling_common.py:2918: test requires Flash Attention 3",
|
| 17805 |
+
"count": 1
|
| 17806 |
+
},
|
| 17807 |
+
{
|
| 17808 |
+
"status": "skipped",
|
| 17809 |
+
"test": "[1] tests/test_modeling_common.py:2876: test requires the kernels library",
|
| 17810 |
+
"count": 1
|
| 17811 |
+
},
|
| 17812 |
+
{
|
| 17813 |
+
"status": "skipped",
|
| 17814 |
+
"test": "[1] tests/test_modeling_common.py:2884: test requires MPS",
|
| 17815 |
+
"count": 1
|
| 17816 |
+
}
|
| 17817 |
+
]
|
| 17818 |
+
},
|
| 17819 |
+
{
|
| 17820 |
+
"model": "megatron_gpt2",
|
| 17821 |
+
"results": []
|
| 17822 |
+
},
|
| 17823 |
+
{
|
| 17824 |
+
"model": "metaclip_2",
|
| 17825 |
+
"results": [
|
| 17826 |
+
{
|
| 17827 |
+
"status": "passed",
|
| 17828 |
+
"test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2VisionModelTest::test_flash_attn_2_fp32_ln",
|
| 17829 |
+
"count": 1
|
| 17830 |
+
},
|
| 17831 |
+
{
|
| 17832 |
+
"status": "passed",
|
| 17833 |
+
"test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2VisionModelTest::test_flash_attn_2_from_config",
|
| 17834 |
+
"count": 1
|
| 17835 |
+
},
|
| 17836 |
+
{
|
| 17837 |
+
"status": "passed",
|
| 17838 |
+
"test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2VisionModelTest::test_flash_attn_2_inference_equivalence",
|
| 17839 |
+
"count": 1
|
| 17840 |
+
},
|
| 17841 |
+
{
|
| 17842 |
+
"status": "passed",
|
| 17843 |
+
"test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2VisionModelTest::test_flash_attn_2_inference_equivalence_right_padding",
|
| 17844 |
+
"count": 1
|
| 17845 |
+
},
|
| 17846 |
+
{
|
| 17847 |
+
"status": "passed",
|
| 17848 |
+
"test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2TextModelTest::test_flash_attn_2_fp32_ln",
|
| 17849 |
+
"count": 1
|
| 17850 |
+
},
|
| 17851 |
+
{
|
| 17852 |
+
"status": "passed",
|
| 17853 |
+
"test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2TextModelTest::test_flash_attn_2_from_config",
|
| 17854 |
+
"count": 1
|
| 17855 |
+
},
|
| 17856 |
+
{
|
| 17857 |
+
"status": "passed",
|
| 17858 |
+
"test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2TextModelTest::test_flash_attn_2_inference_equivalence",
|
| 17859 |
+
"count": 1
|
| 17860 |
+
},
|
| 17861 |
+
{
|
| 17862 |
+
"status": "passed",
|
| 17863 |
+
"test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2TextModelTest::test_flash_attn_2_inference_equivalence_right_padding",
|
| 17864 |
+
"count": 1
|
| 17865 |
+
},
|
| 17866 |
+
{
|
| 17867 |
+
"status": "passed",
|
| 17868 |
+
"test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2ModelTest::test_flash_attn_2_can_dispatch_composite_models",
|
| 17869 |
+
"count": 1
|
| 17870 |
+
},
|
| 17871 |
+
{
|
| 17872 |
+
"status": "passed",
|
| 17873 |
+
"test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2ModelTest::test_flash_attn_2_fp32_ln",
|
| 17874 |
+
"count": 1
|
| 17875 |
+
},
|
| 17876 |
+
{
|
| 17877 |
+
"status": "passed",
|
| 17878 |
+
"test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2ModelTest::test_flash_attn_2_from_config",
|
| 17879 |
+
"count": 1
|
| 17880 |
+
},
|
| 17881 |
+
{
|
| 17882 |
+
"status": "passed",
|
| 17883 |
+
"test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2ModelTest::test_flash_attn_2_inference_equivalence",
|
| 17884 |
+
"count": 1
|
| 17885 |
+
},
|
| 17886 |
+
{
|
| 17887 |
+
"status": "passed",
|
| 17888 |
+
"test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2ModelTest::test_flash_attn_2_inference_equivalence_right_padding",
|
| 17889 |
+
"count": 1
|
| 17890 |
+
},
|
| 17891 |
+
{
|
| 17892 |
+
"status": "passed",
|
| 17893 |
+
"test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2ForImageClassificationModelTest::test_flash_attn_2_can_dispatch_composite_models",
|
| 17894 |
+
"count": 1
|
| 17895 |
+
},
|
| 17896 |
+
{
|
| 17897 |
+
"status": "passed",
|
| 17898 |
+
"test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2ForImageClassificationModelTest::test_flash_attn_2_fp32_ln",
|
| 17899 |
+
"count": 1
|
| 17900 |
+
},
|
| 17901 |
+
{
|
| 17902 |
+
"status": "passed",
|
| 17903 |
+
"test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2ForImageClassificationModelTest::test_flash_attn_2_from_config",
|
| 17904 |
+
"count": 1
|
| 17905 |
+
},
|
| 17906 |
+
{
|
| 17907 |
+
"status": "passed",
|
| 17908 |
+
"test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2ForImageClassificationModelTest::test_flash_attn_2_inference_equivalence",
|
| 17909 |
+
"count": 1
|
| 17910 |
+
},
|
| 17911 |
+
{
|
| 17912 |
+
"status": "passed",
|
| 17913 |
+
"test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2ForImageClassificationModelTest::test_flash_attn_2_inference_equivalence_right_padding",
|
| 17914 |
+
"count": 1
|
| 17915 |
+
},
|
| 17916 |
+
{
|
| 17917 |
+
"status": "skipped",
|
| 17918 |
+
"test": "[1] tests/test_modeling_common.py:3367: MetaClip2VisionModelTest doesn't have the attribute `_torch_compile_train_cls`.",
|
| 17919 |
+
"count": 1
|
| 17920 |
+
},
|
| 17921 |
+
{
|
| 17922 |
+
"status": "skipped",
|
| 17923 |
+
"test": "[2] tests/test_modeling_common.py:3294: This model is not a composite model!",
|
| 17924 |
+
"count": 2
|
| 17925 |
+
},
|
| 17926 |
+
{
|
| 17927 |
+
"status": "skipped",
|
| 17928 |
+
"test": "[4] tests/test_modeling_common.py:3300: test requires Flash Attention 3",
|
| 17929 |
+
"count": 4
|
| 17930 |
+
},
|
| 17931 |
+
{
|
| 17932 |
+
"status": "skipped",
|
| 17933 |
+
"test": "[4] tests/test_modeling_common.py:3469: test requires Flash Attention 3",
|
| 17934 |
+
"count": 4
|
| 17935 |
+
},
|
| 17936 |
+
{
|
| 17937 |
+
"status": "skipped",
|
| 17938 |
+
"test": "[4] tests/test_modeling_common.py:2910: test requires Flash Attention 3",
|
| 17939 |
+
"count": 4
|
| 17940 |
+
},
|
| 17941 |
+
{
|
| 17942 |
+
"status": "skipped",
|
| 17943 |
+
"test": "[4] tests/test_modeling_common.py:2918: test requires Flash Attention 3",
|
| 17944 |
+
"count": 4
|
| 17945 |
+
},
|
| 17946 |
+
{
|
| 17947 |
+
"status": "skipped",
|
| 17948 |
+
"test": "[4] tests/test_modeling_common.py:2876: test requires the kernels library",
|
| 17949 |
+
"count": 4
|
| 17950 |
+
},
|
| 17951 |
+
{
|
| 17952 |
+
"status": "skipped",
|
| 17953 |
+
"test": "[4] tests/test_modeling_common.py:2884: test requires MPS",
|
| 17954 |
+
"count": 4
|
| 17955 |
+
},
|
| 17956 |
+
{
|
| 17957 |
+
"status": "skipped",
|
| 17958 |
+
"test": "[1] tests/test_modeling_common.py:3367: MetaClip2TextModelTest doesn't have the attribute `_torch_compile_train_cls`.",
|
| 17959 |
+
"count": 1
|
| 17960 |
+
},
|
| 17961 |
+
{
|
| 17962 |
+
"status": "skipped",
|
| 17963 |
+
"test": "[1] tests/test_modeling_common.py:3367: MetaClip2ModelTest doesn't have the attribute `_torch_compile_train_cls`.",
|
| 17964 |
+
"count": 1
|
| 17965 |
+
},
|
| 17966 |
+
{
|
| 17967 |
+
"status": "skipped",
|
| 17968 |
+
"test": "[1] tests/test_modeling_common.py:3367: MetaClip2ForImageClassificationModelTest doesn't have the attribute `_torch_compile_train_cls`.",
|
| 17969 |
+
"count": 1
|
| 17970 |
+
}
|
| 17971 |
+
]
|
| 17972 |
+
},
|
| 17973 |
{
|
| 17974 |
"model": "mgp_str",
|
| 17975 |
"results": [
|