hf-transformers-bot commited on
Commit
6c27b81
·
verified ·
1 Parent(s): 9cc8c6e

Upload 2025-11-23/ci_results_run_models_gpu/collated_reports_multi-gpu_bdee088.json with huggingface_hub

Browse files
2025-11-23/ci_results_run_models_gpu/collated_reports_multi-gpu_bdee088.json CHANGED
@@ -3,9 +3,9 @@
3
  "machine_type": "multi-gpu",
4
  "commit_hash": "bdee088",
5
  "total_status_count": {
6
- "passed": 1480,
7
- "failed": 289,
8
- "skipped": 5338,
9
  "error": 0,
10
  "null": 0
11
  },
@@ -17202,6 +17202,594 @@
17202
  }
17203
  ]
17204
  },
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
17205
  {
17206
  "model": "mgp_str",
17207
  "results": [
 
3
  "machine_type": "multi-gpu",
4
  "commit_hash": "bdee088",
5
  "total_status_count": {
6
+ "passed": 1516,
7
+ "failed": 297,
8
+ "skipped": 5440,
9
  "error": 0,
10
  "null": 0
11
  },
 
17202
  }
17203
  ]
17204
  },
17205
+ {
17206
+ "model": "markuplm",
17207
+ "results": [
17208
+ {
17209
+ "status": "passed",
17210
+ "test": "tests/models/markuplm/test_modeling_markuplm.py::MarkupLMModelTest::test_flash_attn_2_fp32_ln",
17211
+ "count": 1
17212
+ },
17213
+ {
17214
+ "status": "passed",
17215
+ "test": "tests/models/markuplm/test_modeling_markuplm.py::MarkupLMModelTest::test_flash_attn_2_from_config",
17216
+ "count": 1
17217
+ },
17218
+ {
17219
+ "status": "skipped",
17220
+ "test": "[1] tests/test_modeling_common.py:3354: MarkupLMModelTest doesn't have the attribute `_torch_compile_train_cls`.",
17221
+ "count": 1
17222
+ },
17223
+ {
17224
+ "status": "skipped",
17225
+ "test": "[1] tests/test_modeling_common.py:3286: This model is not a composite model!",
17226
+ "count": 1
17227
+ },
17228
+ {
17229
+ "status": "skipped",
17230
+ "test": "[1] tests/test_modeling_common.py:2886: Model architecture does not support flash_attention_2, or setting its attention dynamically",
17231
+ "count": 1
17232
+ },
17233
+ {
17234
+ "status": "skipped",
17235
+ "test": "[1] tests/test_modeling_common.py:2894: Model architecture does not support flash_attention_2, or setting its attention dynamically",
17236
+ "count": 1
17237
+ },
17238
+ {
17239
+ "status": "skipped",
17240
+ "test": "[1] tests/test_modeling_common.py:3292: test requires Flash Attention 3",
17241
+ "count": 1
17242
+ },
17243
+ {
17244
+ "status": "skipped",
17245
+ "test": "[1] tests/test_modeling_common.py:3448: test requires Flash Attention 3",
17246
+ "count": 1
17247
+ },
17248
+ {
17249
+ "status": "skipped",
17250
+ "test": "[1] tests/test_modeling_common.py:2902: test requires Flash Attention 3",
17251
+ "count": 1
17252
+ },
17253
+ {
17254
+ "status": "skipped",
17255
+ "test": "[1] tests/test_modeling_common.py:2910: test requires Flash Attention 3",
17256
+ "count": 1
17257
+ },
17258
+ {
17259
+ "status": "skipped",
17260
+ "test": "[1] tests/test_modeling_common.py:2868: test requires the kernels library",
17261
+ "count": 1
17262
+ },
17263
+ {
17264
+ "status": "skipped",
17265
+ "test": "[1] tests/test_modeling_common.py:2876: test requires MPS",
17266
+ "count": 1
17267
+ }
17268
+ ]
17269
+ },
17270
+ {
17271
+ "model": "mask2former",
17272
+ "results": [
17273
+ {
17274
+ "status": "passed",
17275
+ "test": "tests/models/mask2former/test_modeling_mask2former.py::Mask2FormerModelTest::test_flash_attn_2_fp32_ln",
17276
+ "count": 1
17277
+ },
17278
+ {
17279
+ "status": "passed",
17280
+ "test": "tests/models/mask2former/test_modeling_mask2former.py::Mask2FormerModelTest::test_flash_attn_2_from_config",
17281
+ "count": 1
17282
+ },
17283
+ {
17284
+ "status": "skipped",
17285
+ "test": "[1] tests/test_modeling_common.py:3354: Mask2FormerModelTest doesn't have the attribute `_torch_compile_train_cls`.",
17286
+ "count": 1
17287
+ },
17288
+ {
17289
+ "status": "skipped",
17290
+ "test": "[1] tests/test_modeling_common.py:3286: This model is not a composite model!",
17291
+ "count": 1
17292
+ },
17293
+ {
17294
+ "status": "skipped",
17295
+ "test": "[1] tests/test_modeling_common.py:3292: test requires Flash Attention 3",
17296
+ "count": 1
17297
+ },
17298
+ {
17299
+ "status": "skipped",
17300
+ "test": "[1] tests/test_modeling_common.py:3448: test requires Flash Attention 3",
17301
+ "count": 1
17302
+ },
17303
+ {
17304
+ "status": "skipped",
17305
+ "test": "[1] tests/test_modeling_common.py:2902: test requires Flash Attention 3",
17306
+ "count": 1
17307
+ },
17308
+ {
17309
+ "status": "skipped",
17310
+ "test": "[1] tests/test_modeling_common.py:2910: test requires Flash Attention 3",
17311
+ "count": 1
17312
+ },
17313
+ {
17314
+ "status": "skipped",
17315
+ "test": "[1] tests/test_modeling_common.py:2868: test requires the kernels library",
17316
+ "count": 1
17317
+ },
17318
+ {
17319
+ "status": "skipped",
17320
+ "test": "[1] tests/test_modeling_common.py:2876: test requires MPS",
17321
+ "count": 1
17322
+ },
17323
+ {
17324
+ "status": "failed",
17325
+ "test": "tests/models/mask2former/test_modeling_mask2former.py::Mask2FormerModelTest::test_flash_attn_2_inference_equivalence - TypeError: unsupported operand type(s) for //: 'int' and 'list'",
17326
+ "count": 1
17327
+ },
17328
+ {
17329
+ "status": "failed",
17330
+ "test": "tests/models/mask2former/test_modeling_mask2former.py::Mask2FormerModelTest::test_flash_attn_2_inference_equivalence_right_padding - TypeError: unsupported operand type(s) for //: 'int' and 'list'",
17331
+ "count": 1
17332
+ }
17333
+ ]
17334
+ },
17335
+ {
17336
+ "model": "maskformer",
17337
+ "results": [
17338
+ {
17339
+ "status": "passed",
17340
+ "test": "tests/models/maskformer/test_modeling_maskformer.py::MaskFormerModelTest::test_flash_attn_2_fp32_ln",
17341
+ "count": 1
17342
+ },
17343
+ {
17344
+ "status": "passed",
17345
+ "test": "tests/models/maskformer/test_modeling_maskformer.py::MaskFormerModelTest::test_flash_attn_2_from_config",
17346
+ "count": 1
17347
+ },
17348
+ {
17349
+ "status": "passed",
17350
+ "test": "tests/models/maskformer/test_modeling_maskformer_swin.py::MaskFormerSwinModelTest::test_flash_attn_2_fp32_ln",
17351
+ "count": 1
17352
+ },
17353
+ {
17354
+ "status": "passed",
17355
+ "test": "tests/models/maskformer/test_modeling_maskformer_swin.py::MaskFormerSwinModelTest::test_flash_attn_2_from_config",
17356
+ "count": 1
17357
+ },
17358
+ {
17359
+ "status": "skipped",
17360
+ "test": "[1] tests/test_modeling_common.py:3354: MaskFormerModelTest doesn't have the attribute `_torch_compile_train_cls`.",
17361
+ "count": 1
17362
+ },
17363
+ {
17364
+ "status": "skipped",
17365
+ "test": "[2] tests/test_modeling_common.py:3286: This model is not a composite model!",
17366
+ "count": 2
17367
+ },
17368
+ {
17369
+ "status": "skipped",
17370
+ "test": "[2] tests/test_modeling_common.py:3292: test requires Flash Attention 3",
17371
+ "count": 2
17372
+ },
17373
+ {
17374
+ "status": "skipped",
17375
+ "test": "[2] tests/test_modeling_common.py:3448: test requires Flash Attention 3",
17376
+ "count": 2
17377
+ },
17378
+ {
17379
+ "status": "skipped",
17380
+ "test": "[2] tests/test_modeling_common.py:2902: test requires Flash Attention 3",
17381
+ "count": 2
17382
+ },
17383
+ {
17384
+ "status": "skipped",
17385
+ "test": "[2] tests/test_modeling_common.py:2910: test requires Flash Attention 3",
17386
+ "count": 2
17387
+ },
17388
+ {
17389
+ "status": "skipped",
17390
+ "test": "[2] tests/test_modeling_common.py:2868: test requires the kernels library",
17391
+ "count": 2
17392
+ },
17393
+ {
17394
+ "status": "skipped",
17395
+ "test": "[2] tests/test_modeling_common.py:2876: test requires MPS",
17396
+ "count": 2
17397
+ },
17398
+ {
17399
+ "status": "skipped",
17400
+ "test": "[1] tests/test_modeling_common.py:3354: MaskFormerSwinModelTest doesn't have the attribute `_torch_compile_train_cls`.",
17401
+ "count": 1
17402
+ },
17403
+ {
17404
+ "status": "failed",
17405
+ "test": "tests/models/maskformer/test_modeling_maskformer.py::MaskFormerModelTest::test_flash_attn_2_inference_equivalence - TypeError: unsupported operand type(s) for //: 'int' and 'list'",
17406
+ "count": 1
17407
+ },
17408
+ {
17409
+ "status": "failed",
17410
+ "test": "tests/models/maskformer/test_modeling_maskformer.py::MaskFormerModelTest::test_flash_attn_2_inference_equivalence_right_padding - TypeError: unsupported operand type(s) for //: 'int' and 'list'",
17411
+ "count": 1
17412
+ },
17413
+ {
17414
+ "status": "failed",
17415
+ "test": "tests/models/maskformer/test_modeling_maskformer_swin.py::MaskFormerSwinModelTest::test_flash_attn_2_inference_equivalence - TypeError: unsupported operand type(s) for //: 'int' and 'list'",
17416
+ "count": 1
17417
+ },
17418
+ {
17419
+ "status": "failed",
17420
+ "test": "tests/models/maskformer/test_modeling_maskformer_swin.py::MaskFormerSwinModelTest::test_flash_attn_2_inference_equivalence_right_padding - TypeError: unsupported operand type(s) for //: 'int' and 'list'",
17421
+ "count": 1
17422
+ }
17423
+ ]
17424
+ },
17425
+ {
17426
+ "model": "mbart50",
17427
+ "results": []
17428
+ },
17429
+ {
17430
+ "model": "mbart",
17431
+ "results": [
17432
+ {
17433
+ "status": "passed",
17434
+ "test": "tests/models/mbart/test_modeling_mbart.py::MBartModelTest::test_eager_matches_fa2_generate",
17435
+ "count": 1
17436
+ },
17437
+ {
17438
+ "status": "passed",
17439
+ "test": "tests/models/mbart/test_modeling_mbart.py::MBartModelTest::test_flash_attn_2_from_config",
17440
+ "count": 1
17441
+ },
17442
+ {
17443
+ "status": "passed",
17444
+ "test": "tests/models/mbart/test_modeling_mbart.py::MBartModelTest::test_flash_attn_2_inference_equivalence",
17445
+ "count": 1
17446
+ },
17447
+ {
17448
+ "status": "passed",
17449
+ "test": "tests/models/mbart/test_modeling_mbart.py::MBartModelTest::test_flash_attn_2_inference_equivalence_right_padding",
17450
+ "count": 1
17451
+ },
17452
+ {
17453
+ "status": "passed",
17454
+ "test": "tests/models/mbart/test_modeling_mbart.py::MBartStandaloneDecoderModelTest::test_eager_matches_fa2_generate",
17455
+ "count": 1
17456
+ },
17457
+ {
17458
+ "status": "passed",
17459
+ "test": "tests/models/mbart/test_modeling_mbart.py::MBartStandaloneDecoderModelTest::test_flash_attn_2_from_config",
17460
+ "count": 1
17461
+ },
17462
+ {
17463
+ "status": "passed",
17464
+ "test": "tests/models/mbart/test_modeling_mbart.py::MBartStandaloneDecoderModelTest::test_flash_attn_2_inference_equivalence",
17465
+ "count": 1
17466
+ },
17467
+ {
17468
+ "status": "passed",
17469
+ "test": "tests/models/mbart/test_modeling_mbart.py::MBartStandaloneDecoderModelTest::test_flash_attn_2_inference_equivalence_right_padding",
17470
+ "count": 1
17471
+ },
17472
+ {
17473
+ "status": "skipped",
17474
+ "test": "[2] tests/generation/test_utils.py:1857: test requires Flash Attention 3",
17475
+ "count": 2
17476
+ },
17477
+ {
17478
+ "status": "skipped",
17479
+ "test": "[1] tests/generation/test_utils.py:1865: Model is an encoder-decoder",
17480
+ "count": 1
17481
+ },
17482
+ {
17483
+ "status": "skipped",
17484
+ "test": "[1] tests/generation/test_utils.py:2067: MBartForConditionalGeneration does not support new attention mask API",
17485
+ "count": 1
17486
+ },
17487
+ {
17488
+ "status": "skipped",
17489
+ "test": "[1] tests/generation/test_utils.py:2074: MBartForConditionalGeneration does not support new attention mask API",
17490
+ "count": 1
17491
+ },
17492
+ {
17493
+ "status": "skipped",
17494
+ "test": "[2] tests/generation/test_utils.py:2083: test requires Flash Attention 3",
17495
+ "count": 2
17496
+ },
17497
+ {
17498
+ "status": "skipped",
17499
+ "test": "[2] tests/generation/test_utils.py:2090: test requires Flash Attention 3",
17500
+ "count": 2
17501
+ },
17502
+ {
17503
+ "status": "skipped",
17504
+ "test": "[1] tests/test_modeling_common.py:3354: MBartModelTest doesn't have the attribute `_torch_compile_train_cls`.",
17505
+ "count": 1
17506
+ },
17507
+ {
17508
+ "status": "skipped",
17509
+ "test": "[2] tests/test_modeling_common.py:3286: This model is not a composite model!",
17510
+ "count": 2
17511
+ },
17512
+ {
17513
+ "status": "skipped",
17514
+ "test": "[2] tests/test_modeling_common.py:3292: test requires Flash Attention 3",
17515
+ "count": 2
17516
+ },
17517
+ {
17518
+ "status": "skipped",
17519
+ "test": "[2] tests/test_modeling_common.py:3448: test requires Flash Attention 3",
17520
+ "count": 2
17521
+ },
17522
+ {
17523
+ "status": "skipped",
17524
+ "test": "[2] tests/test_modeling_common.py:2902: test requires Flash Attention 3",
17525
+ "count": 2
17526
+ },
17527
+ {
17528
+ "status": "skipped",
17529
+ "test": "[2] tests/test_modeling_common.py:2910: test requires Flash Attention 3",
17530
+ "count": 2
17531
+ },
17532
+ {
17533
+ "status": "skipped",
17534
+ "test": "[2] tests/test_modeling_common.py:2868: test requires the kernels library",
17535
+ "count": 2
17536
+ },
17537
+ {
17538
+ "status": "skipped",
17539
+ "test": "[2] tests/test_modeling_common.py:2876: test requires MPS",
17540
+ "count": 2
17541
+ },
17542
+ {
17543
+ "status": "skipped",
17544
+ "test": "[1] tests/generation/test_utils.py:1865: Model does not support position_ids",
17545
+ "count": 1
17546
+ },
17547
+ {
17548
+ "status": "skipped",
17549
+ "test": "[1] tests/generation/test_utils.py:2067: MBartForCausalLM does not support new attention mask API",
17550
+ "count": 1
17551
+ },
17552
+ {
17553
+ "status": "skipped",
17554
+ "test": "[1] tests/generation/test_utils.py:2074: MBartForCausalLM does not support new attention mask API",
17555
+ "count": 1
17556
+ },
17557
+ {
17558
+ "status": "skipped",
17559
+ "test": "[1] tests/test_modeling_common.py:3354: MBartStandaloneDecoderModelTest doesn't have the attribute `_torch_compile_train_cls`.",
17560
+ "count": 1
17561
+ },
17562
+ {
17563
+ "status": "failed",
17564
+ "test": "tests/models/mbart/test_modeling_mbart.py::MBartModelTest::test_flash_attn_2_fp32_ln - NotImplementedError: Cannot copy out of meta tensor; no data! Please use torch.nn.Module.to_empty() instead of torch.nn.Module.to() when moving module from meta to a different device.",
17565
+ "count": 1
17566
+ },
17567
+ {
17568
+ "status": "failed",
17569
+ "test": "tests/models/mbart/test_modeling_mbart.py::MBartStandaloneDecoderModelTest::test_flash_attn_2_fp32_ln - NotImplementedError: Cannot copy out of meta tensor; no data! Please use torch.nn.Module.to_empty() instead of torch.nn.Module.to() when moving module from meta to a different device.",
17570
+ "count": 1
17571
+ }
17572
+ ]
17573
+ },
17574
+ {
17575
+ "model": "megatron_bert",
17576
+ "results": [
17577
+ {
17578
+ "status": "passed",
17579
+ "test": "tests/models/megatron_bert/test_modeling_megatron_bert.py::MegatronBertModelTest::test_flash_attn_2_fp32_ln",
17580
+ "count": 1
17581
+ },
17582
+ {
17583
+ "status": "passed",
17584
+ "test": "tests/models/megatron_bert/test_modeling_megatron_bert.py::MegatronBertModelTest::test_flash_attn_2_from_config",
17585
+ "count": 1
17586
+ },
17587
+ {
17588
+ "status": "skipped",
17589
+ "test": "[1] tests/test_modeling_common.py:3354: MegatronBertModelTest doesn't have the attribute `_torch_compile_train_cls`.",
17590
+ "count": 1
17591
+ },
17592
+ {
17593
+ "status": "skipped",
17594
+ "test": "[1] tests/test_modeling_common.py:3286: This model is not a composite model!",
17595
+ "count": 1
17596
+ },
17597
+ {
17598
+ "status": "skipped",
17599
+ "test": "[1] tests/test_modeling_common.py:2886: Model architecture does not support flash_attention_2, or setting its attention dynamically",
17600
+ "count": 1
17601
+ },
17602
+ {
17603
+ "status": "skipped",
17604
+ "test": "[1] tests/test_modeling_common.py:2894: Model architecture does not support flash_attention_2, or setting its attention dynamically",
17605
+ "count": 1
17606
+ },
17607
+ {
17608
+ "status": "skipped",
17609
+ "test": "[1] tests/test_modeling_common.py:3292: test requires Flash Attention 3",
17610
+ "count": 1
17611
+ },
17612
+ {
17613
+ "status": "skipped",
17614
+ "test": "[1] tests/test_modeling_common.py:3448: test requires Flash Attention 3",
17615
+ "count": 1
17616
+ },
17617
+ {
17618
+ "status": "skipped",
17619
+ "test": "[1] tests/test_modeling_common.py:2902: test requires Flash Attention 3",
17620
+ "count": 1
17621
+ },
17622
+ {
17623
+ "status": "skipped",
17624
+ "test": "[1] tests/test_modeling_common.py:2910: test requires Flash Attention 3",
17625
+ "count": 1
17626
+ },
17627
+ {
17628
+ "status": "skipped",
17629
+ "test": "[1] tests/test_modeling_common.py:2868: test requires the kernels library",
17630
+ "count": 1
17631
+ },
17632
+ {
17633
+ "status": "skipped",
17634
+ "test": "[1] tests/test_modeling_common.py:2876: test requires MPS",
17635
+ "count": 1
17636
+ }
17637
+ ]
17638
+ },
17639
+ {
17640
+ "model": "megatron_gpt2",
17641
+ "results": []
17642
+ },
17643
+ {
17644
+ "model": "metaclip_2",
17645
+ "results": [
17646
+ {
17647
+ "status": "passed",
17648
+ "test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2VisionModelTest::test_flash_attn_2_fp32_ln",
17649
+ "count": 1
17650
+ },
17651
+ {
17652
+ "status": "passed",
17653
+ "test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2VisionModelTest::test_flash_attn_2_from_config",
17654
+ "count": 1
17655
+ },
17656
+ {
17657
+ "status": "passed",
17658
+ "test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2VisionModelTest::test_flash_attn_2_inference_equivalence",
17659
+ "count": 1
17660
+ },
17661
+ {
17662
+ "status": "passed",
17663
+ "test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2VisionModelTest::test_flash_attn_2_inference_equivalence_right_padding",
17664
+ "count": 1
17665
+ },
17666
+ {
17667
+ "status": "passed",
17668
+ "test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2TextModelTest::test_flash_attn_2_fp32_ln",
17669
+ "count": 1
17670
+ },
17671
+ {
17672
+ "status": "passed",
17673
+ "test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2TextModelTest::test_flash_attn_2_from_config",
17674
+ "count": 1
17675
+ },
17676
+ {
17677
+ "status": "passed",
17678
+ "test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2TextModelTest::test_flash_attn_2_inference_equivalence",
17679
+ "count": 1
17680
+ },
17681
+ {
17682
+ "status": "passed",
17683
+ "test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2TextModelTest::test_flash_attn_2_inference_equivalence_right_padding",
17684
+ "count": 1
17685
+ },
17686
+ {
17687
+ "status": "passed",
17688
+ "test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2ModelTest::test_flash_attn_2_can_dispatch_composite_models",
17689
+ "count": 1
17690
+ },
17691
+ {
17692
+ "status": "passed",
17693
+ "test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2ModelTest::test_flash_attn_2_fp32_ln",
17694
+ "count": 1
17695
+ },
17696
+ {
17697
+ "status": "passed",
17698
+ "test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2ModelTest::test_flash_attn_2_from_config",
17699
+ "count": 1
17700
+ },
17701
+ {
17702
+ "status": "passed",
17703
+ "test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2ModelTest::test_flash_attn_2_inference_equivalence",
17704
+ "count": 1
17705
+ },
17706
+ {
17707
+ "status": "passed",
17708
+ "test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2ModelTest::test_flash_attn_2_inference_equivalence_right_padding",
17709
+ "count": 1
17710
+ },
17711
+ {
17712
+ "status": "passed",
17713
+ "test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2ForImageClassificationModelTest::test_flash_attn_2_can_dispatch_composite_models",
17714
+ "count": 1
17715
+ },
17716
+ {
17717
+ "status": "passed",
17718
+ "test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2ForImageClassificationModelTest::test_flash_attn_2_fp32_ln",
17719
+ "count": 1
17720
+ },
17721
+ {
17722
+ "status": "passed",
17723
+ "test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2ForImageClassificationModelTest::test_flash_attn_2_from_config",
17724
+ "count": 1
17725
+ },
17726
+ {
17727
+ "status": "passed",
17728
+ "test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2ForImageClassificationModelTest::test_flash_attn_2_inference_equivalence",
17729
+ "count": 1
17730
+ },
17731
+ {
17732
+ "status": "passed",
17733
+ "test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2ForImageClassificationModelTest::test_flash_attn_2_inference_equivalence_right_padding",
17734
+ "count": 1
17735
+ },
17736
+ {
17737
+ "status": "skipped",
17738
+ "test": "[1] tests/test_modeling_common.py:3354: MetaClip2VisionModelTest doesn't have the attribute `_torch_compile_train_cls`.",
17739
+ "count": 1
17740
+ },
17741
+ {
17742
+ "status": "skipped",
17743
+ "test": "[2] tests/test_modeling_common.py:3286: This model is not a composite model!",
17744
+ "count": 2
17745
+ },
17746
+ {
17747
+ "status": "skipped",
17748
+ "test": "[4] tests/test_modeling_common.py:3292: test requires Flash Attention 3",
17749
+ "count": 4
17750
+ },
17751
+ {
17752
+ "status": "skipped",
17753
+ "test": "[4] tests/test_modeling_common.py:3448: test requires Flash Attention 3",
17754
+ "count": 4
17755
+ },
17756
+ {
17757
+ "status": "skipped",
17758
+ "test": "[4] tests/test_modeling_common.py:2902: test requires Flash Attention 3",
17759
+ "count": 4
17760
+ },
17761
+ {
17762
+ "status": "skipped",
17763
+ "test": "[4] tests/test_modeling_common.py:2910: test requires Flash Attention 3",
17764
+ "count": 4
17765
+ },
17766
+ {
17767
+ "status": "skipped",
17768
+ "test": "[4] tests/test_modeling_common.py:2868: test requires the kernels library",
17769
+ "count": 4
17770
+ },
17771
+ {
17772
+ "status": "skipped",
17773
+ "test": "[4] tests/test_modeling_common.py:2876: test requires MPS",
17774
+ "count": 4
17775
+ },
17776
+ {
17777
+ "status": "skipped",
17778
+ "test": "[1] tests/test_modeling_common.py:3354: MetaClip2TextModelTest doesn't have the attribute `_torch_compile_train_cls`.",
17779
+ "count": 1
17780
+ },
17781
+ {
17782
+ "status": "skipped",
17783
+ "test": "[1] tests/test_modeling_common.py:3354: MetaClip2ModelTest doesn't have the attribute `_torch_compile_train_cls`.",
17784
+ "count": 1
17785
+ },
17786
+ {
17787
+ "status": "skipped",
17788
+ "test": "[1] tests/test_modeling_common.py:3354: MetaClip2ForImageClassificationModelTest doesn't have the attribute `_torch_compile_train_cls`.",
17789
+ "count": 1
17790
+ }
17791
+ ]
17792
+ },
17793
  {
17794
  "model": "mgp_str",
17795
  "results": [