hf-transformers-bot commited on
Commit
bf19dbf
·
verified ·
1 Parent(s): 3078a84

Upload 2025-11-01/ci_results_run_models_gpu/collated_reports_multi-gpu_a0bf5a8.json with huggingface_hub

Browse files
2025-11-01/ci_results_run_models_gpu/collated_reports_multi-gpu_a0bf5a8.json CHANGED
@@ -3,9 +3,9 @@
3
  "machine_type": "multi-gpu",
4
  "commit_hash": "a0bf5a8",
5
  "total_status_count": {
6
- "passed": 1635,
7
- "failed": 139,
8
- "skipped": 5360,
9
  "error": 0,
10
  "null": 0
11
  },
@@ -17144,6 +17144,375 @@
17144
  }
17145
  ]
17146
  },
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
17147
  {
17148
  "model": "megatron_gpt2",
17149
  "results": []
 
3
  "machine_type": "multi-gpu",
4
  "commit_hash": "a0bf5a8",
5
  "total_status_count": {
6
+ "passed": 1653,
7
+ "failed": 145,
8
+ "skipped": 5422,
9
  "error": 0,
10
  "null": 0
11
  },
 
17144
  }
17145
  ]
17146
  },
17147
+ {
17148
+ "model": "mask2former",
17149
+ "results": [
17150
+ {
17151
+ "status": "passed",
17152
+ "test": "tests/models/mask2former/test_modeling_mask2former.py::Mask2FormerModelTest::test_flash_attn_2_fp32_ln",
17153
+ "count": 1
17154
+ },
17155
+ {
17156
+ "status": "passed",
17157
+ "test": "tests/models/mask2former/test_modeling_mask2former.py::Mask2FormerModelTest::test_flash_attn_2_from_config",
17158
+ "count": 1
17159
+ },
17160
+ {
17161
+ "status": "skipped",
17162
+ "test": "[1] tests/test_modeling_common.py:3236: Mask2FormerModelTest doesn't have the attribute `_torch_compile_train_cls`.",
17163
+ "count": 1
17164
+ },
17165
+ {
17166
+ "status": "skipped",
17167
+ "test": "[1] tests/test_modeling_common.py:3168: This model is not a composite model!",
17168
+ "count": 1
17169
+ },
17170
+ {
17171
+ "status": "skipped",
17172
+ "test": "[1] tests/test_modeling_common.py:3174: test requires Flash Attention 3",
17173
+ "count": 1
17174
+ },
17175
+ {
17176
+ "status": "skipped",
17177
+ "test": "[1] tests/test_modeling_common.py:3330: test requires Flash Attention 3",
17178
+ "count": 1
17179
+ },
17180
+ {
17181
+ "status": "skipped",
17182
+ "test": "[1] tests/test_modeling_common.py:2784: test requires Flash Attention 3",
17183
+ "count": 1
17184
+ },
17185
+ {
17186
+ "status": "skipped",
17187
+ "test": "[1] tests/test_modeling_common.py:2792: test requires Flash Attention 3",
17188
+ "count": 1
17189
+ },
17190
+ {
17191
+ "status": "skipped",
17192
+ "test": "[1] tests/test_modeling_common.py:2750: test requires the kernels library",
17193
+ "count": 1
17194
+ },
17195
+ {
17196
+ "status": "skipped",
17197
+ "test": "[1] tests/test_modeling_common.py:2758: test requires MPS",
17198
+ "count": 1
17199
+ },
17200
+ {
17201
+ "status": "failed",
17202
+ "test": "tests/models/mask2former/test_modeling_mask2former.py::Mask2FormerModelTest::test_flash_attn_2_inference_equivalence - TypeError: unsupported operand type(s) for //: 'int' and 'list'",
17203
+ "count": 1
17204
+ },
17205
+ {
17206
+ "status": "failed",
17207
+ "test": "tests/models/mask2former/test_modeling_mask2former.py::Mask2FormerModelTest::test_flash_attn_2_inference_equivalence_right_padding - TypeError: unsupported operand type(s) for //: 'int' and 'list'",
17208
+ "count": 1
17209
+ }
17210
+ ]
17211
+ },
17212
+ {
17213
+ "model": "maskformer",
17214
+ "results": [
17215
+ {
17216
+ "status": "passed",
17217
+ "test": "tests/models/maskformer/test_modeling_maskformer.py::MaskFormerModelTest::test_flash_attn_2_fp32_ln",
17218
+ "count": 1
17219
+ },
17220
+ {
17221
+ "status": "passed",
17222
+ "test": "tests/models/maskformer/test_modeling_maskformer.py::MaskFormerModelTest::test_flash_attn_2_from_config",
17223
+ "count": 1
17224
+ },
17225
+ {
17226
+ "status": "passed",
17227
+ "test": "tests/models/maskformer/test_modeling_maskformer_swin.py::MaskFormerSwinModelTest::test_flash_attn_2_fp32_ln",
17228
+ "count": 1
17229
+ },
17230
+ {
17231
+ "status": "passed",
17232
+ "test": "tests/models/maskformer/test_modeling_maskformer_swin.py::MaskFormerSwinModelTest::test_flash_attn_2_from_config",
17233
+ "count": 1
17234
+ },
17235
+ {
17236
+ "status": "skipped",
17237
+ "test": "[1] tests/test_modeling_common.py:3236: MaskFormerModelTest doesn't have the attribute `_torch_compile_train_cls`.",
17238
+ "count": 1
17239
+ },
17240
+ {
17241
+ "status": "skipped",
17242
+ "test": "[2] tests/test_modeling_common.py:3168: This model is not a composite model!",
17243
+ "count": 2
17244
+ },
17245
+ {
17246
+ "status": "skipped",
17247
+ "test": "[2] tests/test_modeling_common.py:3174: test requires Flash Attention 3",
17248
+ "count": 2
17249
+ },
17250
+ {
17251
+ "status": "skipped",
17252
+ "test": "[2] tests/test_modeling_common.py:3330: test requires Flash Attention 3",
17253
+ "count": 2
17254
+ },
17255
+ {
17256
+ "status": "skipped",
17257
+ "test": "[2] tests/test_modeling_common.py:2784: test requires Flash Attention 3",
17258
+ "count": 2
17259
+ },
17260
+ {
17261
+ "status": "skipped",
17262
+ "test": "[2] tests/test_modeling_common.py:2792: test requires Flash Attention 3",
17263
+ "count": 2
17264
+ },
17265
+ {
17266
+ "status": "skipped",
17267
+ "test": "[2] tests/test_modeling_common.py:2750: test requires the kernels library",
17268
+ "count": 2
17269
+ },
17270
+ {
17271
+ "status": "skipped",
17272
+ "test": "[2] tests/test_modeling_common.py:2758: test requires MPS",
17273
+ "count": 2
17274
+ },
17275
+ {
17276
+ "status": "skipped",
17277
+ "test": "[1] tests/test_modeling_common.py:3236: MaskFormerSwinModelTest doesn't have the attribute `_torch_compile_train_cls`.",
17278
+ "count": 1
17279
+ },
17280
+ {
17281
+ "status": "failed",
17282
+ "test": "tests/models/maskformer/test_modeling_maskformer.py::MaskFormerModelTest::test_flash_attn_2_inference_equivalence - TypeError: unsupported operand type(s) for //: 'int' and 'list'",
17283
+ "count": 1
17284
+ },
17285
+ {
17286
+ "status": "failed",
17287
+ "test": "tests/models/maskformer/test_modeling_maskformer.py::MaskFormerModelTest::test_flash_attn_2_inference_equivalence_right_padding - TypeError: unsupported operand type(s) for //: 'int' and 'list'",
17288
+ "count": 1
17289
+ },
17290
+ {
17291
+ "status": "failed",
17292
+ "test": "tests/models/maskformer/test_modeling_maskformer_swin.py::MaskFormerSwinModelTest::test_flash_attn_2_inference_equivalence - TypeError: unsupported operand type(s) for //: 'int' and 'list'",
17293
+ "count": 1
17294
+ },
17295
+ {
17296
+ "status": "failed",
17297
+ "test": "tests/models/maskformer/test_modeling_maskformer_swin.py::MaskFormerSwinModelTest::test_flash_attn_2_inference_equivalence_right_padding - TypeError: unsupported operand type(s) for //: 'int' and 'list'",
17298
+ "count": 1
17299
+ }
17300
+ ]
17301
+ },
17302
+ {
17303
+ "model": "mbart50",
17304
+ "results": []
17305
+ },
17306
+ {
17307
+ "model": "mbart",
17308
+ "results": [
17309
+ {
17310
+ "status": "passed",
17311
+ "test": "tests/models/mbart/test_modeling_mbart.py::MBartModelTest::test_eager_matches_fa2_generate",
17312
+ "count": 1
17313
+ },
17314
+ {
17315
+ "status": "passed",
17316
+ "test": "tests/models/mbart/test_modeling_mbart.py::MBartModelTest::test_flash_attn_2_fp32_ln",
17317
+ "count": 1
17318
+ },
17319
+ {
17320
+ "status": "passed",
17321
+ "test": "tests/models/mbart/test_modeling_mbart.py::MBartModelTest::test_flash_attn_2_from_config",
17322
+ "count": 1
17323
+ },
17324
+ {
17325
+ "status": "passed",
17326
+ "test": "tests/models/mbart/test_modeling_mbart.py::MBartModelTest::test_flash_attn_2_inference_equivalence",
17327
+ "count": 1
17328
+ },
17329
+ {
17330
+ "status": "passed",
17331
+ "test": "tests/models/mbart/test_modeling_mbart.py::MBartModelTest::test_flash_attn_2_inference_equivalence_right_padding",
17332
+ "count": 1
17333
+ },
17334
+ {
17335
+ "status": "passed",
17336
+ "test": "tests/models/mbart/test_modeling_mbart.py::MBartStandaloneDecoderModelTest::test_eager_matches_fa2_generate",
17337
+ "count": 1
17338
+ },
17339
+ {
17340
+ "status": "passed",
17341
+ "test": "tests/models/mbart/test_modeling_mbart.py::MBartStandaloneDecoderModelTest::test_flash_attn_2_fp32_ln",
17342
+ "count": 1
17343
+ },
17344
+ {
17345
+ "status": "passed",
17346
+ "test": "tests/models/mbart/test_modeling_mbart.py::MBartStandaloneDecoderModelTest::test_flash_attn_2_from_config",
17347
+ "count": 1
17348
+ },
17349
+ {
17350
+ "status": "passed",
17351
+ "test": "tests/models/mbart/test_modeling_mbart.py::MBartStandaloneDecoderModelTest::test_flash_attn_2_inference_equivalence",
17352
+ "count": 1
17353
+ },
17354
+ {
17355
+ "status": "passed",
17356
+ "test": "tests/models/mbart/test_modeling_mbart.py::MBartStandaloneDecoderModelTest::test_flash_attn_2_inference_equivalence_right_padding",
17357
+ "count": 1
17358
+ },
17359
+ {
17360
+ "status": "skipped",
17361
+ "test": "[2] tests/generation/test_utils.py:1857: test requires Flash Attention 3",
17362
+ "count": 2
17363
+ },
17364
+ {
17365
+ "status": "skipped",
17366
+ "test": "[1] tests/generation/test_utils.py:1865: Model is an encoder-decoder",
17367
+ "count": 1
17368
+ },
17369
+ {
17370
+ "status": "skipped",
17371
+ "test": "[1] tests/generation/test_utils.py:2067: MBartForConditionalGeneration does not support new attention mask API",
17372
+ "count": 1
17373
+ },
17374
+ {
17375
+ "status": "skipped",
17376
+ "test": "[1] tests/generation/test_utils.py:2074: MBartForConditionalGeneration does not support new attention mask API",
17377
+ "count": 1
17378
+ },
17379
+ {
17380
+ "status": "skipped",
17381
+ "test": "[2] tests/generation/test_utils.py:2083: test requires Flash Attention 3",
17382
+ "count": 2
17383
+ },
17384
+ {
17385
+ "status": "skipped",
17386
+ "test": "[2] tests/generation/test_utils.py:2090: test requires Flash Attention 3",
17387
+ "count": 2
17388
+ },
17389
+ {
17390
+ "status": "skipped",
17391
+ "test": "[1] tests/test_modeling_common.py:3236: MBartModelTest doesn't have the attribute `_torch_compile_train_cls`.",
17392
+ "count": 1
17393
+ },
17394
+ {
17395
+ "status": "skipped",
17396
+ "test": "[2] tests/test_modeling_common.py:3168: This model is not a composite model!",
17397
+ "count": 2
17398
+ },
17399
+ {
17400
+ "status": "skipped",
17401
+ "test": "[2] tests/test_modeling_common.py:3174: test requires Flash Attention 3",
17402
+ "count": 2
17403
+ },
17404
+ {
17405
+ "status": "skipped",
17406
+ "test": "[2] tests/test_modeling_common.py:3330: test requires Flash Attention 3",
17407
+ "count": 2
17408
+ },
17409
+ {
17410
+ "status": "skipped",
17411
+ "test": "[2] tests/test_modeling_common.py:2784: test requires Flash Attention 3",
17412
+ "count": 2
17413
+ },
17414
+ {
17415
+ "status": "skipped",
17416
+ "test": "[2] tests/test_modeling_common.py:2792: test requires Flash Attention 3",
17417
+ "count": 2
17418
+ },
17419
+ {
17420
+ "status": "skipped",
17421
+ "test": "[2] tests/test_modeling_common.py:2750: test requires the kernels library",
17422
+ "count": 2
17423
+ },
17424
+ {
17425
+ "status": "skipped",
17426
+ "test": "[2] tests/test_modeling_common.py:2758: test requires MPS",
17427
+ "count": 2
17428
+ },
17429
+ {
17430
+ "status": "skipped",
17431
+ "test": "[1] tests/generation/test_utils.py:1865: Model does not support position_ids",
17432
+ "count": 1
17433
+ },
17434
+ {
17435
+ "status": "skipped",
17436
+ "test": "[1] tests/generation/test_utils.py:2067: MBartForCausalLM does not support new attention mask API",
17437
+ "count": 1
17438
+ },
17439
+ {
17440
+ "status": "skipped",
17441
+ "test": "[1] tests/generation/test_utils.py:2074: MBartForCausalLM does not support new attention mask API",
17442
+ "count": 1
17443
+ },
17444
+ {
17445
+ "status": "skipped",
17446
+ "test": "[1] tests/test_modeling_common.py:3236: MBartStandaloneDecoderModelTest doesn't have the attribute `_torch_compile_train_cls`.",
17447
+ "count": 1
17448
+ }
17449
+ ]
17450
+ },
17451
+ {
17452
+ "model": "megatron_bert",
17453
+ "results": [
17454
+ {
17455
+ "status": "passed",
17456
+ "test": "tests/models/megatron_bert/test_modeling_megatron_bert.py::MegatronBertModelTest::test_flash_attn_2_fp32_ln",
17457
+ "count": 1
17458
+ },
17459
+ {
17460
+ "status": "passed",
17461
+ "test": "tests/models/megatron_bert/test_modeling_megatron_bert.py::MegatronBertModelTest::test_flash_attn_2_from_config",
17462
+ "count": 1
17463
+ },
17464
+ {
17465
+ "status": "skipped",
17466
+ "test": "[1] tests/test_modeling_common.py:3236: MegatronBertModelTest doesn't have the attribute `_torch_compile_train_cls`.",
17467
+ "count": 1
17468
+ },
17469
+ {
17470
+ "status": "skipped",
17471
+ "test": "[1] tests/test_modeling_common.py:3168: This model is not a composite model!",
17472
+ "count": 1
17473
+ },
17474
+ {
17475
+ "status": "skipped",
17476
+ "test": "[1] tests/test_modeling_common.py:2768: Model architecture does not support flash_attention_2, or setting its attention dynamically",
17477
+ "count": 1
17478
+ },
17479
+ {
17480
+ "status": "skipped",
17481
+ "test": "[1] tests/test_modeling_common.py:2776: Model architecture does not support flash_attention_2, or setting its attention dynamically",
17482
+ "count": 1
17483
+ },
17484
+ {
17485
+ "status": "skipped",
17486
+ "test": "[1] tests/test_modeling_common.py:3174: test requires Flash Attention 3",
17487
+ "count": 1
17488
+ },
17489
+ {
17490
+ "status": "skipped",
17491
+ "test": "[1] tests/test_modeling_common.py:3330: test requires Flash Attention 3",
17492
+ "count": 1
17493
+ },
17494
+ {
17495
+ "status": "skipped",
17496
+ "test": "[1] tests/test_modeling_common.py:2784: test requires Flash Attention 3",
17497
+ "count": 1
17498
+ },
17499
+ {
17500
+ "status": "skipped",
17501
+ "test": "[1] tests/test_modeling_common.py:2792: test requires Flash Attention 3",
17502
+ "count": 1
17503
+ },
17504
+ {
17505
+ "status": "skipped",
17506
+ "test": "[1] tests/test_modeling_common.py:2750: test requires the kernels library",
17507
+ "count": 1
17508
+ },
17509
+ {
17510
+ "status": "skipped",
17511
+ "test": "[1] tests/test_modeling_common.py:2758: test requires MPS",
17512
+ "count": 1
17513
+ }
17514
+ ]
17515
+ },
17516
  {
17517
  "model": "megatron_gpt2",
17518
  "results": []