hf-transformers-bot commited on
Commit
aedaa55
·
verified ·
1 Parent(s): a27f529

Upload 2025-11-11/ci_results_run_models_gpu/collated_reports_single-gpu_37d48bb.json with huggingface_hub

Browse files
2025-11-11/ci_results_run_models_gpu/collated_reports_single-gpu_37d48bb.json CHANGED
@@ -3,9 +3,9 @@
3
  "machine_type": "single-gpu",
4
  "commit_hash": "37d48bb",
5
  "total_status_count": {
6
- "passed": 1660,
7
  "failed": 140,
8
- "skipped": 5408,
9
  "error": 0,
10
  "null": 0
11
  },
@@ -17452,6 +17452,71 @@
17452
  }
17453
  ]
17454
  },
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
17455
  {
17456
  "model": "megatron_gpt2",
17457
  "results": []
 
3
  "machine_type": "single-gpu",
4
  "commit_hash": "37d48bb",
5
  "total_status_count": {
6
+ "passed": 1662,
7
  "failed": 140,
8
+ "skipped": 5418,
9
  "error": 0,
10
  "null": 0
11
  },
 
17452
  }
17453
  ]
17454
  },
17455
+ {
17456
+ "model": "megatron_bert",
17457
+ "results": [
17458
+ {
17459
+ "status": "passed",
17460
+ "test": "tests/models/megatron_bert/test_modeling_megatron_bert.py::MegatronBertModelTest::test_flash_attn_2_fp32_ln",
17461
+ "count": 1
17462
+ },
17463
+ {
17464
+ "status": "passed",
17465
+ "test": "tests/models/megatron_bert/test_modeling_megatron_bert.py::MegatronBertModelTest::test_flash_attn_2_from_config",
17466
+ "count": 1
17467
+ },
17468
+ {
17469
+ "status": "skipped",
17470
+ "test": "[1] tests/test_modeling_common.py:3236: MegatronBertModelTest doesn't have the attribute `_torch_compile_train_cls`.",
17471
+ "count": 1
17472
+ },
17473
+ {
17474
+ "status": "skipped",
17475
+ "test": "[1] tests/test_modeling_common.py:3168: This model is not a composite model!",
17476
+ "count": 1
17477
+ },
17478
+ {
17479
+ "status": "skipped",
17480
+ "test": "[1] tests/test_modeling_common.py:2768: Model architecture does not support flash_attention_2, or setting its attention dynamically",
17481
+ "count": 1
17482
+ },
17483
+ {
17484
+ "status": "skipped",
17485
+ "test": "[1] tests/test_modeling_common.py:2776: Model architecture does not support flash_attention_2, or setting its attention dynamically",
17486
+ "count": 1
17487
+ },
17488
+ {
17489
+ "status": "skipped",
17490
+ "test": "[1] tests/test_modeling_common.py:3174: test requires Flash Attention 3",
17491
+ "count": 1
17492
+ },
17493
+ {
17494
+ "status": "skipped",
17495
+ "test": "[1] tests/test_modeling_common.py:3330: test requires Flash Attention 3",
17496
+ "count": 1
17497
+ },
17498
+ {
17499
+ "status": "skipped",
17500
+ "test": "[1] tests/test_modeling_common.py:2784: test requires Flash Attention 3",
17501
+ "count": 1
17502
+ },
17503
+ {
17504
+ "status": "skipped",
17505
+ "test": "[1] tests/test_modeling_common.py:2792: test requires Flash Attention 3",
17506
+ "count": 1
17507
+ },
17508
+ {
17509
+ "status": "skipped",
17510
+ "test": "[1] tests/test_modeling_common.py:2750: test requires the kernels library",
17511
+ "count": 1
17512
+ },
17513
+ {
17514
+ "status": "skipped",
17515
+ "test": "[1] tests/test_modeling_common.py:2758: test requires MPS",
17516
+ "count": 1
17517
+ }
17518
+ ]
17519
+ },
17520
  {
17521
  "model": "megatron_gpt2",
17522
  "results": []