hf-transformers-bot commited on
Commit
82707d2
·
verified ·
1 Parent(s): 9c87fa5

Upload 2025-12-05/ci_results_run_models_gpu/collated_reports_multi-gpu_3a8d291.json with huggingface_hub

Browse files
2025-12-05/ci_results_run_models_gpu/collated_reports_multi-gpu_3a8d291.json CHANGED
@@ -3,9 +3,9 @@
3
  "machine_type": "multi-gpu",
4
  "commit_hash": "3a8d291",
5
  "total_status_count": {
6
- "passed": 1688,
7
  "failed": 110,
8
- "skipped": 5552,
9
  "error": 0,
10
  "null": 0
11
  },
@@ -17602,374 +17602,6 @@
17602
  }
17603
  ]
17604
  },
17605
- {
17606
- "model": "mbart50",
17607
- "results": []
17608
- },
17609
- {
17610
- "model": "mbart",
17611
- "results": [
17612
- {
17613
- "status": "passed",
17614
- "test": "tests/models/mbart/test_modeling_mbart.py::MBartModelTest::test_eager_matches_fa2_generate",
17615
- "count": 1
17616
- },
17617
- {
17618
- "status": "passed",
17619
- "test": "tests/models/mbart/test_modeling_mbart.py::MBartModelTest::test_flash_attn_2_fp32_ln",
17620
- "count": 1
17621
- },
17622
- {
17623
- "status": "passed",
17624
- "test": "tests/models/mbart/test_modeling_mbart.py::MBartModelTest::test_flash_attn_2_from_config",
17625
- "count": 1
17626
- },
17627
- {
17628
- "status": "passed",
17629
- "test": "tests/models/mbart/test_modeling_mbart.py::MBartModelTest::test_flash_attn_2_inference_equivalence",
17630
- "count": 1
17631
- },
17632
- {
17633
- "status": "passed",
17634
- "test": "tests/models/mbart/test_modeling_mbart.py::MBartModelTest::test_flash_attn_2_inference_equivalence_right_padding",
17635
- "count": 1
17636
- },
17637
- {
17638
- "status": "passed",
17639
- "test": "tests/models/mbart/test_modeling_mbart.py::MBartStandaloneDecoderModelTest::test_eager_matches_fa2_generate",
17640
- "count": 1
17641
- },
17642
- {
17643
- "status": "passed",
17644
- "test": "tests/models/mbart/test_modeling_mbart.py::MBartStandaloneDecoderModelTest::test_flash_attn_2_fp32_ln",
17645
- "count": 1
17646
- },
17647
- {
17648
- "status": "passed",
17649
- "test": "tests/models/mbart/test_modeling_mbart.py::MBartStandaloneDecoderModelTest::test_flash_attn_2_from_config",
17650
- "count": 1
17651
- },
17652
- {
17653
- "status": "passed",
17654
- "test": "tests/models/mbart/test_modeling_mbart.py::MBartStandaloneDecoderModelTest::test_flash_attn_2_inference_equivalence",
17655
- "count": 1
17656
- },
17657
- {
17658
- "status": "passed",
17659
- "test": "tests/models/mbart/test_modeling_mbart.py::MBartStandaloneDecoderModelTest::test_flash_attn_2_inference_equivalence_right_padding",
17660
- "count": 1
17661
- },
17662
- {
17663
- "status": "skipped",
17664
- "test": "[2] tests/generation/test_utils.py:1857: test requires Flash Attention 3",
17665
- "count": 2
17666
- },
17667
- {
17668
- "status": "skipped",
17669
- "test": "[1] tests/generation/test_utils.py:1865: Model is an encoder-decoder",
17670
- "count": 1
17671
- },
17672
- {
17673
- "status": "skipped",
17674
- "test": "[1] tests/generation/test_utils.py:2082: MBartForConditionalGeneration does not support new attention mask API",
17675
- "count": 1
17676
- },
17677
- {
17678
- "status": "skipped",
17679
- "test": "[1] tests/generation/test_utils.py:2089: MBartForConditionalGeneration does not support new attention mask API",
17680
- "count": 1
17681
- },
17682
- {
17683
- "status": "skipped",
17684
- "test": "[2] tests/generation/test_utils.py:2098: test requires Flash Attention 3",
17685
- "count": 2
17686
- },
17687
- {
17688
- "status": "skipped",
17689
- "test": "[2] tests/generation/test_utils.py:2105: test requires Flash Attention 3",
17690
- "count": 2
17691
- },
17692
- {
17693
- "status": "skipped",
17694
- "test": "[1] tests/test_modeling_common.py:3367: MBartModelTest doesn't have the attribute `_torch_compile_train_cls`.",
17695
- "count": 1
17696
- },
17697
- {
17698
- "status": "skipped",
17699
- "test": "[2] tests/test_modeling_common.py:3294: This model is not a composite model!",
17700
- "count": 2
17701
- },
17702
- {
17703
- "status": "skipped",
17704
- "test": "[2] tests/test_modeling_common.py:3300: test requires Flash Attention 3",
17705
- "count": 2
17706
- },
17707
- {
17708
- "status": "skipped",
17709
- "test": "[2] tests/test_modeling_common.py:3469: test requires Flash Attention 3",
17710
- "count": 2
17711
- },
17712
- {
17713
- "status": "skipped",
17714
- "test": "[2] tests/test_modeling_common.py:2910: test requires Flash Attention 3",
17715
- "count": 2
17716
- },
17717
- {
17718
- "status": "skipped",
17719
- "test": "[2] tests/test_modeling_common.py:2918: test requires Flash Attention 3",
17720
- "count": 2
17721
- },
17722
- {
17723
- "status": "skipped",
17724
- "test": "[2] tests/test_modeling_common.py:2876: test requires the kernels library",
17725
- "count": 2
17726
- },
17727
- {
17728
- "status": "skipped",
17729
- "test": "[2] tests/test_modeling_common.py:2884: test requires MPS",
17730
- "count": 2
17731
- },
17732
- {
17733
- "status": "skipped",
17734
- "test": "[1] tests/generation/test_utils.py:1865: Model does not support position_ids",
17735
- "count": 1
17736
- },
17737
- {
17738
- "status": "skipped",
17739
- "test": "[1] tests/generation/test_utils.py:2082: MBartForCausalLM does not support new attention mask API",
17740
- "count": 1
17741
- },
17742
- {
17743
- "status": "skipped",
17744
- "test": "[1] tests/generation/test_utils.py:2089: MBartForCausalLM does not support new attention mask API",
17745
- "count": 1
17746
- },
17747
- {
17748
- "status": "skipped",
17749
- "test": "[1] tests/test_modeling_common.py:3367: MBartStandaloneDecoderModelTest doesn't have the attribute `_torch_compile_train_cls`.",
17750
- "count": 1
17751
- }
17752
- ]
17753
- },
17754
- {
17755
- "model": "megatron_bert",
17756
- "results": [
17757
- {
17758
- "status": "passed",
17759
- "test": "tests/models/megatron_bert/test_modeling_megatron_bert.py::MegatronBertModelTest::test_flash_attn_2_fp32_ln",
17760
- "count": 1
17761
- },
17762
- {
17763
- "status": "passed",
17764
- "test": "tests/models/megatron_bert/test_modeling_megatron_bert.py::MegatronBertModelTest::test_flash_attn_2_from_config",
17765
- "count": 1
17766
- },
17767
- {
17768
- "status": "skipped",
17769
- "test": "[1] tests/test_modeling_common.py:3367: MegatronBertModelTest doesn't have the attribute `_torch_compile_train_cls`.",
17770
- "count": 1
17771
- },
17772
- {
17773
- "status": "skipped",
17774
- "test": "[1] tests/test_modeling_common.py:3294: This model is not a composite model!",
17775
- "count": 1
17776
- },
17777
- {
17778
- "status": "skipped",
17779
- "test": "[1] tests/test_modeling_common.py:2894: Model architecture does not support flash_attention_2, or setting its attention dynamically",
17780
- "count": 1
17781
- },
17782
- {
17783
- "status": "skipped",
17784
- "test": "[1] tests/test_modeling_common.py:2902: Model architecture does not support flash_attention_2, or setting its attention dynamically",
17785
- "count": 1
17786
- },
17787
- {
17788
- "status": "skipped",
17789
- "test": "[1] tests/test_modeling_common.py:3300: test requires Flash Attention 3",
17790
- "count": 1
17791
- },
17792
- {
17793
- "status": "skipped",
17794
- "test": "[1] tests/test_modeling_common.py:3469: test requires Flash Attention 3",
17795
- "count": 1
17796
- },
17797
- {
17798
- "status": "skipped",
17799
- "test": "[1] tests/test_modeling_common.py:2910: test requires Flash Attention 3",
17800
- "count": 1
17801
- },
17802
- {
17803
- "status": "skipped",
17804
- "test": "[1] tests/test_modeling_common.py:2918: test requires Flash Attention 3",
17805
- "count": 1
17806
- },
17807
- {
17808
- "status": "skipped",
17809
- "test": "[1] tests/test_modeling_common.py:2876: test requires the kernels library",
17810
- "count": 1
17811
- },
17812
- {
17813
- "status": "skipped",
17814
- "test": "[1] tests/test_modeling_common.py:2884: test requires MPS",
17815
- "count": 1
17816
- }
17817
- ]
17818
- },
17819
- {
17820
- "model": "megatron_gpt2",
17821
- "results": []
17822
- },
17823
- {
17824
- "model": "metaclip_2",
17825
- "results": [
17826
- {
17827
- "status": "passed",
17828
- "test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2VisionModelTest::test_flash_attn_2_fp32_ln",
17829
- "count": 1
17830
- },
17831
- {
17832
- "status": "passed",
17833
- "test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2VisionModelTest::test_flash_attn_2_from_config",
17834
- "count": 1
17835
- },
17836
- {
17837
- "status": "passed",
17838
- "test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2VisionModelTest::test_flash_attn_2_inference_equivalence",
17839
- "count": 1
17840
- },
17841
- {
17842
- "status": "passed",
17843
- "test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2VisionModelTest::test_flash_attn_2_inference_equivalence_right_padding",
17844
- "count": 1
17845
- },
17846
- {
17847
- "status": "passed",
17848
- "test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2TextModelTest::test_flash_attn_2_fp32_ln",
17849
- "count": 1
17850
- },
17851
- {
17852
- "status": "passed",
17853
- "test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2TextModelTest::test_flash_attn_2_from_config",
17854
- "count": 1
17855
- },
17856
- {
17857
- "status": "passed",
17858
- "test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2TextModelTest::test_flash_attn_2_inference_equivalence",
17859
- "count": 1
17860
- },
17861
- {
17862
- "status": "passed",
17863
- "test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2TextModelTest::test_flash_attn_2_inference_equivalence_right_padding",
17864
- "count": 1
17865
- },
17866
- {
17867
- "status": "passed",
17868
- "test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2ModelTest::test_flash_attn_2_can_dispatch_composite_models",
17869
- "count": 1
17870
- },
17871
- {
17872
- "status": "passed",
17873
- "test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2ModelTest::test_flash_attn_2_fp32_ln",
17874
- "count": 1
17875
- },
17876
- {
17877
- "status": "passed",
17878
- "test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2ModelTest::test_flash_attn_2_from_config",
17879
- "count": 1
17880
- },
17881
- {
17882
- "status": "passed",
17883
- "test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2ModelTest::test_flash_attn_2_inference_equivalence",
17884
- "count": 1
17885
- },
17886
- {
17887
- "status": "passed",
17888
- "test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2ModelTest::test_flash_attn_2_inference_equivalence_right_padding",
17889
- "count": 1
17890
- },
17891
- {
17892
- "status": "passed",
17893
- "test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2ForImageClassificationModelTest::test_flash_attn_2_can_dispatch_composite_models",
17894
- "count": 1
17895
- },
17896
- {
17897
- "status": "passed",
17898
- "test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2ForImageClassificationModelTest::test_flash_attn_2_fp32_ln",
17899
- "count": 1
17900
- },
17901
- {
17902
- "status": "passed",
17903
- "test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2ForImageClassificationModelTest::test_flash_attn_2_from_config",
17904
- "count": 1
17905
- },
17906
- {
17907
- "status": "passed",
17908
- "test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2ForImageClassificationModelTest::test_flash_attn_2_inference_equivalence",
17909
- "count": 1
17910
- },
17911
- {
17912
- "status": "passed",
17913
- "test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2ForImageClassificationModelTest::test_flash_attn_2_inference_equivalence_right_padding",
17914
- "count": 1
17915
- },
17916
- {
17917
- "status": "skipped",
17918
- "test": "[1] tests/test_modeling_common.py:3367: MetaClip2VisionModelTest doesn't have the attribute `_torch_compile_train_cls`.",
17919
- "count": 1
17920
- },
17921
- {
17922
- "status": "skipped",
17923
- "test": "[2] tests/test_modeling_common.py:3294: This model is not a composite model!",
17924
- "count": 2
17925
- },
17926
- {
17927
- "status": "skipped",
17928
- "test": "[4] tests/test_modeling_common.py:3300: test requires Flash Attention 3",
17929
- "count": 4
17930
- },
17931
- {
17932
- "status": "skipped",
17933
- "test": "[4] tests/test_modeling_common.py:3469: test requires Flash Attention 3",
17934
- "count": 4
17935
- },
17936
- {
17937
- "status": "skipped",
17938
- "test": "[4] tests/test_modeling_common.py:2910: test requires Flash Attention 3",
17939
- "count": 4
17940
- },
17941
- {
17942
- "status": "skipped",
17943
- "test": "[4] tests/test_modeling_common.py:2918: test requires Flash Attention 3",
17944
- "count": 4
17945
- },
17946
- {
17947
- "status": "skipped",
17948
- "test": "[4] tests/test_modeling_common.py:2876: test requires the kernels library",
17949
- "count": 4
17950
- },
17951
- {
17952
- "status": "skipped",
17953
- "test": "[4] tests/test_modeling_common.py:2884: test requires MPS",
17954
- "count": 4
17955
- },
17956
- {
17957
- "status": "skipped",
17958
- "test": "[1] tests/test_modeling_common.py:3367: MetaClip2TextModelTest doesn't have the attribute `_torch_compile_train_cls`.",
17959
- "count": 1
17960
- },
17961
- {
17962
- "status": "skipped",
17963
- "test": "[1] tests/test_modeling_common.py:3367: MetaClip2ModelTest doesn't have the attribute `_torch_compile_train_cls`.",
17964
- "count": 1
17965
- },
17966
- {
17967
- "status": "skipped",
17968
- "test": "[1] tests/test_modeling_common.py:3367: MetaClip2ForImageClassificationModelTest doesn't have the attribute `_torch_compile_train_cls`.",
17969
- "count": 1
17970
- }
17971
- ]
17972
- },
17973
  {
17974
  "model": "mgp_str",
17975
  "results": [
 
3
  "machine_type": "multi-gpu",
4
  "commit_hash": "3a8d291",
5
  "total_status_count": {
6
+ "passed": 1658,
7
  "failed": 110,
8
+ "skipped": 5484,
9
  "error": 0,
10
  "null": 0
11
  },
 
17602
  }
17603
  ]
17604
  },
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
17605
  {
17606
  "model": "mgp_str",
17607
  "results": [