hf-transformers-bot commited on
Commit
de891a6
·
verified ·
1 Parent(s): 346f5d9

Upload 2025-12-04/ci_results_run_models_gpu/collated_reports_multi-gpu_a48d68c.json with huggingface_hub

Browse files
2025-12-04/ci_results_run_models_gpu/collated_reports_multi-gpu_a48d68c.json CHANGED
@@ -3,9 +3,9 @@
3
  "machine_type": "multi-gpu",
4
  "commit_hash": "a48d68c",
5
  "total_status_count": {
6
- "passed": 1635,
7
  "failed": 110,
8
- "skipped": 5363,
9
  "error": 0,
10
  "null": 0
11
  },
@@ -16942,6 +16942,1034 @@
16942
  }
16943
  ]
16944
  },
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
16945
  {
16946
  "model": "mgp_str",
16947
  "results": [
 
3
  "machine_type": "multi-gpu",
4
  "commit_hash": "a48d68c",
5
  "total_status_count": {
6
+ "passed": 1689,
7
  "failed": 110,
8
+ "skipped": 5551,
9
  "error": 0,
10
  "null": 0
11
  },
 
16942
  }
16943
  ]
16944
  },
16945
+ {
16946
+ "model": "m2m_100",
16947
+ "results": [
16948
+ {
16949
+ "status": "passed",
16950
+ "test": "tests/models/m2m_100/test_modeling_m2m_100.py::M2M100ModelTest::test_eager_matches_fa2_generate",
16951
+ "count": 1
16952
+ },
16953
+ {
16954
+ "status": "passed",
16955
+ "test": "tests/models/m2m_100/test_modeling_m2m_100.py::M2M100ModelTest::test_flash_attn_2_fp32_ln",
16956
+ "count": 1
16957
+ },
16958
+ {
16959
+ "status": "passed",
16960
+ "test": "tests/models/m2m_100/test_modeling_m2m_100.py::M2M100ModelTest::test_flash_attn_2_from_config",
16961
+ "count": 1
16962
+ },
16963
+ {
16964
+ "status": "passed",
16965
+ "test": "tests/models/m2m_100/test_modeling_m2m_100.py::M2M100ModelTest::test_flash_attn_2_inference_equivalence",
16966
+ "count": 1
16967
+ },
16968
+ {
16969
+ "status": "passed",
16970
+ "test": "tests/models/m2m_100/test_modeling_m2m_100.py::M2M100ModelTest::test_flash_attn_2_inference_equivalence_right_padding",
16971
+ "count": 1
16972
+ },
16973
+ {
16974
+ "status": "passed",
16975
+ "test": "tests/models/m2m_100/test_modeling_m2m_100.py::M2M100ModelIntegrationTests::test_flash_attn_2_seq_to_seq_generation",
16976
+ "count": 1
16977
+ },
16978
+ {
16979
+ "status": "skipped",
16980
+ "test": "[1] tests/generation/test_utils.py:1857: test requires Flash Attention 3",
16981
+ "count": 1
16982
+ },
16983
+ {
16984
+ "status": "skipped",
16985
+ "test": "[1] tests/generation/test_utils.py:1865: Model is an encoder-decoder",
16986
+ "count": 1
16987
+ },
16988
+ {
16989
+ "status": "skipped",
16990
+ "test": "[1] tests/generation/test_utils.py:2082: M2M100ForConditionalGeneration does not support new attention mask API",
16991
+ "count": 1
16992
+ },
16993
+ {
16994
+ "status": "skipped",
16995
+ "test": "[1] tests/generation/test_utils.py:2089: M2M100ForConditionalGeneration does not support new attention mask API",
16996
+ "count": 1
16997
+ },
16998
+ {
16999
+ "status": "skipped",
17000
+ "test": "[1] tests/generation/test_utils.py:2098: test requires Flash Attention 3",
17001
+ "count": 1
17002
+ },
17003
+ {
17004
+ "status": "skipped",
17005
+ "test": "[1] tests/generation/test_utils.py:2105: test requires Flash Attention 3",
17006
+ "count": 1
17007
+ },
17008
+ {
17009
+ "status": "skipped",
17010
+ "test": "[1] tests/test_modeling_common.py:3367: M2M100ModelTest doesn't have the attribute `_torch_compile_train_cls`.",
17011
+ "count": 1
17012
+ },
17013
+ {
17014
+ "status": "skipped",
17015
+ "test": "[1] tests/test_modeling_common.py:3294: This model is not a composite model!",
17016
+ "count": 1
17017
+ },
17018
+ {
17019
+ "status": "skipped",
17020
+ "test": "[1] tests/test_modeling_common.py:3300: test requires Flash Attention 3",
17021
+ "count": 1
17022
+ },
17023
+ {
17024
+ "status": "skipped",
17025
+ "test": "[1] tests/test_modeling_common.py:3469: test requires Flash Attention 3",
17026
+ "count": 1
17027
+ },
17028
+ {
17029
+ "status": "skipped",
17030
+ "test": "[1] tests/test_modeling_common.py:2910: test requires Flash Attention 3",
17031
+ "count": 1
17032
+ },
17033
+ {
17034
+ "status": "skipped",
17035
+ "test": "[1] tests/test_modeling_common.py:2918: test requires Flash Attention 3",
17036
+ "count": 1
17037
+ },
17038
+ {
17039
+ "status": "skipped",
17040
+ "test": "[1] tests/test_modeling_common.py:2876: test requires the kernels library",
17041
+ "count": 1
17042
+ },
17043
+ {
17044
+ "status": "skipped",
17045
+ "test": "[1] tests/test_modeling_common.py:2884: test requires MPS",
17046
+ "count": 1
17047
+ }
17048
+ ]
17049
+ },
17050
+ {
17051
+ "model": "mamba2",
17052
+ "results": [
17053
+ {
17054
+ "status": "skipped",
17055
+ "test": "[1] tests/generation/test_utils.py:1849: Mamba2ForCausalLM does not support `attn_implementation=flash_attention_2`",
17056
+ "count": 1
17057
+ },
17058
+ {
17059
+ "status": "skipped",
17060
+ "test": "[1] tests/generation/test_utils.py:1857: test requires Flash Attention 3",
17061
+ "count": 1
17062
+ },
17063
+ {
17064
+ "status": "skipped",
17065
+ "test": "[1] tests/generation/test_utils.py:1865: Mamba2ForCausalLM does not support Flash Attention.",
17066
+ "count": 1
17067
+ },
17068
+ {
17069
+ "status": "skipped",
17070
+ "test": "[1] tests/generation/test_utils.py:2082: Model architecture does not support attentions",
17071
+ "count": 1
17072
+ },
17073
+ {
17074
+ "status": "skipped",
17075
+ "test": "[1] tests/generation/test_utils.py:2089: Model architecture does not support attentions",
17076
+ "count": 1
17077
+ },
17078
+ {
17079
+ "status": "skipped",
17080
+ "test": "[1] tests/generation/test_utils.py:2098: test requires Flash Attention 3",
17081
+ "count": 1
17082
+ },
17083
+ {
17084
+ "status": "skipped",
17085
+ "test": "[1] tests/generation/test_utils.py:2105: test requires Flash Attention 3",
17086
+ "count": 1
17087
+ },
17088
+ {
17089
+ "status": "skipped",
17090
+ "test": "[1] tests/test_modeling_common.py:3367: Mamba2ModelTest doesn't have the attribute `_torch_compile_train_cls`.",
17091
+ "count": 1
17092
+ },
17093
+ {
17094
+ "status": "skipped",
17095
+ "test": "[1] tests/test_modeling_common.py:3294: Model architecture does not support attentions",
17096
+ "count": 1
17097
+ },
17098
+ {
17099
+ "status": "skipped",
17100
+ "test": "[1] tests/test_modeling_common.py:3306: Model architecture does not support attentions",
17101
+ "count": 1
17102
+ },
17103
+ {
17104
+ "status": "skipped",
17105
+ "test": "[1] tests/test_modeling_common.py:3462: Model architecture does not support attentions",
17106
+ "count": 1
17107
+ },
17108
+ {
17109
+ "status": "skipped",
17110
+ "test": "[1] tests/test_modeling_common.py:2894: Model architecture does not support attentions",
17111
+ "count": 1
17112
+ },
17113
+ {
17114
+ "status": "skipped",
17115
+ "test": "[1] tests/test_modeling_common.py:2902: Model architecture does not support attentions",
17116
+ "count": 1
17117
+ },
17118
+ {
17119
+ "status": "skipped",
17120
+ "test": "[1] tests/test_modeling_common.py:3300: test requires Flash Attention 3",
17121
+ "count": 1
17122
+ },
17123
+ {
17124
+ "status": "skipped",
17125
+ "test": "[1] tests/test_modeling_common.py:3469: test requires Flash Attention 3",
17126
+ "count": 1
17127
+ },
17128
+ {
17129
+ "status": "skipped",
17130
+ "test": "[1] tests/test_modeling_common.py:2910: test requires Flash Attention 3",
17131
+ "count": 1
17132
+ },
17133
+ {
17134
+ "status": "skipped",
17135
+ "test": "[1] tests/test_modeling_common.py:2918: test requires Flash Attention 3",
17136
+ "count": 1
17137
+ },
17138
+ {
17139
+ "status": "skipped",
17140
+ "test": "[1] tests/test_modeling_common.py:2876: test requires the kernels library",
17141
+ "count": 1
17142
+ },
17143
+ {
17144
+ "status": "skipped",
17145
+ "test": "[1] tests/test_modeling_common.py:2884: test requires MPS",
17146
+ "count": 1
17147
+ }
17148
+ ]
17149
+ },
17150
+ {
17151
+ "model": "mamba",
17152
+ "results": [
17153
+ {
17154
+ "status": "skipped",
17155
+ "test": "[1] tests/generation/test_utils.py:1849: MambaForCausalLM does not support `attn_implementation=flash_attention_2`",
17156
+ "count": 1
17157
+ },
17158
+ {
17159
+ "status": "skipped",
17160
+ "test": "[1] tests/generation/test_utils.py:1857: test requires Flash Attention 3",
17161
+ "count": 1
17162
+ },
17163
+ {
17164
+ "status": "skipped",
17165
+ "test": "[1] tests/generation/test_utils.py:1865: MambaForCausalLM does not support Flash Attention.",
17166
+ "count": 1
17167
+ },
17168
+ {
17169
+ "status": "skipped",
17170
+ "test": "[1] tests/generation/test_utils.py:2082: Model architecture does not support attentions",
17171
+ "count": 1
17172
+ },
17173
+ {
17174
+ "status": "skipped",
17175
+ "test": "[1] tests/generation/test_utils.py:2089: Model architecture does not support attentions",
17176
+ "count": 1
17177
+ },
17178
+ {
17179
+ "status": "skipped",
17180
+ "test": "[1] tests/generation/test_utils.py:2098: test requires Flash Attention 3",
17181
+ "count": 1
17182
+ },
17183
+ {
17184
+ "status": "skipped",
17185
+ "test": "[1] tests/generation/test_utils.py:2105: test requires Flash Attention 3",
17186
+ "count": 1
17187
+ },
17188
+ {
17189
+ "status": "skipped",
17190
+ "test": "[1] tests/test_modeling_common.py:3367: MambaModelTest doesn't have the attribute `_torch_compile_train_cls`.",
17191
+ "count": 1
17192
+ },
17193
+ {
17194
+ "status": "skipped",
17195
+ "test": "[1] tests/test_modeling_common.py:3294: Model architecture does not support attentions",
17196
+ "count": 1
17197
+ },
17198
+ {
17199
+ "status": "skipped",
17200
+ "test": "[1] tests/test_modeling_common.py:3306: Model architecture does not support attentions",
17201
+ "count": 1
17202
+ },
17203
+ {
17204
+ "status": "skipped",
17205
+ "test": "[1] tests/test_modeling_common.py:3462: Model architecture does not support attentions",
17206
+ "count": 1
17207
+ },
17208
+ {
17209
+ "status": "skipped",
17210
+ "test": "[1] tests/test_modeling_common.py:2894: Model architecture does not support attentions",
17211
+ "count": 1
17212
+ },
17213
+ {
17214
+ "status": "skipped",
17215
+ "test": "[1] tests/test_modeling_common.py:2902: Model architecture does not support attentions",
17216
+ "count": 1
17217
+ },
17218
+ {
17219
+ "status": "skipped",
17220
+ "test": "[1] tests/test_modeling_common.py:3300: test requires Flash Attention 3",
17221
+ "count": 1
17222
+ },
17223
+ {
17224
+ "status": "skipped",
17225
+ "test": "[1] tests/test_modeling_common.py:3469: test requires Flash Attention 3",
17226
+ "count": 1
17227
+ },
17228
+ {
17229
+ "status": "skipped",
17230
+ "test": "[1] tests/test_modeling_common.py:2910: test requires Flash Attention 3",
17231
+ "count": 1
17232
+ },
17233
+ {
17234
+ "status": "skipped",
17235
+ "test": "[1] tests/test_modeling_common.py:2918: test requires Flash Attention 3",
17236
+ "count": 1
17237
+ },
17238
+ {
17239
+ "status": "skipped",
17240
+ "test": "[1] tests/test_modeling_common.py:2876: test requires the kernels library",
17241
+ "count": 1
17242
+ },
17243
+ {
17244
+ "status": "skipped",
17245
+ "test": "[1] tests/test_modeling_common.py:2884: test requires MPS",
17246
+ "count": 1
17247
+ }
17248
+ ]
17249
+ },
17250
+ {
17251
+ "model": "marian",
17252
+ "results": [
17253
+ {
17254
+ "status": "passed",
17255
+ "test": "tests/models/marian/test_modeling_marian.py::MarianModelTest::test_eager_matches_fa2_generate",
17256
+ "count": 1
17257
+ },
17258
+ {
17259
+ "status": "passed",
17260
+ "test": "tests/models/marian/test_modeling_marian.py::MarianModelTest::test_flash_attn_2_fp32_ln",
17261
+ "count": 1
17262
+ },
17263
+ {
17264
+ "status": "passed",
17265
+ "test": "tests/models/marian/test_modeling_marian.py::MarianModelTest::test_flash_attn_2_from_config",
17266
+ "count": 1
17267
+ },
17268
+ {
17269
+ "status": "passed",
17270
+ "test": "tests/models/marian/test_modeling_marian.py::MarianModelTest::test_flash_attn_2_inference_equivalence",
17271
+ "count": 1
17272
+ },
17273
+ {
17274
+ "status": "passed",
17275
+ "test": "tests/models/marian/test_modeling_marian.py::MarianModelTest::test_flash_attn_2_inference_equivalence_right_padding",
17276
+ "count": 1
17277
+ },
17278
+ {
17279
+ "status": "passed",
17280
+ "test": "tests/models/marian/test_modeling_marian.py::MarianStandaloneDecoderModelTest::test_eager_matches_fa2_generate",
17281
+ "count": 1
17282
+ },
17283
+ {
17284
+ "status": "passed",
17285
+ "test": "tests/models/marian/test_modeling_marian.py::MarianStandaloneDecoderModelTest::test_flash_attn_2_fp32_ln",
17286
+ "count": 1
17287
+ },
17288
+ {
17289
+ "status": "passed",
17290
+ "test": "tests/models/marian/test_modeling_marian.py::MarianStandaloneDecoderModelTest::test_flash_attn_2_from_config",
17291
+ "count": 1
17292
+ },
17293
+ {
17294
+ "status": "passed",
17295
+ "test": "tests/models/marian/test_modeling_marian.py::MarianStandaloneDecoderModelTest::test_flash_attn_2_inference_equivalence",
17296
+ "count": 1
17297
+ },
17298
+ {
17299
+ "status": "passed",
17300
+ "test": "tests/models/marian/test_modeling_marian.py::MarianStandaloneDecoderModelTest::test_flash_attn_2_inference_equivalence_right_padding",
17301
+ "count": 1
17302
+ },
17303
+ {
17304
+ "status": "skipped",
17305
+ "test": "[2] tests/generation/test_utils.py:1857: test requires Flash Attention 3",
17306
+ "count": 2
17307
+ },
17308
+ {
17309
+ "status": "skipped",
17310
+ "test": "[1] tests/generation/test_utils.py:1865: Model is an encoder-decoder",
17311
+ "count": 1
17312
+ },
17313
+ {
17314
+ "status": "skipped",
17315
+ "test": "[1] tests/generation/test_utils.py:2082: MarianMTModel does not support new attention mask API",
17316
+ "count": 1
17317
+ },
17318
+ {
17319
+ "status": "skipped",
17320
+ "test": "[1] tests/generation/test_utils.py:2089: MarianMTModel does not support new attention mask API",
17321
+ "count": 1
17322
+ },
17323
+ {
17324
+ "status": "skipped",
17325
+ "test": "[2] tests/generation/test_utils.py:2098: test requires Flash Attention 3",
17326
+ "count": 2
17327
+ },
17328
+ {
17329
+ "status": "skipped",
17330
+ "test": "[2] tests/generation/test_utils.py:2105: test requires Flash Attention 3",
17331
+ "count": 2
17332
+ },
17333
+ {
17334
+ "status": "skipped",
17335
+ "test": "[1] tests/test_modeling_common.py:3367: MarianModelTest doesn't have the attribute `_torch_compile_train_cls`.",
17336
+ "count": 1
17337
+ },
17338
+ {
17339
+ "status": "skipped",
17340
+ "test": "[2] tests/test_modeling_common.py:3294: This model is not a composite model!",
17341
+ "count": 2
17342
+ },
17343
+ {
17344
+ "status": "skipped",
17345
+ "test": "[2] tests/test_modeling_common.py:3300: test requires Flash Attention 3",
17346
+ "count": 2
17347
+ },
17348
+ {
17349
+ "status": "skipped",
17350
+ "test": "[2] tests/test_modeling_common.py:3469: test requires Flash Attention 3",
17351
+ "count": 2
17352
+ },
17353
+ {
17354
+ "status": "skipped",
17355
+ "test": "[2] tests/test_modeling_common.py:2910: test requires Flash Attention 3",
17356
+ "count": 2
17357
+ },
17358
+ {
17359
+ "status": "skipped",
17360
+ "test": "[2] tests/test_modeling_common.py:2918: test requires Flash Attention 3",
17361
+ "count": 2
17362
+ },
17363
+ {
17364
+ "status": "skipped",
17365
+ "test": "[2] tests/test_modeling_common.py:2876: test requires the kernels library",
17366
+ "count": 2
17367
+ },
17368
+ {
17369
+ "status": "skipped",
17370
+ "test": "[2] tests/test_modeling_common.py:2884: test requires MPS",
17371
+ "count": 2
17372
+ },
17373
+ {
17374
+ "status": "skipped",
17375
+ "test": "[1] tests/generation/test_utils.py:1865: Model does not support position_ids",
17376
+ "count": 1
17377
+ },
17378
+ {
17379
+ "status": "skipped",
17380
+ "test": "[1] tests/generation/test_utils.py:2082: MarianForCausalLM does not support new attention mask API",
17381
+ "count": 1
17382
+ },
17383
+ {
17384
+ "status": "skipped",
17385
+ "test": "[1] tests/generation/test_utils.py:2089: MarianForCausalLM does not support new attention mask API",
17386
+ "count": 1
17387
+ },
17388
+ {
17389
+ "status": "skipped",
17390
+ "test": "[1] tests/test_modeling_common.py:3367: MarianStandaloneDecoderModelTest doesn't have the attribute `_torch_compile_train_cls`.",
17391
+ "count": 1
17392
+ }
17393
+ ]
17394
+ },
17395
+ {
17396
+ "model": "markuplm",
17397
+ "results": [
17398
+ {
17399
+ "status": "passed",
17400
+ "test": "tests/models/markuplm/test_modeling_markuplm.py::MarkupLMModelTest::test_flash_attn_2_fp32_ln",
17401
+ "count": 1
17402
+ },
17403
+ {
17404
+ "status": "passed",
17405
+ "test": "tests/models/markuplm/test_modeling_markuplm.py::MarkupLMModelTest::test_flash_attn_2_from_config",
17406
+ "count": 1
17407
+ },
17408
+ {
17409
+ "status": "skipped",
17410
+ "test": "[1] tests/test_modeling_common.py:3367: MarkupLMModelTest doesn't have the attribute `_torch_compile_train_cls`.",
17411
+ "count": 1
17412
+ },
17413
+ {
17414
+ "status": "skipped",
17415
+ "test": "[1] tests/test_modeling_common.py:3294: This model is not a composite model!",
17416
+ "count": 1
17417
+ },
17418
+ {
17419
+ "status": "skipped",
17420
+ "test": "[1] tests/test_modeling_common.py:2894: Model architecture does not support flash_attention_2, or setting its attention dynamically",
17421
+ "count": 1
17422
+ },
17423
+ {
17424
+ "status": "skipped",
17425
+ "test": "[1] tests/test_modeling_common.py:2902: Model architecture does not support flash_attention_2, or setting its attention dynamically",
17426
+ "count": 1
17427
+ },
17428
+ {
17429
+ "status": "skipped",
17430
+ "test": "[1] tests/test_modeling_common.py:3300: test requires Flash Attention 3",
17431
+ "count": 1
17432
+ },
17433
+ {
17434
+ "status": "skipped",
17435
+ "test": "[1] tests/test_modeling_common.py:3469: test requires Flash Attention 3",
17436
+ "count": 1
17437
+ },
17438
+ {
17439
+ "status": "skipped",
17440
+ "test": "[1] tests/test_modeling_common.py:2910: test requires Flash Attention 3",
17441
+ "count": 1
17442
+ },
17443
+ {
17444
+ "status": "skipped",
17445
+ "test": "[1] tests/test_modeling_common.py:2918: test requires Flash Attention 3",
17446
+ "count": 1
17447
+ },
17448
+ {
17449
+ "status": "skipped",
17450
+ "test": "[1] tests/test_modeling_common.py:2876: test requires the kernels library",
17451
+ "count": 1
17452
+ },
17453
+ {
17454
+ "status": "skipped",
17455
+ "test": "[1] tests/test_modeling_common.py:2884: test requires MPS",
17456
+ "count": 1
17457
+ }
17458
+ ]
17459
+ },
17460
+ {
17461
+ "model": "mask2former",
17462
+ "results": [
17463
+ {
17464
+ "status": "passed",
17465
+ "test": "tests/models/mask2former/test_modeling_mask2former.py::Mask2FormerModelTest::test_flash_attn_2_fp32_ln",
17466
+ "count": 1
17467
+ },
17468
+ {
17469
+ "status": "passed",
17470
+ "test": "tests/models/mask2former/test_modeling_mask2former.py::Mask2FormerModelTest::test_flash_attn_2_from_config",
17471
+ "count": 1
17472
+ },
17473
+ {
17474
+ "status": "skipped",
17475
+ "test": "[1] tests/test_modeling_common.py:3367: Mask2FormerModelTest doesn't have the attribute `_torch_compile_train_cls`.",
17476
+ "count": 1
17477
+ },
17478
+ {
17479
+ "status": "skipped",
17480
+ "test": "[1] tests/test_modeling_common.py:3294: This model is not a composite model!",
17481
+ "count": 1
17482
+ },
17483
+ {
17484
+ "status": "skipped",
17485
+ "test": "[1] tests/test_modeling_common.py:2894: Model architecture does not support flash_attention_2, or setting its attention dynamically",
17486
+ "count": 1
17487
+ },
17488
+ {
17489
+ "status": "skipped",
17490
+ "test": "[1] tests/test_modeling_common.py:2902: Model architecture does not support flash_attention_2, or setting its attention dynamically",
17491
+ "count": 1
17492
+ },
17493
+ {
17494
+ "status": "skipped",
17495
+ "test": "[1] tests/test_modeling_common.py:3300: test requires Flash Attention 3",
17496
+ "count": 1
17497
+ },
17498
+ {
17499
+ "status": "skipped",
17500
+ "test": "[1] tests/test_modeling_common.py:3469: test requires Flash Attention 3",
17501
+ "count": 1
17502
+ },
17503
+ {
17504
+ "status": "skipped",
17505
+ "test": "[1] tests/test_modeling_common.py:2910: test requires Flash Attention 3",
17506
+ "count": 1
17507
+ },
17508
+ {
17509
+ "status": "skipped",
17510
+ "test": "[1] tests/test_modeling_common.py:2918: test requires Flash Attention 3",
17511
+ "count": 1
17512
+ },
17513
+ {
17514
+ "status": "skipped",
17515
+ "test": "[1] tests/test_modeling_common.py:2876: test requires the kernels library",
17516
+ "count": 1
17517
+ },
17518
+ {
17519
+ "status": "skipped",
17520
+ "test": "[1] tests/test_modeling_common.py:2884: test requires MPS",
17521
+ "count": 1
17522
+ }
17523
+ ]
17524
+ },
17525
+ {
17526
+ "model": "maskformer",
17527
+ "results": [
17528
+ {
17529
+ "status": "passed",
17530
+ "test": "tests/models/maskformer/test_modeling_maskformer.py::MaskFormerModelTest::test_flash_attn_2_fp32_ln",
17531
+ "count": 1
17532
+ },
17533
+ {
17534
+ "status": "passed",
17535
+ "test": "tests/models/maskformer/test_modeling_maskformer.py::MaskFormerModelTest::test_flash_attn_2_from_config",
17536
+ "count": 1
17537
+ },
17538
+ {
17539
+ "status": "passed",
17540
+ "test": "tests/models/maskformer/test_modeling_maskformer_swin.py::MaskFormerSwinModelTest::test_flash_attn_2_fp32_ln",
17541
+ "count": 1
17542
+ },
17543
+ {
17544
+ "status": "passed",
17545
+ "test": "tests/models/maskformer/test_modeling_maskformer_swin.py::MaskFormerSwinModelTest::test_flash_attn_2_from_config",
17546
+ "count": 1
17547
+ },
17548
+ {
17549
+ "status": "skipped",
17550
+ "test": "[1] tests/test_modeling_common.py:3367: MaskFormerModelTest doesn't have the attribute `_torch_compile_train_cls`.",
17551
+ "count": 1
17552
+ },
17553
+ {
17554
+ "status": "skipped",
17555
+ "test": "[2] tests/test_modeling_common.py:3294: This model is not a composite model!",
17556
+ "count": 2
17557
+ },
17558
+ {
17559
+ "status": "skipped",
17560
+ "test": "[2] tests/test_modeling_common.py:2894: Model architecture does not support flash_attention_2, or setting its attention dynamically",
17561
+ "count": 2
17562
+ },
17563
+ {
17564
+ "status": "skipped",
17565
+ "test": "[2] tests/test_modeling_common.py:2902: Model architecture does not support flash_attention_2, or setting its attention dynamically",
17566
+ "count": 2
17567
+ },
17568
+ {
17569
+ "status": "skipped",
17570
+ "test": "[2] tests/test_modeling_common.py:3300: test requires Flash Attention 3",
17571
+ "count": 2
17572
+ },
17573
+ {
17574
+ "status": "skipped",
17575
+ "test": "[2] tests/test_modeling_common.py:3469: test requires Flash Attention 3",
17576
+ "count": 2
17577
+ },
17578
+ {
17579
+ "status": "skipped",
17580
+ "test": "[2] tests/test_modeling_common.py:2910: test requires Flash Attention 3",
17581
+ "count": 2
17582
+ },
17583
+ {
17584
+ "status": "skipped",
17585
+ "test": "[2] tests/test_modeling_common.py:2918: test requires Flash Attention 3",
17586
+ "count": 2
17587
+ },
17588
+ {
17589
+ "status": "skipped",
17590
+ "test": "[2] tests/test_modeling_common.py:2876: test requires the kernels library",
17591
+ "count": 2
17592
+ },
17593
+ {
17594
+ "status": "skipped",
17595
+ "test": "[2] tests/test_modeling_common.py:2884: test requires MPS",
17596
+ "count": 2
17597
+ },
17598
+ {
17599
+ "status": "skipped",
17600
+ "test": "[1] tests/test_modeling_common.py:3367: MaskFormerSwinModelTest doesn't have the attribute `_torch_compile_train_cls`.",
17601
+ "count": 1
17602
+ }
17603
+ ]
17604
+ },
17605
+ {
17606
+ "model": "mbart50",
17607
+ "results": []
17608
+ },
17609
+ {
17610
+ "model": "mbart",
17611
+ "results": [
17612
+ {
17613
+ "status": "passed",
17614
+ "test": "tests/models/mbart/test_modeling_mbart.py::MBartModelTest::test_eager_matches_fa2_generate",
17615
+ "count": 1
17616
+ },
17617
+ {
17618
+ "status": "passed",
17619
+ "test": "tests/models/mbart/test_modeling_mbart.py::MBartModelTest::test_flash_attn_2_fp32_ln",
17620
+ "count": 1
17621
+ },
17622
+ {
17623
+ "status": "passed",
17624
+ "test": "tests/models/mbart/test_modeling_mbart.py::MBartModelTest::test_flash_attn_2_from_config",
17625
+ "count": 1
17626
+ },
17627
+ {
17628
+ "status": "passed",
17629
+ "test": "tests/models/mbart/test_modeling_mbart.py::MBartModelTest::test_flash_attn_2_inference_equivalence",
17630
+ "count": 1
17631
+ },
17632
+ {
17633
+ "status": "passed",
17634
+ "test": "tests/models/mbart/test_modeling_mbart.py::MBartModelTest::test_flash_attn_2_inference_equivalence_right_padding",
17635
+ "count": 1
17636
+ },
17637
+ {
17638
+ "status": "passed",
17639
+ "test": "tests/models/mbart/test_modeling_mbart.py::MBartStandaloneDecoderModelTest::test_eager_matches_fa2_generate",
17640
+ "count": 1
17641
+ },
17642
+ {
17643
+ "status": "passed",
17644
+ "test": "tests/models/mbart/test_modeling_mbart.py::MBartStandaloneDecoderModelTest::test_flash_attn_2_fp32_ln",
17645
+ "count": 1
17646
+ },
17647
+ {
17648
+ "status": "passed",
17649
+ "test": "tests/models/mbart/test_modeling_mbart.py::MBartStandaloneDecoderModelTest::test_flash_attn_2_from_config",
17650
+ "count": 1
17651
+ },
17652
+ {
17653
+ "status": "passed",
17654
+ "test": "tests/models/mbart/test_modeling_mbart.py::MBartStandaloneDecoderModelTest::test_flash_attn_2_inference_equivalence",
17655
+ "count": 1
17656
+ },
17657
+ {
17658
+ "status": "passed",
17659
+ "test": "tests/models/mbart/test_modeling_mbart.py::MBartStandaloneDecoderModelTest::test_flash_attn_2_inference_equivalence_right_padding",
17660
+ "count": 1
17661
+ },
17662
+ {
17663
+ "status": "skipped",
17664
+ "test": "[2] tests/generation/test_utils.py:1857: test requires Flash Attention 3",
17665
+ "count": 2
17666
+ },
17667
+ {
17668
+ "status": "skipped",
17669
+ "test": "[1] tests/generation/test_utils.py:1865: Model is an encoder-decoder",
17670
+ "count": 1
17671
+ },
17672
+ {
17673
+ "status": "skipped",
17674
+ "test": "[1] tests/generation/test_utils.py:2082: MBartForConditionalGeneration does not support new attention mask API",
17675
+ "count": 1
17676
+ },
17677
+ {
17678
+ "status": "skipped",
17679
+ "test": "[1] tests/generation/test_utils.py:2089: MBartForConditionalGeneration does not support new attention mask API",
17680
+ "count": 1
17681
+ },
17682
+ {
17683
+ "status": "skipped",
17684
+ "test": "[2] tests/generation/test_utils.py:2098: test requires Flash Attention 3",
17685
+ "count": 2
17686
+ },
17687
+ {
17688
+ "status": "skipped",
17689
+ "test": "[2] tests/generation/test_utils.py:2105: test requires Flash Attention 3",
17690
+ "count": 2
17691
+ },
17692
+ {
17693
+ "status": "skipped",
17694
+ "test": "[1] tests/test_modeling_common.py:3367: MBartModelTest doesn't have the attribute `_torch_compile_train_cls`.",
17695
+ "count": 1
17696
+ },
17697
+ {
17698
+ "status": "skipped",
17699
+ "test": "[2] tests/test_modeling_common.py:3294: This model is not a composite model!",
17700
+ "count": 2
17701
+ },
17702
+ {
17703
+ "status": "skipped",
17704
+ "test": "[2] tests/test_modeling_common.py:3300: test requires Flash Attention 3",
17705
+ "count": 2
17706
+ },
17707
+ {
17708
+ "status": "skipped",
17709
+ "test": "[2] tests/test_modeling_common.py:3469: test requires Flash Attention 3",
17710
+ "count": 2
17711
+ },
17712
+ {
17713
+ "status": "skipped",
17714
+ "test": "[2] tests/test_modeling_common.py:2910: test requires Flash Attention 3",
17715
+ "count": 2
17716
+ },
17717
+ {
17718
+ "status": "skipped",
17719
+ "test": "[2] tests/test_modeling_common.py:2918: test requires Flash Attention 3",
17720
+ "count": 2
17721
+ },
17722
+ {
17723
+ "status": "skipped",
17724
+ "test": "[2] tests/test_modeling_common.py:2876: test requires the kernels library",
17725
+ "count": 2
17726
+ },
17727
+ {
17728
+ "status": "skipped",
17729
+ "test": "[2] tests/test_modeling_common.py:2884: test requires MPS",
17730
+ "count": 2
17731
+ },
17732
+ {
17733
+ "status": "skipped",
17734
+ "test": "[1] tests/generation/test_utils.py:1865: Model does not support position_ids",
17735
+ "count": 1
17736
+ },
17737
+ {
17738
+ "status": "skipped",
17739
+ "test": "[1] tests/generation/test_utils.py:2082: MBartForCausalLM does not support new attention mask API",
17740
+ "count": 1
17741
+ },
17742
+ {
17743
+ "status": "skipped",
17744
+ "test": "[1] tests/generation/test_utils.py:2089: MBartForCausalLM does not support new attention mask API",
17745
+ "count": 1
17746
+ },
17747
+ {
17748
+ "status": "skipped",
17749
+ "test": "[1] tests/test_modeling_common.py:3367: MBartStandaloneDecoderModelTest doesn't have the attribute `_torch_compile_train_cls`.",
17750
+ "count": 1
17751
+ }
17752
+ ]
17753
+ },
17754
+ {
17755
+ "model": "megatron_bert",
17756
+ "results": [
17757
+ {
17758
+ "status": "passed",
17759
+ "test": "tests/models/megatron_bert/test_modeling_megatron_bert.py::MegatronBertModelTest::test_flash_attn_2_fp32_ln",
17760
+ "count": 1
17761
+ },
17762
+ {
17763
+ "status": "passed",
17764
+ "test": "tests/models/megatron_bert/test_modeling_megatron_bert.py::MegatronBertModelTest::test_flash_attn_2_from_config",
17765
+ "count": 1
17766
+ },
17767
+ {
17768
+ "status": "skipped",
17769
+ "test": "[1] tests/test_modeling_common.py:3367: MegatronBertModelTest doesn't have the attribute `_torch_compile_train_cls`.",
17770
+ "count": 1
17771
+ },
17772
+ {
17773
+ "status": "skipped",
17774
+ "test": "[1] tests/test_modeling_common.py:3294: This model is not a composite model!",
17775
+ "count": 1
17776
+ },
17777
+ {
17778
+ "status": "skipped",
17779
+ "test": "[1] tests/test_modeling_common.py:2894: Model architecture does not support flash_attention_2, or setting its attention dynamically",
17780
+ "count": 1
17781
+ },
17782
+ {
17783
+ "status": "skipped",
17784
+ "test": "[1] tests/test_modeling_common.py:2902: Model architecture does not support flash_attention_2, or setting its attention dynamically",
17785
+ "count": 1
17786
+ },
17787
+ {
17788
+ "status": "skipped",
17789
+ "test": "[1] tests/test_modeling_common.py:3300: test requires Flash Attention 3",
17790
+ "count": 1
17791
+ },
17792
+ {
17793
+ "status": "skipped",
17794
+ "test": "[1] tests/test_modeling_common.py:3469: test requires Flash Attention 3",
17795
+ "count": 1
17796
+ },
17797
+ {
17798
+ "status": "skipped",
17799
+ "test": "[1] tests/test_modeling_common.py:2910: test requires Flash Attention 3",
17800
+ "count": 1
17801
+ },
17802
+ {
17803
+ "status": "skipped",
17804
+ "test": "[1] tests/test_modeling_common.py:2918: test requires Flash Attention 3",
17805
+ "count": 1
17806
+ },
17807
+ {
17808
+ "status": "skipped",
17809
+ "test": "[1] tests/test_modeling_common.py:2876: test requires the kernels library",
17810
+ "count": 1
17811
+ },
17812
+ {
17813
+ "status": "skipped",
17814
+ "test": "[1] tests/test_modeling_common.py:2884: test requires MPS",
17815
+ "count": 1
17816
+ }
17817
+ ]
17818
+ },
17819
+ {
17820
+ "model": "megatron_gpt2",
17821
+ "results": []
17822
+ },
17823
+ {
17824
+ "model": "metaclip_2",
17825
+ "results": [
17826
+ {
17827
+ "status": "passed",
17828
+ "test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2VisionModelTest::test_flash_attn_2_fp32_ln",
17829
+ "count": 1
17830
+ },
17831
+ {
17832
+ "status": "passed",
17833
+ "test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2VisionModelTest::test_flash_attn_2_from_config",
17834
+ "count": 1
17835
+ },
17836
+ {
17837
+ "status": "passed",
17838
+ "test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2VisionModelTest::test_flash_attn_2_inference_equivalence",
17839
+ "count": 1
17840
+ },
17841
+ {
17842
+ "status": "passed",
17843
+ "test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2VisionModelTest::test_flash_attn_2_inference_equivalence_right_padding",
17844
+ "count": 1
17845
+ },
17846
+ {
17847
+ "status": "passed",
17848
+ "test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2TextModelTest::test_flash_attn_2_fp32_ln",
17849
+ "count": 1
17850
+ },
17851
+ {
17852
+ "status": "passed",
17853
+ "test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2TextModelTest::test_flash_attn_2_from_config",
17854
+ "count": 1
17855
+ },
17856
+ {
17857
+ "status": "passed",
17858
+ "test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2TextModelTest::test_flash_attn_2_inference_equivalence",
17859
+ "count": 1
17860
+ },
17861
+ {
17862
+ "status": "passed",
17863
+ "test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2TextModelTest::test_flash_attn_2_inference_equivalence_right_padding",
17864
+ "count": 1
17865
+ },
17866
+ {
17867
+ "status": "passed",
17868
+ "test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2ModelTest::test_flash_attn_2_can_dispatch_composite_models",
17869
+ "count": 1
17870
+ },
17871
+ {
17872
+ "status": "passed",
17873
+ "test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2ModelTest::test_flash_attn_2_fp32_ln",
17874
+ "count": 1
17875
+ },
17876
+ {
17877
+ "status": "passed",
17878
+ "test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2ModelTest::test_flash_attn_2_from_config",
17879
+ "count": 1
17880
+ },
17881
+ {
17882
+ "status": "passed",
17883
+ "test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2ModelTest::test_flash_attn_2_inference_equivalence",
17884
+ "count": 1
17885
+ },
17886
+ {
17887
+ "status": "passed",
17888
+ "test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2ModelTest::test_flash_attn_2_inference_equivalence_right_padding",
17889
+ "count": 1
17890
+ },
17891
+ {
17892
+ "status": "passed",
17893
+ "test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2ForImageClassificationModelTest::test_flash_attn_2_can_dispatch_composite_models",
17894
+ "count": 1
17895
+ },
17896
+ {
17897
+ "status": "passed",
17898
+ "test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2ForImageClassificationModelTest::test_flash_attn_2_fp32_ln",
17899
+ "count": 1
17900
+ },
17901
+ {
17902
+ "status": "passed",
17903
+ "test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2ForImageClassificationModelTest::test_flash_attn_2_from_config",
17904
+ "count": 1
17905
+ },
17906
+ {
17907
+ "status": "passed",
17908
+ "test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2ForImageClassificationModelTest::test_flash_attn_2_inference_equivalence",
17909
+ "count": 1
17910
+ },
17911
+ {
17912
+ "status": "passed",
17913
+ "test": "tests/models/metaclip_2/test_modeling_metaclip_2.py::MetaClip2ForImageClassificationModelTest::test_flash_attn_2_inference_equivalence_right_padding",
17914
+ "count": 1
17915
+ },
17916
+ {
17917
+ "status": "skipped",
17918
+ "test": "[1] tests/test_modeling_common.py:3367: MetaClip2VisionModelTest doesn't have the attribute `_torch_compile_train_cls`.",
17919
+ "count": 1
17920
+ },
17921
+ {
17922
+ "status": "skipped",
17923
+ "test": "[2] tests/test_modeling_common.py:3294: This model is not a composite model!",
17924
+ "count": 2
17925
+ },
17926
+ {
17927
+ "status": "skipped",
17928
+ "test": "[4] tests/test_modeling_common.py:3300: test requires Flash Attention 3",
17929
+ "count": 4
17930
+ },
17931
+ {
17932
+ "status": "skipped",
17933
+ "test": "[4] tests/test_modeling_common.py:3469: test requires Flash Attention 3",
17934
+ "count": 4
17935
+ },
17936
+ {
17937
+ "status": "skipped",
17938
+ "test": "[4] tests/test_modeling_common.py:2910: test requires Flash Attention 3",
17939
+ "count": 4
17940
+ },
17941
+ {
17942
+ "status": "skipped",
17943
+ "test": "[4] tests/test_modeling_common.py:2918: test requires Flash Attention 3",
17944
+ "count": 4
17945
+ },
17946
+ {
17947
+ "status": "skipped",
17948
+ "test": "[4] tests/test_modeling_common.py:2876: test requires the kernels library",
17949
+ "count": 4
17950
+ },
17951
+ {
17952
+ "status": "skipped",
17953
+ "test": "[4] tests/test_modeling_common.py:2884: test requires MPS",
17954
+ "count": 4
17955
+ },
17956
+ {
17957
+ "status": "skipped",
17958
+ "test": "[1] tests/test_modeling_common.py:3367: MetaClip2TextModelTest doesn't have the attribute `_torch_compile_train_cls`.",
17959
+ "count": 1
17960
+ },
17961
+ {
17962
+ "status": "skipped",
17963
+ "test": "[1] tests/test_modeling_common.py:3367: MetaClip2ModelTest doesn't have the attribute `_torch_compile_train_cls`.",
17964
+ "count": 1
17965
+ },
17966
+ {
17967
+ "status": "skipped",
17968
+ "test": "[1] tests/test_modeling_common.py:3367: MetaClip2ForImageClassificationModelTest doesn't have the attribute `_torch_compile_train_cls`.",
17969
+ "count": 1
17970
+ }
17971
+ ]
17972
+ },
17973
  {
17974
  "model": "mgp_str",
17975
  "results": [