hf-transformers-bot commited on
Commit
e25a4ba
·
verified ·
1 Parent(s): 4939b88

Upload 2025-11-20/ci_results_run_models_gpu/collated_reports_multi-gpu_453a246.json with huggingface_hub

Browse files
2025-11-20/ci_results_run_models_gpu/collated_reports_multi-gpu_453a246.json CHANGED
@@ -5,7 +5,7 @@
5
  "total_status_count": {
6
  "passed": 1657,
7
  "failed": 158,
8
- "skipped": 5434,
9
  "error": 0,
10
  "null": 0
11
  },
@@ -32983,6 +32983,67 @@
32983
  "count": 1
32984
  }
32985
  ]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
32986
  }
32987
  ]
32988
  }
 
5
  "total_status_count": {
6
  "passed": 1657,
7
  "failed": 158,
8
+ "skipped": 5438,
9
  "error": 0,
10
  "null": 0
11
  },
 
32983
  "count": 1
32984
  }
32985
  ]
32986
+ },
32987
+ {
32988
+ "model": "gpu_optimization",
32989
+ "results": []
32990
+ },
32991
+ {
32992
+ "model": "gpu_peft_integration",
32993
+ "results": []
32994
+ },
32995
+ {
32996
+ "model": "gpu_pipelines",
32997
+ "results": []
32998
+ },
32999
+ {
33000
+ "model": "gpu_quantization",
33001
+ "results": [
33002
+ {
33003
+ "status": "skipped",
33004
+ "test": "[1] tests/quantization/autoawq/test_awq.py:450: test requires autoawq",
33005
+ "count": 1
33006
+ },
33007
+ {
33008
+ "status": "skipped",
33009
+ "test": "[1] tests/quantization/autoawq/test_awq.py:368: test requires autoawq",
33010
+ "count": 1
33011
+ },
33012
+ {
33013
+ "status": "skipped",
33014
+ "test": "[1] tests/quantization/autoawq/test_awq.py:397: test requires autoawq",
33015
+ "count": 1
33016
+ },
33017
+ {
33018
+ "status": "skipped",
33019
+ "test": "[1] tests/quantization/autoawq/test_awq.py:492: test requires autoawq",
33020
+ "count": 1
33021
+ }
33022
+ ]
33023
+ },
33024
+ {
33025
+ "model": "gpu_repo_utils",
33026
+ "results": []
33027
+ },
33028
+ {
33029
+ "model": "gpu_sagemaker",
33030
+ "results": []
33031
+ },
33032
+ {
33033
+ "model": "gpu_tensor_parallel",
33034
+ "results": []
33035
+ },
33036
+ {
33037
+ "model": "gpu_tokenization",
33038
+ "results": []
33039
+ },
33040
+ {
33041
+ "model": "gpu_trainer",
33042
+ "results": []
33043
+ },
33044
+ {
33045
+ "model": "gpu_utils",
33046
+ "results": []
33047
  }
33048
  ]
33049
  }