Upload cuda_inference_transformers_image-classification_google/vit-base-patch16-224/benchmark.json with huggingface_hub
Browse files
cuda_inference_transformers_image-classification_google/vit-base-patch16-224/benchmark.json
CHANGED
|
@@ -6,19 +6,17 @@
|
|
| 6 |
"version": "2.2.2+rocm5.7",
|
| 7 |
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
|
| 8 |
"task": "image-classification",
|
| 9 |
-
"model": "google/vit-base-patch16-224",
|
| 10 |
"library": "transformers",
|
|
|
|
|
|
|
| 11 |
"device": "cuda",
|
| 12 |
"device_ids": "0",
|
| 13 |
"seed": 42,
|
| 14 |
"inter_op_num_threads": null,
|
| 15 |
"intra_op_num_threads": null,
|
| 16 |
-
"
|
| 17 |
-
|
| 18 |
-
|
| 19 |
-
"local_files_only": false,
|
| 20 |
-
"trust_remote_code": false
|
| 21 |
-
},
|
| 22 |
"no_weights": true,
|
| 23 |
"device_map": null,
|
| 24 |
"torch_dtype": null,
|
|
@@ -85,7 +83,7 @@
|
|
| 85 |
"gpu_count": 1,
|
| 86 |
"gpu_vram_mb": 68702699520,
|
| 87 |
"optimum_benchmark_version": "0.2.0",
|
| 88 |
-
"optimum_benchmark_commit": "
|
| 89 |
"transformers_version": "4.40.2",
|
| 90 |
"transformers_commit": null,
|
| 91 |
"accelerate_version": "0.30.1",
|
|
@@ -104,174 +102,184 @@
|
|
| 104 |
"forward": {
|
| 105 |
"memory": {
|
| 106 |
"unit": "MB",
|
| 107 |
-
"max_ram":
|
| 108 |
-
"max_global_vram": 778.
|
| 109 |
-
"max_process_vram":
|
| 110 |
"max_reserved": 406.847488,
|
| 111 |
"max_allocated": 355.912704
|
| 112 |
},
|
| 113 |
"latency": {
|
| 114 |
"unit": "s",
|
| 115 |
-
"count":
|
| 116 |
-
"total": 0.
|
| 117 |
-
"mean": 0.
|
| 118 |
-
"stdev": 0.
|
| 119 |
-
"p50": 0.
|
| 120 |
-
"p90": 0.
|
| 121 |
-
"p95": 0.
|
| 122 |
-
"p99": 0.
|
| 123 |
"values": [
|
| 124 |
-
0.
|
| 125 |
-
0.
|
| 126 |
-
0.
|
| 127 |
-
0.
|
| 128 |
-
0.
|
| 129 |
-
0.
|
| 130 |
-
0.
|
| 131 |
-
0.
|
| 132 |
-
0.
|
| 133 |
-
0.
|
| 134 |
-
0.
|
| 135 |
-
0.
|
| 136 |
-
0.
|
| 137 |
-
0.
|
| 138 |
-
0.
|
| 139 |
-
0.
|
| 140 |
-
0.
|
| 141 |
-
0.
|
| 142 |
-
0.
|
| 143 |
-
0.
|
| 144 |
-
0.
|
| 145 |
-
0.
|
| 146 |
-
0.
|
| 147 |
-
0.
|
| 148 |
-
0.
|
| 149 |
-
0.
|
| 150 |
-
0.
|
| 151 |
-
0.
|
| 152 |
-
0.
|
| 153 |
-
0.
|
| 154 |
-
0.
|
| 155 |
-
0.
|
| 156 |
-
0.
|
| 157 |
-
0.
|
| 158 |
-
0.
|
| 159 |
-
0.
|
| 160 |
-
0.
|
| 161 |
-
0.
|
| 162 |
-
0.
|
| 163 |
-
0.
|
| 164 |
-
0.
|
| 165 |
-
0.
|
| 166 |
-
0.
|
| 167 |
-
0.
|
| 168 |
-
0.
|
| 169 |
-
0.
|
| 170 |
-
0.
|
| 171 |
-
0.
|
| 172 |
-
0.
|
| 173 |
-
0.
|
| 174 |
-
0.
|
| 175 |
-
0.
|
| 176 |
-
0.
|
| 177 |
-
0.
|
| 178 |
-
0.
|
| 179 |
-
0.
|
| 180 |
-
0.
|
| 181 |
-
0.
|
| 182 |
-
0.
|
| 183 |
-
0.
|
| 184 |
-
0.
|
| 185 |
-
0.
|
| 186 |
-
0.
|
| 187 |
-
0.
|
| 188 |
-
0.
|
| 189 |
-
0.
|
| 190 |
-
0.
|
| 191 |
-
0.
|
| 192 |
-
0.
|
| 193 |
-
0.
|
| 194 |
-
0.
|
| 195 |
-
0.
|
| 196 |
-
0.
|
| 197 |
-
0.
|
| 198 |
-
0.
|
| 199 |
-
0.
|
| 200 |
-
0.
|
| 201 |
-
0.
|
| 202 |
-
0.
|
| 203 |
-
0.
|
| 204 |
-
0.
|
| 205 |
-
0.
|
| 206 |
-
0.
|
| 207 |
-
0.
|
| 208 |
-
0.
|
| 209 |
-
0.
|
| 210 |
-
0.
|
| 211 |
-
0.
|
| 212 |
-
0.
|
| 213 |
-
0.
|
| 214 |
-
0.
|
| 215 |
-
0.
|
| 216 |
-
0.
|
| 217 |
-
0.
|
| 218 |
-
0.
|
| 219 |
-
0.
|
| 220 |
-
0.
|
| 221 |
-
0.
|
| 222 |
-
0.
|
| 223 |
-
0.
|
| 224 |
-
0.
|
| 225 |
-
0.
|
| 226 |
-
0.
|
| 227 |
-
0.
|
| 228 |
-
0.
|
| 229 |
-
0.
|
| 230 |
-
0.
|
| 231 |
-
0.
|
| 232 |
-
0.
|
| 233 |
-
0.
|
| 234 |
-
0.
|
| 235 |
-
0.
|
| 236 |
-
0.
|
| 237 |
-
0.
|
| 238 |
-
0.
|
| 239 |
-
0.
|
| 240 |
-
0.
|
| 241 |
-
0.
|
| 242 |
-
0.
|
| 243 |
-
0.
|
| 244 |
-
0.
|
| 245 |
-
0.
|
| 246 |
-
0.
|
| 247 |
-
0.
|
| 248 |
-
0.
|
| 249 |
-
0.
|
| 250 |
-
0.
|
| 251 |
-
0.
|
| 252 |
-
0.
|
| 253 |
-
0.
|
| 254 |
-
0.
|
| 255 |
-
0.
|
| 256 |
-
0.
|
| 257 |
-
0.
|
| 258 |
-
0.
|
| 259 |
-
0.
|
| 260 |
-
0.
|
| 261 |
-
0.
|
| 262 |
-
0.
|
| 263 |
-
0.
|
| 264 |
-
0.
|
| 265 |
-
0.
|
| 266 |
-
0.
|
| 267 |
-
0.
|
| 268 |
-
0.
|
| 269 |
-
0.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 270 |
]
|
| 271 |
},
|
| 272 |
"throughput": {
|
| 273 |
"unit": "samples/s",
|
| 274 |
-
"value":
|
| 275 |
},
|
| 276 |
"energy": null,
|
| 277 |
"efficiency": null
|
|
|
|
| 6 |
"version": "2.2.2+rocm5.7",
|
| 7 |
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
|
| 8 |
"task": "image-classification",
|
|
|
|
| 9 |
"library": "transformers",
|
| 10 |
+
"model": "google/vit-base-patch16-224",
|
| 11 |
+
"processor": "google/vit-base-patch16-224",
|
| 12 |
"device": "cuda",
|
| 13 |
"device_ids": "0",
|
| 14 |
"seed": 42,
|
| 15 |
"inter_op_num_threads": null,
|
| 16 |
"intra_op_num_threads": null,
|
| 17 |
+
"model_kwargs": {},
|
| 18 |
+
"processor_kwargs": {},
|
| 19 |
+
"hub_kwargs": {},
|
|
|
|
|
|
|
|
|
|
| 20 |
"no_weights": true,
|
| 21 |
"device_map": null,
|
| 22 |
"torch_dtype": null,
|
|
|
|
| 83 |
"gpu_count": 1,
|
| 84 |
"gpu_vram_mb": 68702699520,
|
| 85 |
"optimum_benchmark_version": "0.2.0",
|
| 86 |
+
"optimum_benchmark_commit": "6fd377459e287bb09e9383ba2516b1b2a271a562",
|
| 87 |
"transformers_version": "4.40.2",
|
| 88 |
"transformers_commit": null,
|
| 89 |
"accelerate_version": "0.30.1",
|
|
|
|
| 102 |
"forward": {
|
| 103 |
"memory": {
|
| 104 |
"unit": "MB",
|
| 105 |
+
"max_ram": 1059.401728,
|
| 106 |
+
"max_global_vram": 778.4448,
|
| 107 |
+
"max_process_vram": 178601.218048,
|
| 108 |
"max_reserved": 406.847488,
|
| 109 |
"max_allocated": 355.912704
|
| 110 |
},
|
| 111 |
"latency": {
|
| 112 |
"unit": "s",
|
| 113 |
+
"count": 156,
|
| 114 |
+
"total": 0.9967217988967896,
|
| 115 |
+
"mean": 0.006389242300620446,
|
| 116 |
+
"stdev": 0.0002269903110995269,
|
| 117 |
+
"p50": 0.006375027894973755,
|
| 118 |
+
"p90": 0.00666902756690979,
|
| 119 |
+
"p95": 0.006712506890296936,
|
| 120 |
+
"p99": 0.006921547079086304,
|
| 121 |
"values": [
|
| 122 |
+
0.006487987995147705,
|
| 123 |
+
0.006144628047943115,
|
| 124 |
+
0.006301908016204834,
|
| 125 |
+
0.006480788230895996,
|
| 126 |
+
0.006539987087249756,
|
| 127 |
+
0.006543988227844238,
|
| 128 |
+
0.006409907817840576,
|
| 129 |
+
0.00637310791015625,
|
| 130 |
+
0.006362867832183838,
|
| 131 |
+
0.006312948226928711,
|
| 132 |
+
0.0062622280120849605,
|
| 133 |
+
0.006284788131713868,
|
| 134 |
+
0.006405588150024414,
|
| 135 |
+
0.006242387771606446,
|
| 136 |
+
0.00621374797821045,
|
| 137 |
+
0.006420467853546143,
|
| 138 |
+
0.0062169480323791505,
|
| 139 |
+
0.006146708965301514,
|
| 140 |
+
0.00617006778717041,
|
| 141 |
+
0.006126708030700684,
|
| 142 |
+
0.00610862922668457,
|
| 143 |
+
0.006160307884216309,
|
| 144 |
+
0.006203028202056884,
|
| 145 |
+
0.006392148017883301,
|
| 146 |
+
0.006138548851013183,
|
| 147 |
+
0.0061659088134765625,
|
| 148 |
+
0.006117588043212891,
|
| 149 |
+
0.006157429218292237,
|
| 150 |
+
0.006160307884216309,
|
| 151 |
+
0.006165108203887939,
|
| 152 |
+
0.006135509014129638,
|
| 153 |
+
0.0061411080360412595,
|
| 154 |
+
0.0061508679389953615,
|
| 155 |
+
0.006147028923034668,
|
| 156 |
+
0.006219828128814697,
|
| 157 |
+
0.00615150785446167,
|
| 158 |
+
0.007533425807952881,
|
| 159 |
+
0.0062908678054809574,
|
| 160 |
+
0.00607214879989624,
|
| 161 |
+
0.006024467945098877,
|
| 162 |
+
0.005968628883361816,
|
| 163 |
+
0.0060537481307983395,
|
| 164 |
+
0.006059349060058594,
|
| 165 |
+
0.006058228015899658,
|
| 166 |
+
0.0060894289016723635,
|
| 167 |
+
0.0060267081260681156,
|
| 168 |
+
0.006039828777313233,
|
| 169 |
+
0.006237748146057129,
|
| 170 |
+
0.006051188945770264,
|
| 171 |
+
0.006585426807403564,
|
| 172 |
+
0.00637694787979126,
|
| 173 |
+
0.00632830810546875,
|
| 174 |
+
0.006423987865447998,
|
| 175 |
+
0.006202548027038574,
|
| 176 |
+
0.006528947830200195,
|
| 177 |
+
0.006326228141784668,
|
| 178 |
+
0.006364627838134766,
|
| 179 |
+
0.006463508129119873,
|
| 180 |
+
0.006349267959594726,
|
| 181 |
+
0.006378868103027344,
|
| 182 |
+
0.006328467845916748,
|
| 183 |
+
0.0062887887954711915,
|
| 184 |
+
0.006424787044525146,
|
| 185 |
+
0.006704627990722656,
|
| 186 |
+
0.006701426982879639,
|
| 187 |
+
0.006709746837615967,
|
| 188 |
+
0.006441587924957275,
|
| 189 |
+
0.006646868228912354,
|
| 190 |
+
0.006469587802886963,
|
| 191 |
+
0.006326387882232666,
|
| 192 |
+
0.006480147838592529,
|
| 193 |
+
0.006402547836303711,
|
| 194 |
+
0.006433107852935791,
|
| 195 |
+
0.006504306793212891,
|
| 196 |
+
0.006323028087615967,
|
| 197 |
+
0.006415507793426514,
|
| 198 |
+
0.006360787868499756,
|
| 199 |
+
0.006282708168029785,
|
| 200 |
+
0.006461748123168945,
|
| 201 |
+
0.0062897481918334965,
|
| 202 |
+
0.006597908020019532,
|
| 203 |
+
0.006577746868133545,
|
| 204 |
+
0.006198069095611572,
|
| 205 |
+
0.0064236679077148435,
|
| 206 |
+
0.006254708766937256,
|
| 207 |
+
0.006333108901977539,
|
| 208 |
+
0.0063531088829040525,
|
| 209 |
+
0.006248788833618164,
|
| 210 |
+
0.00641838788986206,
|
| 211 |
+
0.006204148769378662,
|
| 212 |
+
0.0063575878143310545,
|
| 213 |
+
0.006289907932281494,
|
| 214 |
+
0.006268308162689209,
|
| 215 |
+
0.006348948001861573,
|
| 216 |
+
0.006222067832946777,
|
| 217 |
+
0.006443347930908203,
|
| 218 |
+
0.006433427810668945,
|
| 219 |
+
0.006334708213806152,
|
| 220 |
+
0.006338228225708007,
|
| 221 |
+
0.006522867202758789,
|
| 222 |
+
0.0064255881309509275,
|
| 223 |
+
0.006353588104248047,
|
| 224 |
+
0.006295827865600586,
|
| 225 |
+
0.006342708110809326,
|
| 226 |
+
0.006133588790893554,
|
| 227 |
+
0.006035668849945068,
|
| 228 |
+
0.005979989051818848,
|
| 229 |
+
0.006040627956390381,
|
| 230 |
+
0.006198228836059571,
|
| 231 |
+
0.0064159870147705075,
|
| 232 |
+
0.006248788833618164,
|
| 233 |
+
0.006819507122039795,
|
| 234 |
+
0.006642066955566407,
|
| 235 |
+
0.006388467788696289,
|
| 236 |
+
0.006188148021697998,
|
| 237 |
+
0.0063662281036376955,
|
| 238 |
+
0.0063545479774475095,
|
| 239 |
+
0.006416627883911133,
|
| 240 |
+
0.006018229007720947,
|
| 241 |
+
0.006416468143463134,
|
| 242 |
+
0.006720787048339844,
|
| 243 |
+
0.006770226955413818,
|
| 244 |
+
0.0065729479789733885,
|
| 245 |
+
0.006641107082366943,
|
| 246 |
+
0.006684947013854981,
|
| 247 |
+
0.006678708076477051,
|
| 248 |
+
0.006663826942443848,
|
| 249 |
+
0.006528148174285889,
|
| 250 |
+
0.006525746822357178,
|
| 251 |
+
0.006674228191375732,
|
| 252 |
+
0.006690227031707764,
|
| 253 |
+
0.006659667015075684,
|
| 254 |
+
0.006698708057403565,
|
| 255 |
+
0.006945746898651123,
|
| 256 |
+
0.006653427124023437,
|
| 257 |
+
0.006901747226715088,
|
| 258 |
+
0.006805747032165527,
|
| 259 |
+
0.006789587020874024,
|
| 260 |
+
0.006595828056335449,
|
| 261 |
+
0.006546707153320313,
|
| 262 |
+
0.006594548225402832,
|
| 263 |
+
0.006532948017120361,
|
| 264 |
+
0.006557747840881347,
|
| 265 |
+
0.00654942798614502,
|
| 266 |
+
0.006547506809234619,
|
| 267 |
+
0.006573748111724853,
|
| 268 |
+
0.006563346862792968,
|
| 269 |
+
0.0065539078712463375,
|
| 270 |
+
0.006536948204040527,
|
| 271 |
+
0.0065267071723937985,
|
| 272 |
+
0.006535828113555908,
|
| 273 |
+
0.006539507865905762,
|
| 274 |
+
0.006518065929412842,
|
| 275 |
+
0.006527667999267578,
|
| 276 |
+
0.006541428089141846,
|
| 277 |
+
0.0065315070152282715
|
| 278 |
]
|
| 279 |
},
|
| 280 |
"throughput": {
|
| 281 |
"unit": "samples/s",
|
| 282 |
+
"value": 156.513081355968
|
| 283 |
},
|
| 284 |
"energy": null,
|
| 285 |
"efficiency": null
|