Upload cuda_inference_transformers_image-classification_google/vit-base-patch16-224/benchmark.json with huggingface_hub
Browse files
cuda_inference_transformers_image-classification_google/vit-base-patch16-224/benchmark.json
CHANGED
|
@@ -6,19 +6,17 @@
|
|
| 6 |
"version": "2.3.0+cu121",
|
| 7 |
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
|
| 8 |
"task": "image-classification",
|
| 9 |
-
"model": "google/vit-base-patch16-224",
|
| 10 |
"library": "transformers",
|
|
|
|
|
|
|
| 11 |
"device": "cuda",
|
| 12 |
"device_ids": "0",
|
| 13 |
"seed": 42,
|
| 14 |
"inter_op_num_threads": null,
|
| 15 |
"intra_op_num_threads": null,
|
| 16 |
-
"
|
| 17 |
-
|
| 18 |
-
|
| 19 |
-
"local_files_only": false,
|
| 20 |
-
"trust_remote_code": false
|
| 21 |
-
},
|
| 22 |
"no_weights": true,
|
| 23 |
"device_map": null,
|
| 24 |
"torch_dtype": null,
|
|
@@ -104,7 +102,7 @@
|
|
| 104 |
"forward": {
|
| 105 |
"memory": {
|
| 106 |
"unit": "MB",
|
| 107 |
-
"max_ram": 911.
|
| 108 |
"max_global_vram": 1057.488896,
|
| 109 |
"max_process_vram": 0.0,
|
| 110 |
"max_reserved": 406.847488,
|
|
@@ -112,213 +110,196 @@
|
|
| 112 |
},
|
| 113 |
"latency": {
|
| 114 |
"unit": "s",
|
| 115 |
-
"count":
|
| 116 |
-
"total": 0.
|
| 117 |
-
"mean": 0.
|
| 118 |
-
"stdev": 0.
|
| 119 |
-
"p50": 0.
|
| 120 |
-
"p90": 0.
|
| 121 |
-
"p95": 0.
|
| 122 |
-
"p99": 0.
|
| 123 |
"values": [
|
| 124 |
-
0.
|
| 125 |
-
0.
|
| 126 |
-
0.
|
| 127 |
-
0.
|
| 128 |
-
0.
|
| 129 |
-
0.
|
| 130 |
-
0.
|
| 131 |
-
0.
|
| 132 |
-
0.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 133 |
0.00587775993347168,
|
| 134 |
-
0.
|
| 135 |
-
0.
|
| 136 |
-
0.
|
| 137 |
-
0.
|
| 138 |
-
0.
|
| 139 |
-
0.
|
| 140 |
-
0.
|
| 141 |
-
0.
|
| 142 |
-
0.
|
| 143 |
-
0.
|
| 144 |
-
0.
|
| 145 |
-
0.
|
| 146 |
-
0.
|
| 147 |
-
0.
|
| 148 |
-
0.
|
| 149 |
-
0.
|
| 150 |
-
0.
|
| 151 |
-
0.
|
| 152 |
-
0.
|
| 153 |
-
0.
|
| 154 |
-
0.
|
| 155 |
-
0.
|
| 156 |
-
0.
|
| 157 |
-
0.
|
| 158 |
-
0.
|
| 159 |
-
0.
|
| 160 |
-
0.
|
| 161 |
-
0.005412864208221436,
|
| 162 |
-
0.005410816192626953,
|
| 163 |
-
0.005405695915222168,
|
| 164 |
-
0.005392384052276611,
|
| 165 |
-
0.005399551868438721,
|
| 166 |
-
0.005431263923645019,
|
| 167 |
-
0.005380095958709717,
|
| 168 |
-
0.005403647899627686,
|
| 169 |
-
0.005449728012084961,
|
| 170 |
-
0.005573631763458252,
|
| 171 |
-
0.0054271998405456545,
|
| 172 |
-
0.005396480083465576,
|
| 173 |
-
0.005419007778167725,
|
| 174 |
-
0.005434368133544922,
|
| 175 |
-
0.0053975038528442385,
|
| 176 |
-
0.005400576114654541,
|
| 177 |
-
0.005408768177032471,
|
| 178 |
-
0.005400576114654541,
|
| 179 |
-
0.005430272102355957,
|
| 180 |
-
0.00550707197189331,
|
| 181 |
-
0.005395455837249756,
|
| 182 |
-
0.005391424179077149,
|
| 183 |
-
0.0054241280555725096,
|
| 184 |
-
0.005385216236114502,
|
| 185 |
-
0.005404672145843506,
|
| 186 |
-
0.005404640197753907,
|
| 187 |
-
0.005398528099060058,
|
| 188 |
-
0.005445631980895996,
|
| 189 |
-
0.005385216236114502,
|
| 190 |
-
0.0053944320678710935,
|
| 191 |
-
0.005416959762573242,
|
| 192 |
-
0.005390336036682129,
|
| 193 |
-
0.005396480083465576,
|
| 194 |
-
0.005421055793762207,
|
| 195 |
-
0.005427135944366455,
|
| 196 |
-
0.005398528099060058,
|
| 197 |
-
0.005404672145843506,
|
| 198 |
-
0.005386240005493164,
|
| 199 |
-
0.005379072189331055,
|
| 200 |
-
0.00536678409576416,
|
| 201 |
-
0.005391359806060791,
|
| 202 |
-
0.005410816192626953,
|
| 203 |
-
0.005378047943115235,
|
| 204 |
-
0.005401599884033203,
|
| 205 |
-
0.005406720161437988,
|
| 206 |
-
0.005390336036682129,
|
| 207 |
-
0.005411839962005615,
|
| 208 |
-
0.0054241280555725096,
|
| 209 |
-
0.005380095958709717,
|
| 210 |
-
0.0053647360801696775,
|
| 211 |
-
0.005394400119781494,
|
| 212 |
-
0.005380095958709717,
|
| 213 |
-
0.005390336036682129,
|
| 214 |
-
0.005349376201629639,
|
| 215 |
-
0.00536678409576416,
|
| 216 |
-
0.005384191989898681,
|
| 217 |
-
0.0053647360801696775,
|
| 218 |
-
0.0053637118339538575,
|
| 219 |
-
0.005389311790466309,
|
| 220 |
-
0.00535756778717041,
|
| 221 |
-
0.005421055793762207,
|
| 222 |
-
0.005517312049865723,
|
| 223 |
-
0.0054609918594360355,
|
| 224 |
-
0.0054568958282470706,
|
| 225 |
-
0.005498784065246582,
|
| 226 |
-
0.005447679996490478,
|
| 227 |
-
0.005450719833374023,
|
| 228 |
-
0.005430272102355957,
|
| 229 |
-
0.005447679996490478,
|
| 230 |
-
0.005477375984191894,
|
| 231 |
-
0.005453824043273926,
|
| 232 |
-
0.005453824043273926,
|
| 233 |
-
0.0054579200744628905,
|
| 234 |
-
0.005458943843841553,
|
| 235 |
-
0.0054568958282470706,
|
| 236 |
-
0.005436416149139404,
|
| 237 |
-
0.005450751781463623,
|
| 238 |
-
0.005453824043273926,
|
| 239 |
-
0.005429247856140137,
|
| 240 |
-
0.005453760147094726,
|
| 241 |
-
0.005489664077758789,
|
| 242 |
-
0.005451776027679443,
|
| 243 |
-
0.005413824081420898,
|
| 244 |
-
0.005452799797058106,
|
| 245 |
-
0.005454847812652588,
|
| 246 |
-
0.005462080001831055,
|
| 247 |
-
0.0056720318794250485,
|
| 248 |
-
0.005734399795532226,
|
| 249 |
0.005752831935882568,
|
| 250 |
-
0.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 251 |
0.005736447811126709,
|
| 252 |
-
0.
|
| 253 |
-
0.
|
| 254 |
-
0.
|
| 255 |
-
0.
|
| 256 |
-
0.005478400230407715,
|
| 257 |
-
0.005477375984191894,
|
| 258 |
-
0.005473279953002929,
|
| 259 |
-
0.005417984008789062,
|
| 260 |
-
0.00545692777633667,
|
| 261 |
-
0.005434368133544922,
|
| 262 |
-
0.005436416149139404,
|
| 263 |
-
0.005432320117950439,
|
| 264 |
-
0.005451776027679443,
|
| 265 |
-
0.005438464164733887,
|
| 266 |
-
0.005463039875030518,
|
| 267 |
-
0.005410816192626953,
|
| 268 |
-
0.005442560195922852,
|
| 269 |
-
0.0054609918594360355,
|
| 270 |
-
0.005446656227111817,
|
| 271 |
-
0.005448703765869141,
|
| 272 |
-
0.005458911895751953,
|
| 273 |
-
0.005431295871734619,
|
| 274 |
-
0.005455872058868408,
|
| 275 |
-
0.005447679996490478,
|
| 276 |
-
0.005454847812652588,
|
| 277 |
-
0.005458943843841553,
|
| 278 |
-
0.005419007778167725,
|
| 279 |
-
0.0054241280555725096,
|
| 280 |
-
0.005455872058868408,
|
| 281 |
-
0.005432320117950439,
|
| 282 |
-
0.005454847812652588,
|
| 283 |
-
0.005463103771209717,
|
| 284 |
-
0.005452799797058106,
|
| 285 |
-
0.005435391902923584,
|
| 286 |
-
0.005450751781463623,
|
| 287 |
-
0.005430272102355957,
|
| 288 |
-
0.005462016105651855,
|
| 289 |
-
0.005440512180328369,
|
| 290 |
-
0.005458943843841553,
|
| 291 |
-
0.005477375984191894,
|
| 292 |
-
0.005445631980895996,
|
| 293 |
-
0.005444608211517334,
|
| 294 |
-
0.005471231937408447,
|
| 295 |
-
0.005446656227111817,
|
| 296 |
-
0.005409791946411133,
|
| 297 |
-
0.005472256183624268,
|
| 298 |
-
0.005493760108947754,
|
| 299 |
-
0.005467135906219483,
|
| 300 |
-
0.005448703765869141,
|
| 301 |
-
0.005446656227111817,
|
| 302 |
-
0.00546611213684082,
|
| 303 |
-
0.005444608211517334,
|
| 304 |
-
0.0054733438491821286,
|
| 305 |
-
0.005432320117950439
|
| 306 |
]
|
| 307 |
},
|
| 308 |
"throughput": {
|
| 309 |
"unit": "samples/s",
|
| 310 |
-
"value":
|
| 311 |
},
|
| 312 |
"energy": {
|
| 313 |
"unit": "kWh",
|
| 314 |
-
"cpu": 6.
|
| 315 |
-
"ram": 3.
|
| 316 |
-
"gpu": 3.
|
| 317 |
-
"total": 4.
|
| 318 |
},
|
| 319 |
"efficiency": {
|
| 320 |
"unit": "samples/kWh",
|
| 321 |
-
"value":
|
| 322 |
}
|
| 323 |
}
|
| 324 |
}
|
|
|
|
| 6 |
"version": "2.3.0+cu121",
|
| 7 |
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
|
| 8 |
"task": "image-classification",
|
|
|
|
| 9 |
"library": "transformers",
|
| 10 |
+
"model": "google/vit-base-patch16-224",
|
| 11 |
+
"processor": "google/vit-base-patch16-224",
|
| 12 |
"device": "cuda",
|
| 13 |
"device_ids": "0",
|
| 14 |
"seed": 42,
|
| 15 |
"inter_op_num_threads": null,
|
| 16 |
"intra_op_num_threads": null,
|
| 17 |
+
"model_kwargs": {},
|
| 18 |
+
"processor_kwargs": {},
|
| 19 |
+
"hub_kwargs": {},
|
|
|
|
|
|
|
|
|
|
| 20 |
"no_weights": true,
|
| 21 |
"device_map": null,
|
| 22 |
"torch_dtype": null,
|
|
|
|
| 102 |
"forward": {
|
| 103 |
"memory": {
|
| 104 |
"unit": "MB",
|
| 105 |
+
"max_ram": 911.835136,
|
| 106 |
"max_global_vram": 1057.488896,
|
| 107 |
"max_process_vram": 0.0,
|
| 108 |
"max_reserved": 406.847488,
|
|
|
|
| 110 |
},
|
| 111 |
"latency": {
|
| 112 |
"unit": "s",
|
| 113 |
+
"count": 165,
|
| 114 |
+
"total": 0.9947032976150514,
|
| 115 |
+
"mean": 0.006028504834030614,
|
| 116 |
+
"stdev": 0.0002600844592717947,
|
| 117 |
+
"p50": 0.006110208034515381,
|
| 118 |
+
"p90": 0.006268928050994873,
|
| 119 |
+
"p95": 0.006333235263824463,
|
| 120 |
+
"p99": 0.00666234872817993,
|
| 121 |
"values": [
|
| 122 |
+
0.0067573761940002445,
|
| 123 |
+
0.006318079948425293,
|
| 124 |
+
0.006372384071350098,
|
| 125 |
+
0.006284287929534912,
|
| 126 |
+
0.006309887886047363,
|
| 127 |
+
0.006328320026397705,
|
| 128 |
+
0.006487040042877197,
|
| 129 |
+
0.006168575763702393,
|
| 130 |
+
0.006179776191711426,
|
| 131 |
+
0.00618390417098999,
|
| 132 |
+
0.00616755199432373,
|
| 133 |
+
0.006231008052825927,
|
| 134 |
+
0.006564864158630371,
|
| 135 |
+
0.0062259202003479,
|
| 136 |
+
0.00620851182937622,
|
| 137 |
+
0.006128608226776123,
|
| 138 |
+
0.006090752124786377,
|
| 139 |
+
0.006053887844085694,
|
| 140 |
+
0.006115327835083008,
|
| 141 |
+
0.006147071838378906,
|
| 142 |
+
0.006446080207824707,
|
| 143 |
+
0.006203392028808594,
|
| 144 |
+
0.006253568172454834,
|
| 145 |
+
0.006356031894683838,
|
| 146 |
+
0.006171648025512695,
|
| 147 |
+
0.006148096084594727,
|
| 148 |
+
0.006110208034515381,
|
| 149 |
+
0.006188032150268555,
|
| 150 |
+
0.006184959888458252,
|
| 151 |
+
0.006200384140014648,
|
| 152 |
+
0.0062566399574279785,
|
| 153 |
+
0.006241312026977539,
|
| 154 |
+
0.006227968215942382,
|
| 155 |
+
0.006160384178161621,
|
| 156 |
+
0.0061859841346740725,
|
| 157 |
+
0.006198272228240966,
|
| 158 |
+
0.0061972479820251464,
|
| 159 |
+
0.00617574405670166,
|
| 160 |
+
0.006189055919647217,
|
| 161 |
+
0.006220799922943115,
|
| 162 |
+
0.006137856006622314,
|
| 163 |
+
0.006087679862976075,
|
| 164 |
+
0.006096896171569824,
|
| 165 |
+
0.006143968105316162,
|
| 166 |
+
0.006072319984436035,
|
| 167 |
+
0.006113279819488526,
|
| 168 |
+
0.006144000053405762,
|
| 169 |
+
0.006105088233947754,
|
| 170 |
+
0.006164480209350586,
|
| 171 |
+
0.006237184047698975,
|
| 172 |
+
0.006198272228240966,
|
| 173 |
+
0.006334464073181153,
|
| 174 |
+
0.006281184196472168,
|
| 175 |
+
0.006165503978729248,
|
| 176 |
+
0.006173696041107178,
|
| 177 |
+
0.006157311916351318,
|
| 178 |
+
0.006319104194641113,
|
| 179 |
+
0.0062197761535644535,
|
| 180 |
+
0.006195199966430664,
|
| 181 |
+
0.006178815841674804,
|
| 182 |
+
0.006176767826080322,
|
| 183 |
+
0.006102015972137451,
|
| 184 |
+
0.00611737585067749,
|
| 185 |
+
0.006150144100189209,
|
| 186 |
+
0.006146048069000244,
|
| 187 |
+
0.006148096084594727,
|
| 188 |
+
0.006191103935241699,
|
| 189 |
+
0.0062494721412658695,
|
| 190 |
+
0.00607539176940918,
|
| 191 |
+
0.006100992202758789,
|
| 192 |
+
0.006168575763702393,
|
| 193 |
+
0.006124544143676758,
|
| 194 |
+
0.006132736206054688,
|
| 195 |
+
0.006120448112487793,
|
| 196 |
+
0.006144991874694824,
|
| 197 |
+
0.006277120113372803,
|
| 198 |
+
0.006139904022216797,
|
| 199 |
+
0.006164480209350586,
|
| 200 |
+
0.006277120113372803,
|
| 201 |
+
0.00623308801651001,
|
| 202 |
+
0.006245376110076905,
|
| 203 |
+
0.006244351863861084,
|
| 204 |
+
0.006168575763702393,
|
| 205 |
+
0.006160384178161621,
|
| 206 |
+
0.006160384178161621,
|
| 207 |
+
0.006111231803894043,
|
| 208 |
+
0.006076416015625,
|
| 209 |
+
0.006250495910644531,
|
| 210 |
+
0.006048736095428467,
|
| 211 |
+
0.006184000015258789,
|
| 212 |
+
0.005964799880981446,
|
| 213 |
+
0.005893119812011719,
|
| 214 |
+
0.006077439785003662,
|
| 215 |
+
0.0061562881469726565,
|
| 216 |
+
0.006110208034515381,
|
| 217 |
+
0.005825535774230957,
|
| 218 |
+
0.005806079864501953,
|
| 219 |
+
0.00594323205947876,
|
| 220 |
+
0.005863423824310303,
|
| 221 |
+
0.005913599967956543,
|
| 222 |
+
0.005911551952362061,
|
| 223 |
+
0.00591871976852417,
|
| 224 |
0.00587775993347168,
|
| 225 |
+
0.006001664161682129,
|
| 226 |
+
0.0060067839622497555,
|
| 227 |
+
0.006153215885162353,
|
| 228 |
+
0.005880832195281982,
|
| 229 |
+
0.005858304023742676,
|
| 230 |
+
0.005793791770935058,
|
| 231 |
+
0.005898240089416504,
|
| 232 |
+
0.005822463989257813,
|
| 233 |
+
0.005861408233642578,
|
| 234 |
+
0.005826560020446778,
|
| 235 |
+
0.0057794561386108395,
|
| 236 |
+
0.005715968132019043,
|
| 237 |
+
0.005728256225585937,
|
| 238 |
+
0.005720064163208007,
|
| 239 |
+
0.005763072013854981,
|
| 240 |
+
0.005740543842315674,
|
| 241 |
+
0.00576204776763916,
|
| 242 |
+
0.006071296215057373,
|
| 243 |
+
0.007379968166351319,
|
| 244 |
+
0.006608895778656006,
|
| 245 |
+
0.005900288105010986,
|
| 246 |
+
0.005836800098419189,
|
| 247 |
+
0.005860352039337159,
|
| 248 |
+
0.0058122239112854005,
|
| 249 |
+
0.005750783920288086,
|
| 250 |
+
0.0057118721008300784,
|
| 251 |
+
0.005697535991668701,
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 252 |
0.005752831935882568,
|
| 253 |
+
0.005695487976074219,
|
| 254 |
+
0.0056893758773803715,
|
| 255 |
+
0.005683199882507324,
|
| 256 |
+
0.00572822380065918,
|
| 257 |
+
0.00570470380783081,
|
| 258 |
+
0.005739520072937012,
|
| 259 |
+
0.005695456027984619,
|
| 260 |
+
0.005726208209991455,
|
| 261 |
+
0.005677055835723877,
|
| 262 |
+
0.005687295913696289,
|
| 263 |
+
0.005716991901397705,
|
| 264 |
+
0.00568012809753418,
|
| 265 |
+
0.005681151866912842,
|
| 266 |
+
0.0057190399169921875,
|
| 267 |
+
0.005692416191101074,
|
| 268 |
+
0.005690368175506591,
|
| 269 |
+
0.00567193603515625,
|
| 270 |
+
0.0057149438858032225,
|
| 271 |
+
0.005670911788940429,
|
| 272 |
+
0.005715968132019043,
|
| 273 |
+
0.005700607776641845,
|
| 274 |
+
0.005700607776641845,
|
| 275 |
+
0.005695487976074219,
|
| 276 |
+
0.005665791988372803,
|
| 277 |
+
0.005707808017730713,
|
| 278 |
+
0.005677055835723877,
|
| 279 |
+
0.005735424041748047,
|
| 280 |
+
0.005666816234588623,
|
| 281 |
+
0.005695487976074219,
|
| 282 |
0.005736447811126709,
|
| 283 |
+
0.005662720203399658,
|
| 284 |
+
0.00570470380783081,
|
| 285 |
+
0.00566374397277832,
|
| 286 |
+
0.005697535991668701
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 287 |
]
|
| 288 |
},
|
| 289 |
"throughput": {
|
| 290 |
"unit": "samples/s",
|
| 291 |
+
"value": 165.87860962722448
|
| 292 |
},
|
| 293 |
"energy": {
|
| 294 |
"unit": "kWh",
|
| 295 |
+
"cpu": 6.777853530550761e-08,
|
| 296 |
+
"ram": 3.705675496062862e-08,
|
| 297 |
+
"gpu": 3.4072566940571446e-07,
|
| 298 |
+
"total": 4.4556095967185066e-07
|
| 299 |
},
|
| 300 |
"efficiency": {
|
| 301 |
"unit": "samples/kWh",
|
| 302 |
+
"value": 2244361.8057032777
|
| 303 |
}
|
| 304 |
}
|
| 305 |
}
|