Upload cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base/benchmark.json with huggingface_hub
Browse files
cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base/benchmark.json
CHANGED
|
@@ -3,7 +3,7 @@
|
|
| 3 |
"name": "cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base",
|
| 4 |
"backend": {
|
| 5 |
"name": "pytorch",
|
| 6 |
-
"version": "2.
|
| 7 |
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
|
| 8 |
"task": "multiple-choice",
|
| 9 |
"library": "transformers",
|
|
@@ -103,7 +103,7 @@
|
|
| 103 |
"optimum_commit": null,
|
| 104 |
"timm_version": "1.0.9",
|
| 105 |
"timm_commit": null,
|
| 106 |
-
"peft_version":
|
| 107 |
"peft_commit": null
|
| 108 |
}
|
| 109 |
},
|
|
@@ -111,24 +111,24 @@
|
|
| 111 |
"load": {
|
| 112 |
"memory": {
|
| 113 |
"unit": "MB",
|
| 114 |
-
"max_ram":
|
| 115 |
"max_global_vram": 68702.69952,
|
| 116 |
-
"max_process_vram":
|
| 117 |
"max_reserved": 555.74528,
|
| 118 |
"max_allocated": 499.37152
|
| 119 |
},
|
| 120 |
"latency": {
|
| 121 |
"unit": "s",
|
| 122 |
"count": 1,
|
| 123 |
-
"total": 7.
|
| 124 |
-
"mean": 7.
|
| 125 |
"stdev": 0.0,
|
| 126 |
-
"p50": 7.
|
| 127 |
-
"p90": 7.
|
| 128 |
-
"p95": 7.
|
| 129 |
-
"p99": 7.
|
| 130 |
"values": [
|
| 131 |
-
7.
|
| 132 |
]
|
| 133 |
},
|
| 134 |
"throughput": null,
|
|
@@ -138,159 +138,159 @@
|
|
| 138 |
"forward": {
|
| 139 |
"memory": {
|
| 140 |
"unit": "MB",
|
| 141 |
-
"max_ram":
|
| 142 |
"max_global_vram": 68702.69952,
|
| 143 |
-
"max_process_vram":
|
| 144 |
"max_reserved": 555.74528,
|
| 145 |
"max_allocated": 499.5072
|
| 146 |
},
|
| 147 |
"latency": {
|
| 148 |
"unit": "s",
|
| 149 |
"count": 131,
|
| 150 |
-
"total":
|
| 151 |
-
"mean": 0.
|
| 152 |
-
"stdev": 0.
|
| 153 |
-
"p50": 0.
|
| 154 |
-
"p90": 0.
|
| 155 |
-
"p95": 0.
|
| 156 |
-
"p99": 0.
|
| 157 |
"values": [
|
| 158 |
-
0.
|
| 159 |
-
0.
|
| 160 |
-
0.
|
| 161 |
-
0.
|
| 162 |
-
0.
|
| 163 |
-
0.
|
| 164 |
-
0.
|
| 165 |
-
0.
|
| 166 |
-
0.
|
| 167 |
-
0.
|
| 168 |
-
0.
|
| 169 |
-
0.
|
| 170 |
-
0.
|
| 171 |
-
0.
|
| 172 |
-
0.
|
| 173 |
-
0.
|
| 174 |
-
0.
|
| 175 |
-
0.
|
| 176 |
-
0.
|
| 177 |
-
0.
|
| 178 |
-
0.
|
| 179 |
-
0.
|
| 180 |
-
0.
|
| 181 |
-
0.
|
| 182 |
-
0.
|
| 183 |
-
0.
|
| 184 |
-
0.
|
| 185 |
-
0.
|
| 186 |
-
0.
|
| 187 |
-
0.
|
| 188 |
-
0.
|
| 189 |
-
0.
|
| 190 |
-
0.
|
| 191 |
-
0.
|
| 192 |
-
0.
|
| 193 |
-
0.
|
| 194 |
-
0.
|
| 195 |
-
0.
|
| 196 |
-
0.
|
| 197 |
-
0.
|
| 198 |
-
0.
|
| 199 |
-
0.
|
| 200 |
-
0.
|
| 201 |
-
0.
|
| 202 |
-
0.
|
| 203 |
-
0.
|
| 204 |
-
0.
|
| 205 |
-
0.
|
| 206 |
-
0.
|
| 207 |
-
0.
|
| 208 |
-
0.
|
| 209 |
-
0.
|
| 210 |
-
0.
|
| 211 |
-
0.
|
| 212 |
-
0.
|
| 213 |
-
0.
|
| 214 |
-
0.
|
| 215 |
-
0.
|
| 216 |
-
0.
|
| 217 |
-
0.007944337844848633,
|
| 218 |
-
0.00747489595413208,
|
| 219 |
-
0.007304975986480713,
|
| 220 |
-
0.0072344160079956055,
|
| 221 |
0.007236495971679687,
|
| 222 |
-
0.
|
| 223 |
-
0.
|
| 224 |
-
0.
|
| 225 |
-
0.
|
| 226 |
-
0.
|
| 227 |
-
0.
|
| 228 |
-
0.
|
| 229 |
-
0.
|
| 230 |
-
0.
|
| 231 |
-
0.
|
| 232 |
-
0.
|
| 233 |
-
0.
|
| 234 |
-
0.
|
| 235 |
-
0.
|
| 236 |
-
0.
|
| 237 |
-
0.
|
| 238 |
-
0.
|
| 239 |
-
0.
|
| 240 |
-
0.
|
| 241 |
-
0.
|
| 242 |
-
0.
|
| 243 |
-
0.007273615837097168,
|
| 244 |
-
0.0072508959770202636,
|
| 245 |
-
0.007359055995941162,
|
| 246 |
-
0.007612977027893066,
|
| 247 |
-
0.007435697078704834,
|
| 248 |
-
0.007696656227111817,
|
| 249 |
-
0.007633296012878418,
|
| 250 |
-
0.007731216907501221,
|
| 251 |
-
0.0072632160186767575,
|
| 252 |
0.007249616146087647,
|
| 253 |
-
0.
|
| 254 |
-
0.
|
| 255 |
-
0.
|
| 256 |
-
0.
|
| 257 |
-
0.
|
| 258 |
-
0.
|
| 259 |
-
0.
|
| 260 |
-
0.
|
| 261 |
-
0.
|
| 262 |
-
0.
|
| 263 |
-
0.
|
| 264 |
-
0.
|
| 265 |
-
0.
|
| 266 |
-
0.
|
| 267 |
-
0.
|
| 268 |
-
0.
|
| 269 |
-
0.
|
| 270 |
-
0.
|
| 271 |
-
0.
|
| 272 |
-
0.
|
| 273 |
-
0.
|
| 274 |
-
0.
|
| 275 |
-
0.
|
| 276 |
-
0.
|
| 277 |
-
0.
|
| 278 |
-
0.
|
| 279 |
-
0.
|
| 280 |
-
0.
|
| 281 |
-
0.
|
| 282 |
-
0.
|
| 283 |
-
0.
|
| 284 |
-
0.
|
| 285 |
-
0.
|
| 286 |
-
0.
|
| 287 |
-
0.
|
| 288 |
-
0.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 289 |
]
|
| 290 |
},
|
| 291 |
"throughput": {
|
| 292 |
"unit": "samples/s",
|
| 293 |
-
"value":
|
| 294 |
},
|
| 295 |
"energy": null,
|
| 296 |
"efficiency": null
|
|
|
|
| 3 |
"name": "cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base",
|
| 4 |
"backend": {
|
| 5 |
"name": "pytorch",
|
| 6 |
+
"version": "2.3.1+rocm5.7",
|
| 7 |
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
|
| 8 |
"task": "multiple-choice",
|
| 9 |
"library": "transformers",
|
|
|
|
| 103 |
"optimum_commit": null,
|
| 104 |
"timm_version": "1.0.9",
|
| 105 |
"timm_commit": null,
|
| 106 |
+
"peft_version": null,
|
| 107 |
"peft_commit": null
|
| 108 |
}
|
| 109 |
},
|
|
|
|
| 111 |
"load": {
|
| 112 |
"memory": {
|
| 113 |
"unit": "MB",
|
| 114 |
+
"max_ram": 1031.028736,
|
| 115 |
"max_global_vram": 68702.69952,
|
| 116 |
+
"max_process_vram": 43248.467968,
|
| 117 |
"max_reserved": 555.74528,
|
| 118 |
"max_allocated": 499.37152
|
| 119 |
},
|
| 120 |
"latency": {
|
| 121 |
"unit": "s",
|
| 122 |
"count": 1,
|
| 123 |
+
"total": 7.70135302734375,
|
| 124 |
+
"mean": 7.70135302734375,
|
| 125 |
"stdev": 0.0,
|
| 126 |
+
"p50": 7.70135302734375,
|
| 127 |
+
"p90": 7.70135302734375,
|
| 128 |
+
"p95": 7.70135302734375,
|
| 129 |
+
"p99": 7.70135302734375,
|
| 130 |
"values": [
|
| 131 |
+
7.70135302734375
|
| 132 |
]
|
| 133 |
},
|
| 134 |
"throughput": null,
|
|
|
|
| 138 |
"forward": {
|
| 139 |
"memory": {
|
| 140 |
"unit": "MB",
|
| 141 |
+
"max_ram": 1148.33408,
|
| 142 |
"max_global_vram": 68702.69952,
|
| 143 |
+
"max_process_vram": 211202.048,
|
| 144 |
"max_reserved": 555.74528,
|
| 145 |
"max_allocated": 499.5072
|
| 146 |
},
|
| 147 |
"latency": {
|
| 148 |
"unit": "s",
|
| 149 |
"count": 131,
|
| 150 |
+
"total": 1.0011928222179418,
|
| 151 |
+
"mean": 0.007642693299373597,
|
| 152 |
+
"stdev": 0.0018883905585774827,
|
| 153 |
+
"p50": 0.007370256900787354,
|
| 154 |
+
"p90": 0.008073779106140136,
|
| 155 |
+
"p95": 0.008486579418182373,
|
| 156 |
+
"p99": 0.018086472034454286,
|
| 157 |
"values": [
|
| 158 |
+
0.008713299751281739,
|
| 159 |
+
0.008611379623413086,
|
| 160 |
+
0.008302099227905273,
|
| 161 |
+
0.008216338157653809,
|
| 162 |
+
0.008281139373779297,
|
| 163 |
+
0.00856369972229004,
|
| 164 |
+
0.008110418319702148,
|
| 165 |
+
0.008102898597717284,
|
| 166 |
+
0.008000337600708007,
|
| 167 |
+
0.007983538150787353,
|
| 168 |
+
0.007876177787780762,
|
| 169 |
+
0.007830577850341797,
|
| 170 |
+
0.014402112007141114,
|
| 171 |
+
0.007352497100830078,
|
| 172 |
+
0.007817937850952148,
|
| 173 |
+
0.00781073808670044,
|
| 174 |
+
0.00783201789855957,
|
| 175 |
+
0.007724017143249512,
|
| 176 |
+
0.007753617763519287,
|
| 177 |
+
0.00781777811050415,
|
| 178 |
+
0.007734736919403076,
|
| 179 |
+
0.007731537818908691,
|
| 180 |
+
0.007690257072448731,
|
| 181 |
+
0.011262745857238769,
|
| 182 |
+
0.008409459114074707,
|
| 183 |
+
0.00799793815612793,
|
| 184 |
+
0.007993138790130616,
|
| 185 |
+
0.007896657943725585,
|
| 186 |
+
0.007542256832122803,
|
| 187 |
+
0.007531857013702393,
|
| 188 |
+
0.007606417179107666,
|
| 189 |
+
0.007795217990875244,
|
| 190 |
+
0.007558577060699463,
|
| 191 |
+
0.007485456943511963,
|
| 192 |
+
0.007700977802276611,
|
| 193 |
+
0.007449616909027099,
|
| 194 |
+
0.007288176059722901,
|
| 195 |
+
0.007359857082366943,
|
| 196 |
+
0.007501136779785156,
|
| 197 |
+
0.007589137077331543,
|
| 198 |
+
0.007709617137908935,
|
| 199 |
+
0.007591856956481934,
|
| 200 |
+
0.007747858047485351,
|
| 201 |
+
0.007729776859283447,
|
| 202 |
+
0.007542097091674805,
|
| 203 |
+
0.007617776870727539,
|
| 204 |
+
0.007573936939239502,
|
| 205 |
+
0.00773809814453125,
|
| 206 |
+
0.007599536895751953,
|
| 207 |
+
0.007647377014160156,
|
| 208 |
+
0.007632018089294434,
|
| 209 |
+
0.007370256900787354,
|
| 210 |
+
0.007216815948486328,
|
| 211 |
+
0.007310737133026123,
|
| 212 |
+
0.0072612957954406735,
|
| 213 |
+
0.0072440171241760255,
|
| 214 |
+
0.007186895847320557,
|
| 215 |
+
0.007284336090087891,
|
| 216 |
+
0.007280655860900879,
|
|
|
|
|
|
|
|
|
|
|
|
|
| 217 |
0.007236495971679687,
|
| 218 |
+
0.007180816173553467,
|
| 219 |
+
0.00732209587097168,
|
| 220 |
+
0.0072742562294006346,
|
| 221 |
+
0.0072411360740661625,
|
| 222 |
+
0.007207376003265381,
|
| 223 |
+
0.020342605590820313,
|
| 224 |
+
0.0025214459896087645,
|
| 225 |
+
0.0043776102066040035,
|
| 226 |
+
0.007239696025848389,
|
| 227 |
+
0.007212976932525635,
|
| 228 |
+
0.007189136028289795,
|
| 229 |
+
0.007223375797271728,
|
| 230 |
+
0.007229297161102295,
|
| 231 |
+
0.007177775859832764,
|
| 232 |
+
0.007160016059875488,
|
| 233 |
+
0.007139697074890137,
|
| 234 |
+
0.007213295936584473,
|
| 235 |
+
0.007211535930633545,
|
| 236 |
+
0.00716785717010498,
|
| 237 |
+
0.00716977596282959,
|
| 238 |
+
0.007239376068115234,
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 239 |
0.007249616146087647,
|
| 240 |
+
0.0072276959419250485,
|
| 241 |
+
0.007265777111053467,
|
| 242 |
+
0.007277935981750488,
|
| 243 |
+
0.007252655982971192,
|
| 244 |
+
0.007018735885620117,
|
| 245 |
+
0.007043375968933106,
|
| 246 |
+
0.007052015781402588,
|
| 247 |
+
0.0070070562362670894,
|
| 248 |
+
0.0070428957939147946,
|
| 249 |
+
0.007088016033172608,
|
| 250 |
+
0.007200175762176513,
|
| 251 |
+
0.007226096153259277,
|
| 252 |
+
0.007044976234436035,
|
| 253 |
+
0.007054736137390136,
|
| 254 |
+
0.007033616065979004,
|
| 255 |
+
0.007064335823059082,
|
| 256 |
+
0.00720033597946167,
|
| 257 |
+
0.007015535831451416,
|
| 258 |
+
0.007015855789184571,
|
| 259 |
+
0.00803489875793457,
|
| 260 |
+
0.007971377849578857,
|
| 261 |
+
0.00798545789718628,
|
| 262 |
+
0.008073779106140136,
|
| 263 |
+
0.008064977645874024,
|
| 264 |
+
0.0076849770545959475,
|
| 265 |
+
0.0075641770362854,
|
| 266 |
+
0.007361135959625244,
|
| 267 |
+
0.008009617805480956,
|
| 268 |
+
0.019665483474731446,
|
| 269 |
+
0.0025139260292053223,
|
| 270 |
+
0.00343776798248291,
|
| 271 |
+
0.007988977909088134,
|
| 272 |
+
0.00798145818710327,
|
| 273 |
+
0.007908338069915772,
|
| 274 |
+
0.007409136772155761,
|
| 275 |
+
0.007478416919708252,
|
| 276 |
+
0.007379056930541993,
|
| 277 |
+
0.007392656803131103,
|
| 278 |
+
0.007257135868072509,
|
| 279 |
+
0.007009776115417481,
|
| 280 |
+
0.007010575771331787,
|
| 281 |
+
0.007044976234436035,
|
| 282 |
+
0.0073627371788024904,
|
| 283 |
+
0.007231056213378906,
|
| 284 |
+
0.007019375801086426,
|
| 285 |
+
0.007048495769500732,
|
| 286 |
+
0.00704753589630127,
|
| 287 |
+
0.0070523362159729,
|
| 288 |
+
0.007031374931335449
|
| 289 |
]
|
| 290 |
},
|
| 291 |
"throughput": {
|
| 292 |
"unit": "samples/s",
|
| 293 |
+
"value": 130.8439264574389
|
| 294 |
},
|
| 295 |
"energy": null,
|
| 296 |
"efficiency": null
|