Upload cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base/benchmark.json with huggingface_hub
Browse files
cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base/benchmark.json
CHANGED
|
@@ -3,7 +3,7 @@
|
|
| 3 |
"name": "cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base",
|
| 4 |
"backend": {
|
| 5 |
"name": "pytorch",
|
| 6 |
-
"version": "2.
|
| 7 |
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
|
| 8 |
"task": "multiple-choice",
|
| 9 |
"library": "transformers",
|
|
@@ -65,7 +65,7 @@
|
|
| 65 |
"name": "process",
|
| 66 |
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
|
| 67 |
"device_isolation": true,
|
| 68 |
-
"device_isolation_action": "
|
| 69 |
"numactl": false,
|
| 70 |
"numactl_kwargs": {},
|
| 71 |
"start_method": "spawn"
|
|
@@ -95,7 +95,7 @@
|
|
| 95 |
"optimum_benchmark_commit": null,
|
| 96 |
"transformers_version": "4.44.2",
|
| 97 |
"transformers_commit": null,
|
| 98 |
-
"accelerate_version": "0.
|
| 99 |
"accelerate_commit": null,
|
| 100 |
"diffusers_version": "0.30.2",
|
| 101 |
"diffusers_commit": null,
|
|
@@ -111,24 +111,24 @@
|
|
| 111 |
"load": {
|
| 112 |
"memory": {
|
| 113 |
"unit": "MB",
|
| 114 |
-
"max_ram":
|
| 115 |
-
"max_global_vram":
|
| 116 |
-
"max_process_vram":
|
| 117 |
"max_reserved": 555.74528,
|
| 118 |
"max_allocated": 499.37152
|
| 119 |
},
|
| 120 |
"latency": {
|
| 121 |
"unit": "s",
|
| 122 |
"count": 1,
|
| 123 |
-
"total": 7.
|
| 124 |
-
"mean": 7.
|
| 125 |
"stdev": 0.0,
|
| 126 |
-
"p50": 7.
|
| 127 |
-
"p90": 7.
|
| 128 |
-
"p95": 7.
|
| 129 |
-
"p99": 7.
|
| 130 |
"values": [
|
| 131 |
-
7.
|
| 132 |
]
|
| 133 |
},
|
| 134 |
"throughput": null,
|
|
@@ -138,150 +138,152 @@
|
|
| 138 |
"forward": {
|
| 139 |
"memory": {
|
| 140 |
"unit": "MB",
|
| 141 |
-
"max_ram":
|
| 142 |
-
"max_global_vram":
|
| 143 |
-
"max_process_vram":
|
| 144 |
"max_reserved": 555.74528,
|
| 145 |
"max_allocated": 499.5072
|
| 146 |
},
|
| 147 |
"latency": {
|
| 148 |
"unit": "s",
|
| 149 |
-
"count":
|
| 150 |
-
"total":
|
| 151 |
-
"mean": 0.
|
| 152 |
-
"stdev": 0.
|
| 153 |
-
"p50": 0.
|
| 154 |
-
"p90": 0.
|
| 155 |
-
"p95": 0.
|
| 156 |
-
"p99": 0.
|
| 157 |
"values": [
|
| 158 |
-
0.
|
| 159 |
-
0.
|
| 160 |
-
0.
|
| 161 |
-
0.
|
| 162 |
-
0.
|
| 163 |
-
0.
|
| 164 |
-
0.
|
| 165 |
-
0.
|
| 166 |
-
0.
|
| 167 |
-
0.
|
| 168 |
-
0.
|
| 169 |
-
0.
|
| 170 |
-
0.
|
| 171 |
-
0.
|
| 172 |
-
0.
|
| 173 |
-
0.
|
| 174 |
-
0.
|
| 175 |
-
0.
|
| 176 |
-
0.
|
| 177 |
-
0.
|
| 178 |
-
0.
|
| 179 |
-
0.
|
| 180 |
-
0.
|
| 181 |
-
0.
|
| 182 |
-
0.
|
| 183 |
-
0.
|
| 184 |
-
0.
|
| 185 |
-
0.
|
| 186 |
-
0.
|
| 187 |
-
0.
|
| 188 |
-
0.
|
| 189 |
-
0.
|
| 190 |
-
0.
|
| 191 |
-
0.
|
| 192 |
-
0.
|
| 193 |
-
0.
|
| 194 |
-
0.
|
| 195 |
-
0.
|
| 196 |
-
0.
|
| 197 |
-
0.
|
| 198 |
-
0.
|
| 199 |
-
0.
|
| 200 |
-
0.
|
| 201 |
-
0.
|
| 202 |
-
0.
|
| 203 |
-
0.
|
| 204 |
-
0.
|
| 205 |
-
0.
|
| 206 |
-
0.
|
| 207 |
-
0.
|
| 208 |
-
0.
|
| 209 |
-
0.
|
| 210 |
-
0.
|
| 211 |
-
0.
|
| 212 |
-
0.
|
| 213 |
-
0.
|
| 214 |
-
0.
|
| 215 |
-
0.
|
| 216 |
-
0.
|
| 217 |
-
0.
|
| 218 |
-
0.
|
| 219 |
-
0.
|
| 220 |
-
0.
|
| 221 |
-
0.
|
| 222 |
-
0.
|
| 223 |
-
0.
|
| 224 |
-
0.
|
| 225 |
-
0.
|
| 226 |
-
0.
|
| 227 |
-
0.
|
| 228 |
-
0.
|
| 229 |
-
0.
|
| 230 |
-
0.
|
| 231 |
-
0.
|
| 232 |
-
0.
|
| 233 |
-
0.
|
| 234 |
-
0.
|
| 235 |
-
0.
|
| 236 |
-
0.
|
| 237 |
-
0.
|
| 238 |
-
0.
|
| 239 |
-
0.
|
| 240 |
-
0.
|
| 241 |
-
0.
|
| 242 |
-
0.
|
| 243 |
-
0.
|
| 244 |
-
0.
|
| 245 |
-
0.
|
| 246 |
-
0.
|
| 247 |
-
0.
|
| 248 |
-
0.
|
| 249 |
-
0.
|
| 250 |
-
0.
|
| 251 |
-
0.
|
| 252 |
-
0.
|
| 253 |
-
0.
|
| 254 |
-
0.
|
| 255 |
-
0.
|
| 256 |
-
0.
|
| 257 |
-
0.
|
| 258 |
-
0.
|
| 259 |
-
0.
|
| 260 |
-
0.
|
| 261 |
-
0.
|
| 262 |
-
0.
|
| 263 |
-
0.
|
| 264 |
-
0.
|
| 265 |
-
0.
|
| 266 |
-
0.
|
| 267 |
-
0.
|
| 268 |
-
0.
|
| 269 |
-
0.
|
| 270 |
-
0.
|
| 271 |
-
0.
|
| 272 |
-
0.
|
| 273 |
-
0.
|
| 274 |
-
0.
|
| 275 |
-
0.
|
| 276 |
-
0.
|
| 277 |
-
0.
|
| 278 |
-
0.
|
| 279 |
-
0.
|
|
|
|
|
|
|
| 280 |
]
|
| 281 |
},
|
| 282 |
"throughput": {
|
| 283 |
"unit": "samples/s",
|
| 284 |
-
"value":
|
| 285 |
},
|
| 286 |
"energy": null,
|
| 287 |
"efficiency": null
|
|
|
|
| 3 |
"name": "cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base",
|
| 4 |
"backend": {
|
| 5 |
"name": "pytorch",
|
| 6 |
+
"version": "2.3.1+rocm5.7",
|
| 7 |
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
|
| 8 |
"task": "multiple-choice",
|
| 9 |
"library": "transformers",
|
|
|
|
| 65 |
"name": "process",
|
| 66 |
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
|
| 67 |
"device_isolation": true,
|
| 68 |
+
"device_isolation_action": "warn",
|
| 69 |
"numactl": false,
|
| 70 |
"numactl_kwargs": {},
|
| 71 |
"start_method": "spawn"
|
|
|
|
| 95 |
"optimum_benchmark_commit": null,
|
| 96 |
"transformers_version": "4.44.2",
|
| 97 |
"transformers_commit": null,
|
| 98 |
+
"accelerate_version": "0.34.0",
|
| 99 |
"accelerate_commit": null,
|
| 100 |
"diffusers_version": "0.30.2",
|
| 101 |
"diffusers_commit": null,
|
|
|
|
| 111 |
"load": {
|
| 112 |
"memory": {
|
| 113 |
"unit": "MB",
|
| 114 |
+
"max_ram": 1031.069696,
|
| 115 |
+
"max_global_vram": 68702.69952,
|
| 116 |
+
"max_process_vram": 44370.382848,
|
| 117 |
"max_reserved": 555.74528,
|
| 118 |
"max_allocated": 499.37152
|
| 119 |
},
|
| 120 |
"latency": {
|
| 121 |
"unit": "s",
|
| 122 |
"count": 1,
|
| 123 |
+
"total": 7.6900244140625,
|
| 124 |
+
"mean": 7.6900244140625,
|
| 125 |
"stdev": 0.0,
|
| 126 |
+
"p50": 7.6900244140625,
|
| 127 |
+
"p90": 7.6900244140625,
|
| 128 |
+
"p95": 7.6900244140625,
|
| 129 |
+
"p99": 7.6900244140625,
|
| 130 |
"values": [
|
| 131 |
+
7.6900244140625
|
| 132 |
]
|
| 133 |
},
|
| 134 |
"throughput": null,
|
|
|
|
| 138 |
"forward": {
|
| 139 |
"memory": {
|
| 140 |
"unit": "MB",
|
| 141 |
+
"max_ram": 1148.645376,
|
| 142 |
+
"max_global_vram": 68702.69952,
|
| 143 |
+
"max_process_vram": 218992.996352,
|
| 144 |
"max_reserved": 555.74528,
|
| 145 |
"max_allocated": 499.5072
|
| 146 |
},
|
| 147 |
"latency": {
|
| 148 |
"unit": "s",
|
| 149 |
+
"count": 124,
|
| 150 |
+
"total": 1.000955594778061,
|
| 151 |
+
"mean": 0.00807222253853275,
|
| 152 |
+
"stdev": 0.003566930112279063,
|
| 153 |
+
"p50": 0.007740370035171509,
|
| 154 |
+
"p90": 0.008459327030181884,
|
| 155 |
+
"p95": 0.008549190759658813,
|
| 156 |
+
"p99": 0.016620282993316644,
|
| 157 |
"values": [
|
| 158 |
+
0.00851772689819336,
|
| 159 |
+
0.008453726768493652,
|
| 160 |
+
0.015384099960327148,
|
| 161 |
+
0.003182547092437744,
|
| 162 |
+
0.008228768348693847,
|
| 163 |
+
0.008207326889038085,
|
| 164 |
+
0.007884768962860108,
|
| 165 |
+
0.0077406101226806644,
|
| 166 |
+
0.00764093017578125,
|
| 167 |
+
0.007619329929351807,
|
| 168 |
+
0.007591169834136963,
|
| 169 |
+
0.007517889976501465,
|
| 170 |
+
0.007517570972442627,
|
| 171 |
+
0.007464611053466797,
|
| 172 |
+
0.007456450939178467,
|
| 173 |
+
0.007446209907531739,
|
| 174 |
+
0.0074393310546875,
|
| 175 |
+
0.007457571029663086,
|
| 176 |
+
0.007472290992736817,
|
| 177 |
+
0.0074210910797119145,
|
| 178 |
+
0.007447010993957519,
|
| 179 |
+
0.007367650985717774,
|
| 180 |
+
0.00742733097076416,
|
| 181 |
+
0.007370370864868164,
|
| 182 |
+
0.01160987377166748,
|
| 183 |
+
0.008309087753295898,
|
| 184 |
+
0.008166208267211914,
|
| 185 |
+
0.008127009391784668,
|
| 186 |
+
0.00817404842376709,
|
| 187 |
+
0.008165727615356446,
|
| 188 |
+
0.008128607749938965,
|
| 189 |
+
0.008145407676696777,
|
| 190 |
+
0.008169407844543458,
|
| 191 |
+
0.008159808158874512,
|
| 192 |
+
0.00819644832611084,
|
| 193 |
+
0.0081430082321167,
|
| 194 |
+
0.008157407760620117,
|
| 195 |
+
0.008161249160766602,
|
| 196 |
+
0.008207167625427246,
|
| 197 |
+
0.010982996940612793,
|
| 198 |
+
0.008545247077941895,
|
| 199 |
+
0.008217887878417969,
|
| 200 |
+
0.008247008323669433,
|
| 201 |
+
0.016989532470703124,
|
| 202 |
+
0.0061417360305786135,
|
| 203 |
+
0.00880556583404541,
|
| 204 |
+
0.008549886703491211,
|
| 205 |
+
0.008483806610107421,
|
| 206 |
+
0.008439167022705078,
|
| 207 |
+
0.008466366767883302,
|
| 208 |
+
0.00851628589630127,
|
| 209 |
+
0.008461727142333984,
|
| 210 |
+
0.008438687324523925,
|
| 211 |
+
0.008299487113952637,
|
| 212 |
+
0.008359807014465332,
|
| 213 |
+
0.008265727996826172,
|
| 214 |
+
0.008273567199707031,
|
| 215 |
+
0.008260767936706543,
|
| 216 |
+
0.008260287284851075,
|
| 217 |
+
0.008246527671813966,
|
| 218 |
+
0.007683810234069824,
|
| 219 |
+
0.0076154909133911135,
|
| 220 |
+
0.00746605110168457,
|
| 221 |
+
0.007571809768676758,
|
| 222 |
+
0.007495650768280029,
|
| 223 |
+
0.007429891109466553,
|
| 224 |
+
0.007495649814605713,
|
| 225 |
+
0.007470530986785889,
|
| 226 |
+
0.0075154900550842285,
|
| 227 |
+
0.007436770915985108,
|
| 228 |
+
0.007534369945526123,
|
| 229 |
+
0.007440451145172119,
|
| 230 |
+
0.007617889881134033,
|
| 231 |
+
0.007469730854034424,
|
| 232 |
+
0.007565730094909668,
|
| 233 |
+
0.007740129947662353,
|
| 234 |
+
0.007682370185852051,
|
| 235 |
+
0.007647329807281494,
|
| 236 |
+
0.00757213020324707,
|
| 237 |
+
0.007619649887084961,
|
| 238 |
+
0.007486851215362549,
|
| 239 |
+
0.007480610847473144,
|
| 240 |
+
0.00749229097366333,
|
| 241 |
+
0.007492930889129639,
|
| 242 |
+
0.007509089946746826,
|
| 243 |
+
0.0074572510719299314,
|
| 244 |
+
0.007632929801940918,
|
| 245 |
+
0.007506050109863281,
|
| 246 |
+
0.007494690895080566,
|
| 247 |
+
0.00789436912536621,
|
| 248 |
+
0.008231488227844239,
|
| 249 |
+
0.007648449897766113,
|
| 250 |
+
0.007699170112609863,
|
| 251 |
+
0.007676770210266113,
|
| 252 |
+
0.042911670684814454,
|
| 253 |
+
0.002612468957901001,
|
| 254 |
+
0.0024998300075531007,
|
| 255 |
+
0.002457911014556885,
|
| 256 |
+
0.0024388699531555174,
|
| 257 |
+
0.00244783091545105,
|
| 258 |
+
0.005583178043365478,
|
| 259 |
+
0.008260766983032226,
|
| 260 |
+
0.008216927528381348,
|
| 261 |
+
0.007931328773498536,
|
| 262 |
+
0.008147968292236327,
|
| 263 |
+
0.008182687759399414,
|
| 264 |
+
0.00820268726348877,
|
| 265 |
+
0.008158687591552735,
|
| 266 |
+
0.00816316795349121,
|
| 267 |
+
0.008184927940368653,
|
| 268 |
+
0.008196767807006835,
|
| 269 |
+
0.008193086624145508,
|
| 270 |
+
0.00818988800048828,
|
| 271 |
+
0.00819388771057129,
|
| 272 |
+
0.00815548801422119,
|
| 273 |
+
0.00819996738433838,
|
| 274 |
+
0.008195648193359375,
|
| 275 |
+
0.008007168769836426,
|
| 276 |
+
0.007720929145812988,
|
| 277 |
+
0.007697249889373779,
|
| 278 |
+
0.00771357011795044,
|
| 279 |
+
0.007247490882873535,
|
| 280 |
+
0.007275651931762695,
|
| 281 |
+
0.007265091896057129
|
| 282 |
]
|
| 283 |
},
|
| 284 |
"throughput": {
|
| 285 |
"unit": "samples/s",
|
| 286 |
+
"value": 123.88161937143092
|
| 287 |
},
|
| 288 |
"energy": null,
|
| 289 |
"efficiency": null
|