Upload cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base/benchmark.json with huggingface_hub
Browse files
cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base/benchmark.json
CHANGED
|
@@ -3,7 +3,7 @@
|
|
| 3 |
"name": "cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base",
|
| 4 |
"backend": {
|
| 5 |
"name": "pytorch",
|
| 6 |
-
"version": "2.
|
| 7 |
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
|
| 8 |
"task": "multiple-choice",
|
| 9 |
"library": "transformers",
|
|
@@ -104,24 +104,24 @@
|
|
| 104 |
"load": {
|
| 105 |
"memory": {
|
| 106 |
"unit": "MB",
|
| 107 |
-
"max_ram":
|
| 108 |
-
"max_global_vram":
|
| 109 |
-
"max_process_vram":
|
| 110 |
"max_reserved": 555.74528,
|
| 111 |
"max_allocated": 499.37152
|
| 112 |
},
|
| 113 |
"latency": {
|
| 114 |
"unit": "s",
|
| 115 |
"count": 1,
|
| 116 |
-
"total": 7.
|
| 117 |
-
"mean": 7.
|
| 118 |
"stdev": 0.0,
|
| 119 |
-
"p50": 7.
|
| 120 |
-
"p90": 7.
|
| 121 |
-
"p95": 7.
|
| 122 |
-
"p99": 7.
|
| 123 |
"values": [
|
| 124 |
-
7.
|
| 125 |
]
|
| 126 |
},
|
| 127 |
"throughput": null,
|
|
@@ -131,149 +131,153 @@
|
|
| 131 |
"forward": {
|
| 132 |
"memory": {
|
| 133 |
"unit": "MB",
|
| 134 |
-
"max_ram":
|
| 135 |
-
"max_global_vram":
|
| 136 |
-
"max_process_vram":
|
| 137 |
"max_reserved": 555.74528,
|
| 138 |
"max_allocated": 499.5072
|
| 139 |
},
|
| 140 |
"latency": {
|
| 141 |
"unit": "s",
|
| 142 |
-
"count":
|
| 143 |
-
"total": 0.
|
| 144 |
-
"mean": 0.
|
| 145 |
-
"stdev": 0.
|
| 146 |
-
"p50": 0.
|
| 147 |
-
"p90": 0.
|
| 148 |
-
"p95": 0.
|
| 149 |
-
"p99": 0.
|
| 150 |
"values": [
|
| 151 |
-
0.
|
| 152 |
-
0.
|
| 153 |
-
0.
|
| 154 |
-
0.
|
| 155 |
-
0.
|
| 156 |
-
0.
|
| 157 |
-
0.
|
| 158 |
-
0.
|
| 159 |
-
0.
|
| 160 |
-
0.
|
| 161 |
-
0.
|
| 162 |
-
0.
|
| 163 |
-
0.
|
| 164 |
-
0.
|
| 165 |
-
0.
|
| 166 |
-
0.
|
| 167 |
-
0.
|
| 168 |
-
0.
|
| 169 |
-
0.
|
| 170 |
-
0.
|
| 171 |
-
0.
|
| 172 |
-
0.
|
| 173 |
-
0.
|
| 174 |
-
0.
|
| 175 |
-
0.
|
| 176 |
-
0.
|
| 177 |
-
0.
|
| 178 |
-
0.
|
| 179 |
-
0.
|
| 180 |
-
0.
|
| 181 |
-
0.
|
| 182 |
-
0.
|
| 183 |
-
0.
|
| 184 |
-
0.
|
| 185 |
-
0.
|
| 186 |
-
0.
|
| 187 |
-
0.
|
| 188 |
-
0.
|
| 189 |
-
0.
|
| 190 |
-
0.
|
| 191 |
-
0.
|
| 192 |
-
0.
|
| 193 |
-
0.
|
| 194 |
-
0.
|
| 195 |
-
0.
|
| 196 |
-
0.
|
| 197 |
-
0.
|
| 198 |
-
0.
|
| 199 |
-
0.
|
| 200 |
-
0.
|
| 201 |
-
0.
|
| 202 |
-
0.
|
| 203 |
-
0.
|
| 204 |
-
0.
|
| 205 |
-
0.
|
| 206 |
-
0.
|
| 207 |
-
0.
|
| 208 |
-
0.
|
| 209 |
-
0.
|
| 210 |
-
0.
|
| 211 |
-
0.
|
| 212 |
-
0.
|
| 213 |
-
0.
|
| 214 |
-
0.
|
| 215 |
-
0.
|
| 216 |
-
0.
|
| 217 |
-
0.
|
| 218 |
-
0.
|
| 219 |
-
0.
|
| 220 |
-
0.
|
| 221 |
-
0.
|
| 222 |
-
0.
|
| 223 |
-
0.
|
| 224 |
-
0.
|
| 225 |
-
0.
|
| 226 |
-
0.
|
| 227 |
-
0.
|
| 228 |
-
0.
|
| 229 |
-
0.
|
| 230 |
-
0.
|
| 231 |
-
0.
|
| 232 |
-
0.
|
| 233 |
-
0.
|
| 234 |
-
0.
|
| 235 |
-
0.
|
| 236 |
-
0.
|
| 237 |
-
0.
|
| 238 |
-
0.
|
| 239 |
-
0.
|
| 240 |
-
0.
|
| 241 |
-
0.
|
| 242 |
-
0.
|
| 243 |
-
0.
|
| 244 |
-
0.
|
| 245 |
-
0.
|
| 246 |
-
0.
|
| 247 |
-
0.
|
| 248 |
-
0.
|
| 249 |
-
0.
|
| 250 |
-
0.
|
| 251 |
-
0.
|
| 252 |
-
0.
|
| 253 |
-
0.
|
| 254 |
-
0.
|
| 255 |
-
0.
|
| 256 |
-
0.
|
| 257 |
-
0.
|
| 258 |
-
0.
|
| 259 |
-
0.
|
| 260 |
-
0.
|
| 261 |
-
0.
|
| 262 |
-
0.
|
| 263 |
-
0.
|
| 264 |
-
0.
|
| 265 |
-
0.
|
| 266 |
-
0.
|
| 267 |
-
0.
|
| 268 |
-
0.
|
| 269 |
-
0.
|
| 270 |
-
0.
|
| 271 |
-
0.
|
|
|
|
|
|
|
|
|
|
|
|
|
| 272 |
]
|
| 273 |
},
|
| 274 |
"throughput": {
|
| 275 |
"unit": "samples/s",
|
| 276 |
-
"value":
|
| 277 |
},
|
| 278 |
"energy": null,
|
| 279 |
"efficiency": null
|
|
|
|
| 3 |
"name": "cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base",
|
| 4 |
"backend": {
|
| 5 |
"name": "pytorch",
|
| 6 |
+
"version": "2.2.2+rocm5.7",
|
| 7 |
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
|
| 8 |
"task": "multiple-choice",
|
| 9 |
"library": "transformers",
|
|
|
|
| 104 |
"load": {
|
| 105 |
"memory": {
|
| 106 |
"unit": "MB",
|
| 107 |
+
"max_ram": 905.58464,
|
| 108 |
+
"max_global_vram": 841.764864,
|
| 109 |
+
"max_process_vram": 44807.573504,
|
| 110 |
"max_reserved": 555.74528,
|
| 111 |
"max_allocated": 499.37152
|
| 112 |
},
|
| 113 |
"latency": {
|
| 114 |
"unit": "s",
|
| 115 |
"count": 1,
|
| 116 |
+
"total": 7.1981728515625,
|
| 117 |
+
"mean": 7.1981728515625,
|
| 118 |
"stdev": 0.0,
|
| 119 |
+
"p50": 7.1981728515625,
|
| 120 |
+
"p90": 7.1981728515625,
|
| 121 |
+
"p95": 7.1981728515625,
|
| 122 |
+
"p99": 7.1981728515625,
|
| 123 |
"values": [
|
| 124 |
+
7.1981728515625
|
| 125 |
]
|
| 126 |
},
|
| 127 |
"throughput": null,
|
|
|
|
| 131 |
"forward": {
|
| 132 |
"memory": {
|
| 133 |
"unit": "MB",
|
| 134 |
+
"max_ram": 1015.414784,
|
| 135 |
+
"max_global_vram": 898.269184,
|
| 136 |
+
"max_process_vram": 218894.667776,
|
| 137 |
"max_reserved": 555.74528,
|
| 138 |
"max_allocated": 499.5072
|
| 139 |
},
|
| 140 |
"latency": {
|
| 141 |
"unit": "s",
|
| 142 |
+
"count": 125,
|
| 143 |
+
"total": 0.9963846902847285,
|
| 144 |
+
"mean": 0.007971077522277832,
|
| 145 |
+
"stdev": 0.0007331559968759307,
|
| 146 |
+
"p50": 0.007840778827667236,
|
| 147 |
+
"p90": 0.008722056388854982,
|
| 148 |
+
"p95": 0.00900333595275879,
|
| 149 |
+
"p99": 0.009678912391662601,
|
| 150 |
"values": [
|
| 151 |
+
0.008753096580505371,
|
| 152 |
+
0.008942537307739258,
|
| 153 |
+
0.009018535614013673,
|
| 154 |
+
0.009047016143798828,
|
| 155 |
+
0.00908973503112793,
|
| 156 |
+
0.009133416175842285,
|
| 157 |
+
0.009071334838867188,
|
| 158 |
+
0.009851174354553223,
|
| 159 |
+
0.00875389575958252,
|
| 160 |
+
0.008851016044616699,
|
| 161 |
+
0.008799016952514649,
|
| 162 |
+
0.008846055984497071,
|
| 163 |
+
0.008675496101379394,
|
| 164 |
+
0.00845213794708252,
|
| 165 |
+
0.008345577239990234,
|
| 166 |
+
0.008319977760314942,
|
| 167 |
+
0.008330376625061035,
|
| 168 |
+
0.008364458084106446,
|
| 169 |
+
0.00835085678100586,
|
| 170 |
+
0.00847997760772705,
|
| 171 |
+
0.00836589813232422,
|
| 172 |
+
0.008303977966308594,
|
| 173 |
+
0.008316938400268554,
|
| 174 |
+
0.008310697555541991,
|
| 175 |
+
0.008345577239990234,
|
| 176 |
+
0.008294537544250488,
|
| 177 |
+
0.008319658279418945,
|
| 178 |
+
0.00830253791809082,
|
| 179 |
+
0.008283978462219238,
|
| 180 |
+
0.008259178161621094,
|
| 181 |
+
0.008320937156677246,
|
| 182 |
+
0.008278377532958984,
|
| 183 |
+
0.007942698955535888,
|
| 184 |
+
0.008075018882751465,
|
| 185 |
+
0.008315977096557617,
|
| 186 |
+
0.008304298400878906,
|
| 187 |
+
0.008298057556152344,
|
| 188 |
+
0.008298856735229492,
|
| 189 |
+
0.00833421802520752,
|
| 190 |
+
0.008318697929382324,
|
| 191 |
+
0.00783117914199829,
|
| 192 |
+
0.007840778827667236,
|
| 193 |
+
0.007843499183654784,
|
| 194 |
+
0.007860138893127441,
|
| 195 |
+
0.00790685796737671,
|
| 196 |
+
0.0077487797737121585,
|
| 197 |
+
0.007526059150695801,
|
| 198 |
+
0.0074809398651123045,
|
| 199 |
+
0.007532299995422364,
|
| 200 |
+
0.0075518198013305666,
|
| 201 |
+
0.0075018987655639646,
|
| 202 |
+
0.007574540138244629,
|
| 203 |
+
0.007522220134735107,
|
| 204 |
+
0.0073225398063659665,
|
| 205 |
+
0.007338860034942627,
|
| 206 |
+
0.007332780838012695,
|
| 207 |
+
0.007328781127929687,
|
| 208 |
+
0.007329100131988525,
|
| 209 |
+
0.007349100112915039,
|
| 210 |
+
0.00734749984741211,
|
| 211 |
+
0.007346701145172119,
|
| 212 |
+
0.007363979816436768,
|
| 213 |
+
0.007343180179595947,
|
| 214 |
+
0.007353579998016358,
|
| 215 |
+
0.00728718090057373,
|
| 216 |
+
0.0075513401031494145,
|
| 217 |
+
0.007344620227813721,
|
| 218 |
+
0.007328620910644531,
|
| 219 |
+
0.007329740047454834,
|
| 220 |
+
0.007349899768829346,
|
| 221 |
+
0.00729725980758667,
|
| 222 |
+
0.007347980976104737,
|
| 223 |
+
0.0073580598831176755,
|
| 224 |
+
0.007381740093231201,
|
| 225 |
+
0.007370059967041016,
|
| 226 |
+
0.00741726016998291,
|
| 227 |
+
0.013565724372863769,
|
| 228 |
+
0.006860140800476074,
|
| 229 |
+
0.007444139957427979,
|
| 230 |
+
0.007424620151519776,
|
| 231 |
+
0.007429259777069092,
|
| 232 |
+
0.007548460006713867,
|
| 233 |
+
0.007392139911651612,
|
| 234 |
+
0.007597899913787841,
|
| 235 |
+
0.007853418827056884,
|
| 236 |
+
0.007895978927612304,
|
| 237 |
+
0.007876139163970947,
|
| 238 |
+
0.00805965805053711,
|
| 239 |
+
0.008069578170776368,
|
| 240 |
+
0.008212298393249511,
|
| 241 |
+
0.00815181827545166,
|
| 242 |
+
0.008466697692871094,
|
| 243 |
+
0.008303017616271972,
|
| 244 |
+
0.008314698219299317,
|
| 245 |
+
0.008339177131652832,
|
| 246 |
+
0.008228938102722167,
|
| 247 |
+
0.008234857559204101,
|
| 248 |
+
0.007965578079223632,
|
| 249 |
+
0.007995338916778564,
|
| 250 |
+
0.0077575788497924805,
|
| 251 |
+
0.00777341890335083,
|
| 252 |
+
0.00776957893371582,
|
| 253 |
+
0.008112937927246093,
|
| 254 |
+
0.00833309841156006,
|
| 255 |
+
0.008306378364562989,
|
| 256 |
+
0.007726058959960937,
|
| 257 |
+
0.007640140056610107,
|
| 258 |
+
0.007686379909515381,
|
| 259 |
+
0.0076103792190551755,
|
| 260 |
+
0.0077166190147399906,
|
| 261 |
+
0.007756619930267334,
|
| 262 |
+
0.00777966022491455,
|
| 263 |
+
0.007792298793792725,
|
| 264 |
+
0.007614058971405029,
|
| 265 |
+
0.007431980133056641,
|
| 266 |
+
0.007464300155639649,
|
| 267 |
+
0.007458380222320557,
|
| 268 |
+
0.007469580173492431,
|
| 269 |
+
0.007443339824676514,
|
| 270 |
+
0.007457419872283936,
|
| 271 |
+
0.007420939922332764,
|
| 272 |
+
0.007431340217590332,
|
| 273 |
+
0.007448619842529297,
|
| 274 |
+
0.007426061153411865,
|
| 275 |
+
0.007421259880065918
|
| 276 |
]
|
| 277 |
},
|
| 278 |
"throughput": {
|
| 279 |
"unit": "samples/s",
|
| 280 |
+
"value": 125.45355345060524
|
| 281 |
},
|
| 282 |
"energy": null,
|
| 283 |
"efficiency": null
|