Upload cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base/benchmark_report.json with huggingface_hub
Browse files
cuda_inference_transformers_multiple-choice_FacebookAI/roberta-base/benchmark_report.json
CHANGED
|
@@ -2,7 +2,7 @@
|
|
| 2 |
"forward": {
|
| 3 |
"memory": {
|
| 4 |
"unit": "MB",
|
| 5 |
-
"max_ram": 907.
|
| 6 |
"max_global_vram": 1195.900928,
|
| 7 |
"max_process_vram": 0.0,
|
| 8 |
"max_reserved": 555.74528,
|
|
@@ -10,167 +10,163 @@
|
|
| 10 |
},
|
| 11 |
"latency": {
|
| 12 |
"unit": "s",
|
| 13 |
-
"count":
|
| 14 |
-
"total":
|
| 15 |
-
"mean": 0.
|
| 16 |
-
"stdev": 0.
|
| 17 |
-
"p50": 0.
|
| 18 |
-
"p90": 0.
|
| 19 |
-
"p95": 0.
|
| 20 |
-
"p99": 0.
|
| 21 |
"values": [
|
| 22 |
-
0.
|
| 23 |
-
0.
|
| 24 |
-
0.
|
| 25 |
-
0.
|
| 26 |
-
0.
|
| 27 |
-
0.
|
| 28 |
-
0.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 29 |
0.007953407764434815,
|
| 30 |
-
0.
|
| 31 |
-
0.
|
| 32 |
-
0.
|
| 33 |
-
0.
|
| 34 |
-
0.
|
| 35 |
-
0.
|
| 36 |
-
0.
|
| 37 |
-
0.
|
| 38 |
-
0.
|
| 39 |
-
0.
|
| 40 |
-
0.
|
| 41 |
-
0.
|
| 42 |
-
0.
|
| 43 |
-
0.
|
| 44 |
-
0.
|
| 45 |
-
0.
|
| 46 |
-
0.
|
| 47 |
-
0.
|
| 48 |
-
0.
|
| 49 |
-
0.
|
| 50 |
-
0.
|
| 51 |
-
0.
|
| 52 |
-
0.
|
| 53 |
-
0.
|
| 54 |
-
0.
|
| 55 |
-
0.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 56 |
0.007165952205657959,
|
| 57 |
-
0.
|
| 58 |
-
0.
|
| 59 |
-
0.
|
| 60 |
-
0.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 61 |
0.007120895862579346,
|
| 62 |
-
0.
|
| 63 |
-
0.
|
| 64 |
-
0.
|
| 65 |
-
0.
|
| 66 |
-
0.
|
| 67 |
-
0.
|
| 68 |
-
0.
|
| 69 |
-
0.
|
| 70 |
-
0.
|
| 71 |
-
0.
|
| 72 |
-
0.007127039909362793,
|
| 73 |
-
0.007135231971740722,
|
| 74 |
-
0.007097343921661377,
|
| 75 |
-
0.0071157760620117185,
|
| 76 |
-
0.007090176105499267,
|
| 77 |
-
0.007107583999633789,
|
| 78 |
-
0.007072768211364746,
|
| 79 |
-
0.007087103843688965,
|
| 80 |
-
0.00710041618347168,
|
| 81 |
-
0.007102464199066162,
|
| 82 |
-
0.007109632015228271,
|
| 83 |
-
0.007048096179962158,
|
| 84 |
-
0.007110655784606934,
|
| 85 |
-
0.007104512214660645,
|
| 86 |
-
0.007130112171173096,
|
| 87 |
-
0.007090176105499267,
|
| 88 |
-
0.007677951812744141,
|
| 89 |
-
0.007126016139984131,
|
| 90 |
-
0.007117824077606201,
|
| 91 |
-
0.00709119987487793,
|
| 92 |
-
0.007107583999633789,
|
| 93 |
-
0.007158783912658692,
|
| 94 |
-
0.007016448020935059,
|
| 95 |
-
0.007031807899475098,
|
| 96 |
-
0.007076863765716553,
|
| 97 |
-
0.0070594558715820314,
|
| 98 |
-
0.007203839778900147,
|
| 99 |
-
0.0071198720932006835,
|
| 100 |
-
0.007114783763885498,
|
| 101 |
-
0.007094272136688232,
|
| 102 |
-
0.0070696320533752445,
|
| 103 |
-
0.007101439952850342,
|
| 104 |
-
0.007097248077392578,
|
| 105 |
-
0.00713318395614624,
|
| 106 |
-
0.007074816226959229,
|
| 107 |
-
0.007047167778015137,
|
| 108 |
-
0.007093247890472412,
|
| 109 |
-
0.007098368167877197,
|
| 110 |
-
0.00708403205871582,
|
| 111 |
-
0.007057407855987549,
|
| 112 |
-
0.007116799831390381,
|
| 113 |
-
0.007096320152282715,
|
| 114 |
-
0.0074741759300231934,
|
| 115 |
-
0.00748748779296875,
|
| 116 |
-
0.007504896163940429,
|
| 117 |
-
0.007418879985809326,
|
| 118 |
-
0.0074332160949707035,
|
| 119 |
-
0.007423999786376953,
|
| 120 |
-
0.007457791805267334,
|
| 121 |
-
0.007437312126159668,
|
| 122 |
-
0.007344128131866455,
|
| 123 |
-
0.007351295948028564,
|
| 124 |
-
0.00729804801940918,
|
| 125 |
-
0.007367680072784424,
|
| 126 |
-
0.007304192066192627,
|
| 127 |
-
0.007365568161010742,
|
| 128 |
-
0.007395328044891358,
|
| 129 |
-
0.0074414081573486324,
|
| 130 |
-
0.007409632205963135,
|
| 131 |
-
0.007386112213134765,
|
| 132 |
-
0.007419904232025146,
|
| 133 |
-
0.007468031883239746,
|
| 134 |
-
0.007437312126159668,
|
| 135 |
-
0.0074414081573486324,
|
| 136 |
-
0.007425024032592774,
|
| 137 |
-
0.007431168079376221,
|
| 138 |
-
0.007395328044891358,
|
| 139 |
-
0.007450623989105225,
|
| 140 |
-
0.007457791805267334,
|
| 141 |
-
0.00746284818649292,
|
| 142 |
-
0.00745472002029419,
|
| 143 |
-
0.007435264110565186,
|
| 144 |
-
0.007419904232025146,
|
| 145 |
-
0.007398399829864502,
|
| 146 |
-
0.007386112213134765,
|
| 147 |
-
0.007457791805267334,
|
| 148 |
-
0.007423999786376953,
|
| 149 |
-
0.007426047801971435,
|
| 150 |
-
0.007421919822692871,
|
| 151 |
-
0.007434239864349365,
|
| 152 |
-
0.007459839820861816,
|
| 153 |
-
0.007425024032592774,
|
| 154 |
-
0.007428095817565918,
|
| 155 |
-
0.007431168079376221,
|
| 156 |
-
0.007445504188537597,
|
| 157 |
-
0.00744652795791626
|
| 158 |
]
|
| 159 |
},
|
| 160 |
"throughput": {
|
| 161 |
"unit": "samples/s",
|
| 162 |
-
"value":
|
| 163 |
},
|
| 164 |
"energy": {
|
| 165 |
"unit": "kWh",
|
| 166 |
-
"cpu": 8.
|
| 167 |
-
"ram": 4.
|
| 168 |
-
"gpu": 1.
|
| 169 |
-
"total": 2.
|
| 170 |
},
|
| 171 |
"efficiency": {
|
| 172 |
"unit": "samples/kWh",
|
| 173 |
-
"value":
|
| 174 |
}
|
| 175 |
}
|
| 176 |
}
|
|
|
|
| 2 |
"forward": {
|
| 3 |
"memory": {
|
| 4 |
"unit": "MB",
|
| 5 |
+
"max_ram": 907.358208,
|
| 6 |
"max_global_vram": 1195.900928,
|
| 7 |
"max_process_vram": 0.0,
|
| 8 |
"max_reserved": 555.74528,
|
|
|
|
| 10 |
},
|
| 11 |
"latency": {
|
| 12 |
"unit": "s",
|
| 13 |
+
"count": 132,
|
| 14 |
+
"total": 1.0019654092788697,
|
| 15 |
+
"mean": 0.007590647039991437,
|
| 16 |
+
"stdev": 0.0005809214402145674,
|
| 17 |
+
"p50": 0.007497215986251832,
|
| 18 |
+
"p90": 0.007849881744384765,
|
| 19 |
+
"p95": 0.008180326747894286,
|
| 20 |
+
"p99": 0.010368777799606323,
|
| 21 |
"values": [
|
| 22 |
+
0.010809344291687012,
|
| 23 |
+
0.010365951538085937,
|
| 24 |
+
0.010370047569274902,
|
| 25 |
+
0.010334207534790038,
|
| 26 |
+
0.008263680458068847,
|
| 27 |
+
0.008077247619628906,
|
| 28 |
+
0.007772160053253174,
|
| 29 |
+
0.007849984169006348,
|
| 30 |
+
0.007860223770141601,
|
| 31 |
+
0.00775980806350708,
|
| 32 |
+
0.007813119888305664,
|
| 33 |
+
0.007738368034362793,
|
| 34 |
+
0.00760422420501709,
|
| 35 |
+
0.007623680114746094,
|
| 36 |
+
0.007647232055664062,
|
| 37 |
+
0.007670783996582031,
|
| 38 |
+
0.0077281279563903805,
|
| 39 |
+
0.007759871959686279,
|
| 40 |
+
0.007775231838226319,
|
| 41 |
+
0.007679999828338623,
|
| 42 |
+
0.007666687965393066,
|
| 43 |
+
0.00744755220413208,
|
| 44 |
+
0.007442431926727295,
|
| 45 |
+
0.007530432224273681,
|
| 46 |
+
0.007450623989105225,
|
| 47 |
+
0.00744755220413208,
|
| 48 |
+
0.0073697280883789065,
|
| 49 |
+
0.007373824119567871,
|
| 50 |
+
0.00738099193572998,
|
| 51 |
+
0.007883776187896728,
|
| 52 |
+
0.007751679897308349,
|
| 53 |
+
0.007715839862823487,
|
| 54 |
+
0.007575551986694336,
|
| 55 |
+
0.007740416049957275,
|
| 56 |
+
0.007640063762664795,
|
| 57 |
+
0.007618559837341309,
|
| 58 |
+
0.007633920192718506,
|
| 59 |
+
0.007709695816040039,
|
| 60 |
+
0.007714848041534424,
|
| 61 |
+
0.007730175971984863,
|
| 62 |
+
0.007783423900604248,
|
| 63 |
+
0.007670783996582031,
|
| 64 |
+
0.007659520149230957,
|
| 65 |
+
0.007577600002288819,
|
| 66 |
+
0.007701504230499268,
|
| 67 |
+
0.007727136135101318,
|
| 68 |
+
0.007686143875122071,
|
| 69 |
+
0.007473152160644531,
|
| 70 |
+
0.007404543876647949,
|
| 71 |
+
0.007310336112976074,
|
| 72 |
+
0.007361536026000977,
|
| 73 |
+
0.007288832187652588,
|
| 74 |
+
0.007318528175354004,
|
| 75 |
+
0.007401408195495606,
|
| 76 |
+
0.007353343963623047,
|
| 77 |
+
0.007592959880828858,
|
| 78 |
+
0.007602176189422607,
|
| 79 |
+
0.007684095859527588,
|
| 80 |
+
0.007584767818450928,
|
| 81 |
+
0.007624703884124756,
|
| 82 |
+
0.007709695816040039,
|
| 83 |
+
0.007574528217315674,
|
| 84 |
+
0.007775231838226319,
|
| 85 |
+
0.007469056129455566,
|
| 86 |
+
0.007442431926727295,
|
| 87 |
+
0.007435264110565186,
|
| 88 |
+
0.00808140754699707,
|
| 89 |
+
0.008516608238220215,
|
| 90 |
0.007953407764434815,
|
| 91 |
+
0.007694272041320801,
|
| 92 |
+
0.00774348783493042,
|
| 93 |
+
0.007689216136932373,
|
| 94 |
+
0.007505919933319092,
|
| 95 |
+
0.007451648235321045,
|
| 96 |
+
0.007628896236419678,
|
| 97 |
+
0.007448575973510742,
|
| 98 |
+
0.0072499198913574215,
|
| 99 |
+
0.007223296165466309,
|
| 100 |
+
0.0074403839111328125,
|
| 101 |
+
0.007488512039184571,
|
| 102 |
+
0.007554048061370849,
|
| 103 |
+
0.007553023815155029,
|
| 104 |
+
0.007576576232910156,
|
| 105 |
+
0.007797760009765625,
|
| 106 |
+
0.007848959922790527,
|
| 107 |
+
0.007590911865234375,
|
| 108 |
+
0.007327744007110596,
|
| 109 |
+
0.008112128257751466,
|
| 110 |
+
0.007512063980102539,
|
| 111 |
+
0.0072837119102478025,
|
| 112 |
+
0.007245823860168457,
|
| 113 |
+
0.00724889612197876,
|
| 114 |
+
0.0071495680809020995,
|
| 115 |
+
0.007219200134277344,
|
| 116 |
+
0.007226367950439453,
|
| 117 |
+
0.007227424144744873,
|
| 118 |
+
0.007808000087738037,
|
| 119 |
+
0.007456768035888672,
|
| 120 |
+
0.007237631797790528,
|
| 121 |
+
0.0071823358535766605,
|
| 122 |
+
0.007185408115386963,
|
| 123 |
+
0.00719974422454834,
|
| 124 |
+
0.007180287837982178,
|
| 125 |
+
0.007193600177764893,
|
| 126 |
+
0.007158783912658692,
|
| 127 |
+
0.007160799980163574,
|
| 128 |
+
0.00838144016265869,
|
| 129 |
+
0.007326720237731933,
|
| 130 |
+
0.007231488227844239,
|
| 131 |
+
0.0071987838745117185,
|
| 132 |
0.007165952205657959,
|
| 133 |
+
0.007201791763305664,
|
| 134 |
+
0.007181312084197998,
|
| 135 |
+
0.007197696208953858,
|
| 136 |
+
0.007192575931549072,
|
| 137 |
+
0.0071792640686035155,
|
| 138 |
+
0.007153664112091064,
|
| 139 |
+
0.007153632164001465,
|
| 140 |
+
0.007222271919250488,
|
| 141 |
+
0.0072120318412780765,
|
| 142 |
+
0.007187456130981445,
|
| 143 |
0.007120895862579346,
|
| 144 |
+
0.007156735897064209,
|
| 145 |
+
0.007221216201782226,
|
| 146 |
+
0.007202816009521484,
|
| 147 |
+
0.007206912040710449,
|
| 148 |
+
0.007145472049713135,
|
| 149 |
+
0.007170048236846924,
|
| 150 |
+
0.007160831928253173,
|
| 151 |
+
0.007176191806793213,
|
| 152 |
+
0.007171072006225586,
|
| 153 |
+
0.007153664112091064
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 154 |
]
|
| 155 |
},
|
| 156 |
"throughput": {
|
| 157 |
"unit": "samples/s",
|
| 158 |
+
"value": 131.74107486904413
|
| 159 |
},
|
| 160 |
"energy": {
|
| 161 |
"unit": "kWh",
|
| 162 |
+
"cpu": 8.530816633543142e-08,
|
| 163 |
+
"ram": 4.662791786273805e-08,
|
| 164 |
+
"gpu": 1.6067358976977895e-07,
|
| 165 |
+
"total": 2.926096739679484e-07
|
| 166 |
},
|
| 167 |
"efficiency": {
|
| 168 |
"unit": "samples/kWh",
|
| 169 |
+
"value": 3417522.0061573805
|
| 170 |
}
|
| 171 |
}
|
| 172 |
}
|