Upload cuda_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark_report.json with huggingface_hub
Browse files
cuda_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark_report.json
CHANGED
|
@@ -2,167 +2,174 @@
|
|
| 2 |
"forward": {
|
| 3 |
"memory": {
|
| 4 |
"unit": "MB",
|
| 5 |
-
"max_ram":
|
| 6 |
"max_global_vram": 897.14688,
|
| 7 |
-
"max_process_vram":
|
| 8 |
"max_reserved": 555.74528,
|
| 9 |
"max_allocated": 499.443712
|
| 10 |
},
|
| 11 |
"latency": {
|
| 12 |
"unit": "s",
|
| 13 |
-
"count":
|
| 14 |
-
"total": 0.
|
| 15 |
-
"mean": 0.
|
| 16 |
-
"stdev": 0.
|
| 17 |
-
"p50": 0.
|
| 18 |
-
"p90": 0.
|
| 19 |
-
"p95": 0.
|
| 20 |
-
"p99": 0.
|
| 21 |
"values": [
|
| 22 |
-
0.
|
| 23 |
-
0.
|
| 24 |
-
0.
|
| 25 |
-
0.
|
| 26 |
-
0.
|
| 27 |
-
0.
|
| 28 |
-
0.
|
| 29 |
-
0.
|
| 30 |
-
0.
|
| 31 |
-
0.
|
| 32 |
-
0.
|
| 33 |
-
0.
|
| 34 |
-
0.
|
| 35 |
-
0.
|
| 36 |
-
0.
|
| 37 |
-
0.
|
| 38 |
-
0.
|
| 39 |
-
0.
|
| 40 |
-
0.
|
| 41 |
-
0.
|
| 42 |
-
0.
|
| 43 |
-
0.
|
| 44 |
-
0.
|
| 45 |
-
0.
|
| 46 |
-
0.
|
| 47 |
-
0.
|
| 48 |
-
0.
|
| 49 |
-
0.
|
| 50 |
-
0.
|
| 51 |
-
0.
|
| 52 |
-
0.
|
| 53 |
-
0.
|
| 54 |
-
0.
|
| 55 |
-
0.
|
| 56 |
-
0.
|
| 57 |
-
0.
|
| 58 |
-
0.
|
| 59 |
-
0.
|
| 60 |
-
0.
|
| 61 |
-
0.
|
| 62 |
-
0.
|
| 63 |
-
0.
|
| 64 |
-
0.
|
| 65 |
-
0.
|
| 66 |
-
0.
|
| 67 |
-
0.
|
| 68 |
-
0.
|
| 69 |
-
0.
|
| 70 |
-
0.
|
| 71 |
-
0.
|
| 72 |
-
0.
|
| 73 |
-
0.
|
| 74 |
-
0.
|
| 75 |
-
0.
|
| 76 |
-
0.
|
| 77 |
-
0.
|
| 78 |
-
0.
|
| 79 |
-
0.
|
| 80 |
-
0.
|
| 81 |
-
0.
|
| 82 |
-
0.
|
| 83 |
-
0.
|
| 84 |
-
0.
|
| 85 |
-
0.
|
| 86 |
-
0.
|
| 87 |
-
0.
|
| 88 |
-
0.
|
| 89 |
-
0.
|
| 90 |
-
0.
|
| 91 |
-
0.
|
| 92 |
-
0.
|
| 93 |
-
0.
|
| 94 |
-
0.
|
| 95 |
-
0.
|
| 96 |
-
0.
|
| 97 |
-
0.
|
| 98 |
-
0.
|
| 99 |
-
0.
|
| 100 |
-
0.
|
| 101 |
-
0.
|
| 102 |
-
0.
|
| 103 |
-
0.
|
| 104 |
-
0.
|
| 105 |
-
0.
|
| 106 |
-
0.
|
| 107 |
-
0.
|
| 108 |
-
0.
|
| 109 |
-
0.
|
| 110 |
-
0.
|
| 111 |
-
0.
|
| 112 |
-
0.
|
| 113 |
-
0.
|
| 114 |
-
0.
|
| 115 |
-
0.
|
| 116 |
-
0.
|
| 117 |
-
0.
|
| 118 |
-
0.
|
| 119 |
-
0.
|
| 120 |
-
0.
|
| 121 |
-
0.
|
| 122 |
-
0.
|
| 123 |
-
0.
|
| 124 |
-
0.
|
| 125 |
-
0.
|
| 126 |
-
0.
|
| 127 |
-
0.
|
| 128 |
-
0.
|
| 129 |
-
0.
|
| 130 |
-
0.
|
| 131 |
-
0.
|
| 132 |
-
0.
|
| 133 |
-
0.
|
| 134 |
-
0.
|
| 135 |
-
0.
|
| 136 |
-
0.
|
| 137 |
-
0.
|
| 138 |
-
0.
|
| 139 |
-
0.
|
| 140 |
-
0.
|
| 141 |
-
0.
|
| 142 |
-
0.
|
| 143 |
-
0.
|
| 144 |
-
0.
|
| 145 |
-
0.
|
| 146 |
-
0.
|
| 147 |
-
0.
|
| 148 |
-
0.
|
| 149 |
-
0.
|
| 150 |
-
0.
|
| 151 |
-
0.
|
| 152 |
-
0.
|
| 153 |
-
0.
|
| 154 |
-
0.
|
| 155 |
-
0.
|
| 156 |
-
0.
|
| 157 |
-
0.
|
| 158 |
-
0.
|
| 159 |
-
0.
|
| 160 |
-
0.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 161 |
]
|
| 162 |
},
|
| 163 |
"throughput": {
|
| 164 |
"unit": "samples/s",
|
| 165 |
-
"value":
|
| 166 |
},
|
| 167 |
"energy": null,
|
| 168 |
"efficiency": null
|
|
|
|
| 2 |
"forward": {
|
| 3 |
"memory": {
|
| 4 |
"unit": "MB",
|
| 5 |
+
"max_ram": 1007.775744,
|
| 6 |
"max_global_vram": 897.14688,
|
| 7 |
+
"max_process_vram": 207166.001152,
|
| 8 |
"max_reserved": 555.74528,
|
| 9 |
"max_allocated": 499.443712
|
| 10 |
},
|
| 11 |
"latency": {
|
| 12 |
"unit": "s",
|
| 13 |
+
"count": 146,
|
| 14 |
+
"total": 0.9992842464447022,
|
| 15 |
+
"mean": 0.0068444126468815225,
|
| 16 |
+
"stdev": 0.00033736598203322157,
|
| 17 |
+
"p50": 0.006653344392776489,
|
| 18 |
+
"p90": 0.0071704633235931394,
|
| 19 |
+
"p95": 0.007282543063163757,
|
| 20 |
+
"p99": 0.007604917860031129,
|
| 21 |
"values": [
|
| 22 |
+
0.006918064117431641,
|
| 23 |
+
0.007029102802276611,
|
| 24 |
+
0.007152944087982178,
|
| 25 |
+
0.006964622974395752,
|
| 26 |
+
0.007236303806304932,
|
| 27 |
+
0.007088462829589844,
|
| 28 |
+
0.007127984046936035,
|
| 29 |
+
0.007232143878936768,
|
| 30 |
+
0.007292782783508301,
|
| 31 |
+
0.007223503112792969,
|
| 32 |
+
0.00718798303604126,
|
| 33 |
+
0.007133743762969971,
|
| 34 |
+
0.007087663173675537,
|
| 35 |
+
0.00706382417678833,
|
| 36 |
+
0.007042063236236572,
|
| 37 |
+
0.007122543811798096,
|
| 38 |
+
0.007009744167327881,
|
| 39 |
+
0.007002543926239013,
|
| 40 |
+
0.007245742797851563,
|
| 41 |
+
0.006998383998870849,
|
| 42 |
+
0.006989583969116211,
|
| 43 |
+
0.007004623889923096,
|
| 44 |
+
0.007168943881988525,
|
| 45 |
+
0.006998544216156006,
|
| 46 |
+
0.006997104167938232,
|
| 47 |
+
0.006998223781585693,
|
| 48 |
+
0.00698414421081543,
|
| 49 |
+
0.006962543964385987,
|
| 50 |
+
0.006985583782196045,
|
| 51 |
+
0.006987984180450439,
|
| 52 |
+
0.00703934383392334,
|
| 53 |
+
0.0070198230743408205,
|
| 54 |
+
0.0070985441207885745,
|
| 55 |
+
0.00706894302368164,
|
| 56 |
+
0.007102863788604736,
|
| 57 |
+
0.007087983131408691,
|
| 58 |
+
0.007055503845214844,
|
| 59 |
+
0.007051183223724365,
|
| 60 |
+
0.007036464214324951,
|
| 61 |
+
0.007053584098815918,
|
| 62 |
+
0.007057263851165772,
|
| 63 |
+
0.007028943061828614,
|
| 64 |
+
0.007054384231567383,
|
| 65 |
+
0.007629901885986328,
|
| 66 |
+
0.009487338066101074,
|
| 67 |
+
0.006663185119628906,
|
| 68 |
+
0.006615503787994385,
|
| 69 |
+
0.006653104782104492,
|
| 70 |
+
0.006624144077301025,
|
| 71 |
+
0.006624625205993652,
|
| 72 |
+
0.006623505115509033,
|
| 73 |
+
0.006619823932647705,
|
| 74 |
+
0.006612464904785156,
|
| 75 |
+
0.0066391839981079104,
|
| 76 |
+
0.0066057448387146,
|
| 77 |
+
0.0066187047958374025,
|
| 78 |
+
0.006615824222564697,
|
| 79 |
+
0.006615345001220703,
|
| 80 |
+
0.006597905158996582,
|
| 81 |
+
0.006640625,
|
| 82 |
+
0.0066175851821899416,
|
| 83 |
+
0.0066251039505004885,
|
| 84 |
+
0.006625424861907959,
|
| 85 |
+
0.00663630485534668,
|
| 86 |
+
0.0066390252113342286,
|
| 87 |
+
0.006674705028533935,
|
| 88 |
+
0.006625264167785644,
|
| 89 |
+
0.00732990312576294,
|
| 90 |
+
0.006650064945220947,
|
| 91 |
+
0.006653584003448486,
|
| 92 |
+
0.006635984897613525,
|
| 93 |
+
0.006619344234466553,
|
| 94 |
+
0.006628465175628662,
|
| 95 |
+
0.006609745025634765,
|
| 96 |
+
0.006622223854064941,
|
| 97 |
+
0.006624145030975342,
|
| 98 |
+
0.006631184101104737,
|
| 99 |
+
0.006604625225067139,
|
| 100 |
+
0.006635664939880371,
|
| 101 |
+
0.006852143764495849,
|
| 102 |
+
0.006651504039764404,
|
| 103 |
+
0.006669425010681153,
|
| 104 |
+
0.00663102388381958,
|
| 105 |
+
0.006626224994659423,
|
| 106 |
+
0.006619345188140869,
|
| 107 |
+
0.006622543811798096,
|
| 108 |
+
0.006588144779205322,
|
| 109 |
+
0.006623504161834717,
|
| 110 |
+
0.006612945079803467,
|
| 111 |
+
0.006645744800567627,
|
| 112 |
+
0.006630224227905273,
|
| 113 |
+
0.006673425197601318,
|
| 114 |
+
0.006667984008789062,
|
| 115 |
+
0.0066297450065612796,
|
| 116 |
+
0.006641263961791992,
|
| 117 |
+
0.00665102481842041,
|
| 118 |
+
0.006656464099884033,
|
| 119 |
+
0.006607824802398681,
|
| 120 |
+
0.006591344833374023,
|
| 121 |
+
0.006606544017791748,
|
| 122 |
+
0.006615345001220703,
|
| 123 |
+
0.006618384838104248,
|
| 124 |
+
0.006600304126739502,
|
| 125 |
+
0.006599024772644043,
|
| 126 |
+
0.0066276640892028804,
|
| 127 |
+
0.006648145198822022,
|
| 128 |
+
0.006631824016571045,
|
| 129 |
+
0.006607824802398681,
|
| 130 |
+
0.006624625205993652,
|
| 131 |
+
0.006631824016571045,
|
| 132 |
+
0.006592784881591797,
|
| 133 |
+
0.006606385231018066,
|
| 134 |
+
0.0066156649589538576,
|
| 135 |
+
0.006617744922637939,
|
| 136 |
+
0.006616623878479004,
|
| 137 |
+
0.006602224826812744,
|
| 138 |
+
0.007171982765197754,
|
| 139 |
+
0.007283183097839355,
|
| 140 |
+
0.007023663997650147,
|
| 141 |
+
0.006892623901367187,
|
| 142 |
+
0.006971663951873779,
|
| 143 |
+
0.007280622959136963,
|
| 144 |
+
0.007574381828308105,
|
| 145 |
+
0.007570542812347412,
|
| 146 |
+
0.007541741847991943,
|
| 147 |
+
0.007143663883209229,
|
| 148 |
+
0.00696062421798706,
|
| 149 |
+
0.007090864181518555,
|
| 150 |
+
0.007088624000549317,
|
| 151 |
+
0.00674430513381958,
|
| 152 |
+
0.0067118239402771,
|
| 153 |
+
0.0067790250778198245,
|
| 154 |
+
0.0067556648254394535,
|
| 155 |
+
0.006706863880157471,
|
| 156 |
+
0.0067460651397705075,
|
| 157 |
+
0.006621904850006103,
|
| 158 |
+
0.0065900650024414064,
|
| 159 |
+
0.006601103782653809,
|
| 160 |
+
0.006603344917297364,
|
| 161 |
+
0.006604625225067139,
|
| 162 |
+
0.006611663818359375,
|
| 163 |
+
0.006619665145874023,
|
| 164 |
+
0.006598865032196045,
|
| 165 |
+
0.006625583171844482,
|
| 166 |
+
0.006638384819030761,
|
| 167 |
+
0.006608784198760986
|
| 168 |
]
|
| 169 |
},
|
| 170 |
"throughput": {
|
| 171 |
"unit": "samples/s",
|
| 172 |
+
"value": 146.1045748689077
|
| 173 |
},
|
| 174 |
"energy": null,
|
| 175 |
"efficiency": null
|