Upload cuda_inference_transformers_token-classification_microsoft/deberta-v3-base/benchmark.json with huggingface_hub
Browse files
cuda_inference_transformers_token-classification_microsoft/deberta-v3-base/benchmark.json
CHANGED
|
@@ -102,7 +102,7 @@
|
|
| 102 |
"forward": {
|
| 103 |
"memory": {
|
| 104 |
"unit": "MB",
|
| 105 |
-
"max_ram": 975.
|
| 106 |
"max_global_vram": 1434.976256,
|
| 107 |
"max_process_vram": 0.0,
|
| 108 |
"max_reserved": 794.820608,
|
|
@@ -110,102 +110,101 @@
|
|
| 110 |
},
|
| 111 |
"latency": {
|
| 112 |
"unit": "s",
|
| 113 |
-
"count":
|
| 114 |
-
"total": 1.
|
| 115 |
-
"mean": 0.
|
| 116 |
-
"stdev": 0.
|
| 117 |
-
"p50": 0.
|
| 118 |
-
"p90": 0.
|
| 119 |
-
"p95": 0.
|
| 120 |
-
"p99": 0.
|
| 121 |
"values": [
|
| 122 |
-
0.
|
| 123 |
-
0.
|
| 124 |
-
0.
|
| 125 |
-
0.
|
| 126 |
-
0.
|
| 127 |
-
0.
|
| 128 |
-
0.
|
| 129 |
-
0.
|
| 130 |
-
0.
|
| 131 |
-
0.
|
| 132 |
-
0.
|
| 133 |
-
0.013394944190979004,
|
| 134 |
-
0.013288448333740235,
|
| 135 |
-
0.013904895782470703,
|
| 136 |
-
0.013757439613342285,
|
| 137 |
-
0.01375222396850586,
|
| 138 |
-
0.013810688018798829,
|
| 139 |
-
0.013786111831665039,
|
| 140 |
-
0.013772768020629883,
|
| 141 |
-
0.013760512351989745,
|
| 142 |
-
0.013472767829895019,
|
| 143 |
-
0.013379584312438965,
|
| 144 |
-
0.013337568283081054,
|
| 145 |
-
0.013616127967834473,
|
| 146 |
-
0.013947839736938477,
|
| 147 |
-
0.013999103546142578,
|
| 148 |
-
0.01396224021911621,
|
| 149 |
-
0.013993984222412109,
|
| 150 |
-
0.013989888191223144,
|
| 151 |
-
0.014013440132141113,
|
| 152 |
-
0.013957119941711426,
|
| 153 |
-
0.013981696128845214,
|
| 154 |
-
0.013912063598632812,
|
| 155 |
-
0.01395404815673828,
|
| 156 |
-
0.01387724781036377,
|
| 157 |
-
0.013880319595336914,
|
| 158 |
-
0.013839360237121581,
|
| 159 |
-
0.01380352020263672,
|
| 160 |
-
0.013737983703613281,
|
| 161 |
-
0.013824000358581542,
|
| 162 |
-
0.01389363193511963,
|
| 163 |
-
0.013955072402954101,
|
| 164 |
-
0.013940735816955567,
|
| 165 |
-
0.014006272315979004,
|
| 166 |
-
0.01448140811920166,
|
| 167 |
-
0.014317567825317384,
|
| 168 |
-
0.014907391548156738,
|
| 169 |
-
0.014527487754821777,
|
| 170 |
-
0.01415987205505371,
|
| 171 |
-
0.014041088104248046,
|
| 172 |
-
0.014166015625,
|
| 173 |
0.014106623649597168,
|
| 174 |
-
0.
|
| 175 |
-
0.
|
| 176 |
-
0.
|
| 177 |
-
0.
|
| 178 |
-
0.
|
| 179 |
-
0.
|
| 180 |
-
0.
|
| 181 |
-
0.
|
| 182 |
-
0.
|
| 183 |
-
0.
|
| 184 |
-
0.
|
| 185 |
-
0.
|
| 186 |
-
0.
|
| 187 |
-
0.
|
| 188 |
-
0.
|
| 189 |
-
0.
|
| 190 |
-
0.
|
| 191 |
-
0.
|
| 192 |
-
0.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 193 |
]
|
| 194 |
},
|
| 195 |
"throughput": {
|
| 196 |
"unit": "samples/s",
|
| 197 |
-
"value":
|
| 198 |
},
|
| 199 |
"energy": {
|
| 200 |
"unit": "kWh",
|
| 201 |
-
"cpu": 1.
|
| 202 |
-
"ram":
|
| 203 |
-
"gpu": 3.
|
| 204 |
-
"total": 5.
|
| 205 |
},
|
| 206 |
"efficiency": {
|
| 207 |
"unit": "samples/kWh",
|
| 208 |
-
"value":
|
| 209 |
}
|
| 210 |
}
|
| 211 |
}
|
|
|
|
| 102 |
"forward": {
|
| 103 |
"memory": {
|
| 104 |
"unit": "MB",
|
| 105 |
+
"max_ram": 975.536128,
|
| 106 |
"max_global_vram": 1434.976256,
|
| 107 |
"max_process_vram": 0.0,
|
| 108 |
"max_reserved": 794.820608,
|
|
|
|
| 110 |
},
|
| 111 |
"latency": {
|
| 112 |
"unit": "s",
|
| 113 |
+
"count": 70,
|
| 114 |
+
"total": 1.0018957738876342,
|
| 115 |
+
"mean": 0.014312796769823345,
|
| 116 |
+
"stdev": 0.000471044854333861,
|
| 117 |
+
"p50": 0.014249471664428711,
|
| 118 |
+
"p90": 0.014793625259399413,
|
| 119 |
+
"p95": 0.01523875856399536,
|
| 120 |
+
"p99": 0.01614117872238159,
|
| 121 |
"values": [
|
| 122 |
+
0.015731712341308594,
|
| 123 |
+
0.01640755271911621,
|
| 124 |
+
0.016021503448486327,
|
| 125 |
+
0.015520768165588379,
|
| 126 |
+
0.01480191993713379,
|
| 127 |
+
0.014835712432861328,
|
| 128 |
+
0.014611424446105957,
|
| 129 |
+
0.014012415885925293,
|
| 130 |
+
0.014156800270080566,
|
| 131 |
+
0.014161919593811035,
|
| 132 |
+
0.014011360168457032,
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 133 |
0.014106623649597168,
|
| 134 |
+
0.014136256217956543,
|
| 135 |
+
0.01419059181213379,
|
| 136 |
+
0.013983743667602539,
|
| 137 |
+
0.014202879905700684,
|
| 138 |
+
0.014042112350463867,
|
| 139 |
+
0.014048255920410157,
|
| 140 |
+
0.013971455574035644,
|
| 141 |
+
0.014018560409545898,
|
| 142 |
+
0.01414252758026123,
|
| 143 |
+
0.01437996768951416,
|
| 144 |
+
0.014134271621704102,
|
| 145 |
+
0.014041024208068848,
|
| 146 |
+
0.014183423995971679,
|
| 147 |
+
0.01409126377105713,
|
| 148 |
+
0.013991935729980469,
|
| 149 |
+
0.013936639785766602,
|
| 150 |
+
0.014241791725158692,
|
| 151 |
+
0.013945856094360352,
|
| 152 |
+
0.013956095695495606,
|
| 153 |
+
0.013971455574035644,
|
| 154 |
+
0.013947903633117676,
|
| 155 |
+
0.014299136161804199,
|
| 156 |
+
0.01436569595336914,
|
| 157 |
+
0.014169088363647461,
|
| 158 |
+
0.014048255920410157,
|
| 159 |
+
0.014057472229003906,
|
| 160 |
+
0.013931520462036133,
|
| 161 |
+
0.01386291217803955,
|
| 162 |
+
0.013939711570739746,
|
| 163 |
+
0.01371343994140625,
|
| 164 |
+
0.01421820831298828,
|
| 165 |
+
0.014449664115905762,
|
| 166 |
+
0.014491647720336913,
|
| 167 |
+
0.014784511566162109,
|
| 168 |
+
0.01489408016204834,
|
| 169 |
+
0.013894656181335448,
|
| 170 |
+
0.013721599578857421,
|
| 171 |
+
0.014261247634887696,
|
| 172 |
+
0.014329855918884277,
|
| 173 |
+
0.014292991638183594,
|
| 174 |
+
0.014353407859802245,
|
| 175 |
+
0.014359552383422852,
|
| 176 |
+
0.014307328224182129,
|
| 177 |
+
0.01429100799560547,
|
| 178 |
+
0.014792703628540039,
|
| 179 |
+
0.014342144012451171,
|
| 180 |
+
0.014360544204711913,
|
| 181 |
+
0.014297087669372559,
|
| 182 |
+
0.014299136161804199,
|
| 183 |
+
0.014363648414611817,
|
| 184 |
+
0.014320639610290528,
|
| 185 |
+
0.014285823822021485,
|
| 186 |
+
0.014332927703857423,
|
| 187 |
+
0.014326784133911133,
|
| 188 |
+
0.014334976196289062,
|
| 189 |
+
0.014330880165100097,
|
| 190 |
+
0.01425715160369873,
|
| 191 |
+
0.014276608467102051
|
| 192 |
]
|
| 193 |
},
|
| 194 |
"throughput": {
|
| 195 |
"unit": "samples/s",
|
| 196 |
+
"value": 69.86754692893906
|
| 197 |
},
|
| 198 |
"energy": {
|
| 199 |
"unit": "kWh",
|
| 200 |
+
"cpu": 1.6576348585479056e-07,
|
| 201 |
+
"ram": 9.058178719290784e-08,
|
| 202 |
+
"gpu": 3.3549023752777475e-07,
|
| 203 |
+
"total": 5.918355105754732e-07
|
| 204 |
},
|
| 205 |
"efficiency": {
|
| 206 |
"unit": "samples/kWh",
|
| 207 |
+
"value": 1689658.6671990107
|
| 208 |
}
|
| 209 |
}
|
| 210 |
}
|