Upload cuda_inference_transformers_token-classification_microsoft/deberta-v3-base/benchmark_report.json with huggingface_hub
Browse files
cuda_inference_transformers_token-classification_microsoft/deberta-v3-base/benchmark_report.json
CHANGED
|
@@ -2,7 +2,7 @@
|
|
| 2 |
"forward": {
|
| 3 |
"memory": {
|
| 4 |
"unit": "MB",
|
| 5 |
-
"max_ram":
|
| 6 |
"max_global_vram": 1434.976256,
|
| 7 |
"max_process_vram": 0.0,
|
| 8 |
"max_reserved": 794.820608,
|
|
@@ -10,106 +10,101 @@
|
|
| 10 |
},
|
| 11 |
"latency": {
|
| 12 |
"unit": "s",
|
| 13 |
-
"count":
|
| 14 |
-
"total": 1.
|
| 15 |
-
"mean": 0.
|
| 16 |
-
"stdev": 0.
|
| 17 |
-
"p50": 0.
|
| 18 |
-
"p90": 0.
|
| 19 |
-
"p95": 0.
|
| 20 |
-
"p99": 0.
|
| 21 |
"values": [
|
| 22 |
-
0.
|
| 23 |
-
0.
|
| 24 |
-
0.
|
| 25 |
-
0.
|
| 26 |
-
0.
|
| 27 |
-
0.
|
| 28 |
-
0.
|
| 29 |
-
0.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 30 |
0.013684736251831055,
|
| 31 |
-
0.
|
| 32 |
-
0.
|
| 33 |
-
0.
|
| 34 |
-
0.
|
| 35 |
-
0.
|
| 36 |
-
0.
|
| 37 |
-
0.
|
| 38 |
-
0.
|
| 39 |
-
0.
|
| 40 |
-
0.
|
| 41 |
-
0.
|
| 42 |
-
0.
|
| 43 |
-
0.
|
| 44 |
-
0.013972479820251465,
|
| 45 |
-
0.013706239700317382,
|
| 46 |
-
0.013917183876037598,
|
| 47 |
-
0.01376972770690918,
|
| 48 |
-
0.013674495697021484,
|
| 49 |
-
0.01375641632080078,
|
| 50 |
-
0.013767680168151856,
|
| 51 |
-
0.013882368087768555,
|
| 52 |
-
0.013604864120483399,
|
| 53 |
-
0.013857791900634766,
|
| 54 |
-
0.013698047637939453,
|
| 55 |
-
0.013813759803771973,
|
| 56 |
-
0.014121983528137207,
|
| 57 |
-
0.014346240043640136,
|
| 58 |
-
0.013826047897338867,
|
| 59 |
-
0.01372054386138916,
|
| 60 |
-
0.013690879821777344,
|
| 61 |
-
0.013595647811889648,
|
| 62 |
-
0.01369600009918213,
|
| 63 |
-
0.014807040214538575,
|
| 64 |
-
0.01598259162902832,
|
| 65 |
-
0.013386752128601074,
|
| 66 |
-
0.013181952476501465,
|
| 67 |
-
0.013023232460021973,
|
| 68 |
-
0.012668928146362305,
|
| 69 |
-
0.012690431594848632,
|
| 70 |
-
0.01273036766052246,
|
| 71 |
-
0.012975104331970215,
|
| 72 |
-
0.012660736083984376,
|
| 73 |
-
0.012609536170959473,
|
| 74 |
-
0.01266483211517334,
|
| 75 |
-
0.012606464385986327,
|
| 76 |
-
0.012621824264526366,
|
| 77 |
-
0.012969984054565429,
|
| 78 |
-
0.012698623657226562,
|
| 79 |
-
0.012641280174255372,
|
| 80 |
-
0.012712960243225097,
|
| 81 |
-
0.01260540771484375,
|
| 82 |
-
0.012668928146362305,
|
| 83 |
-
0.012727295875549317,
|
| 84 |
-
0.012659711837768555,
|
| 85 |
-
0.012614656448364257,
|
| 86 |
-
0.012609536170959473,
|
| 87 |
-
0.012731391906738282,
|
| 88 |
-
0.012596223831176758,
|
| 89 |
-
0.012621824264526366,
|
| 90 |
-
0.012692480087280274,
|
| 91 |
-
0.012637184143066407,
|
| 92 |
-
0.012582912445068359,
|
| 93 |
-
0.012596223831176758,
|
| 94 |
-
0.012613632202148438,
|
| 95 |
-
0.012593152046203614,
|
| 96 |
-
0.012577792167663575
|
| 97 |
]
|
| 98 |
},
|
| 99 |
"throughput": {
|
| 100 |
"unit": "samples/s",
|
| 101 |
-
"value":
|
| 102 |
},
|
| 103 |
"energy": {
|
| 104 |
"unit": "kWh",
|
| 105 |
-
"cpu": 1.
|
| 106 |
-
"ram":
|
| 107 |
-
"gpu": 3.
|
| 108 |
-
"total": 5.
|
| 109 |
},
|
| 110 |
"efficiency": {
|
| 111 |
"unit": "samples/kWh",
|
| 112 |
-
"value":
|
| 113 |
}
|
| 114 |
}
|
| 115 |
}
|
|
|
|
| 2 |
"forward": {
|
| 3 |
"memory": {
|
| 4 |
"unit": "MB",
|
| 5 |
+
"max_ram": 975.151104,
|
| 6 |
"max_global_vram": 1434.976256,
|
| 7 |
"max_process_vram": 0.0,
|
| 8 |
"max_reserved": 794.820608,
|
|
|
|
| 10 |
},
|
| 11 |
"latency": {
|
| 12 |
"unit": "s",
|
| 13 |
+
"count": 70,
|
| 14 |
+
"total": 1.0027070407867433,
|
| 15 |
+
"mean": 0.014324386296953474,
|
| 16 |
+
"stdev": 0.0007211542333931602,
|
| 17 |
+
"p50": 0.014160352230072022,
|
| 18 |
+
"p90": 0.015174377822875977,
|
| 19 |
+
"p95": 0.015343468952178954,
|
| 20 |
+
"p99": 0.017134304504394533,
|
| 21 |
"values": [
|
| 22 |
+
0.01528217601776123,
|
| 23 |
+
0.015981568336486816,
|
| 24 |
+
0.015334527969360351,
|
| 25 |
+
0.014557184219360352,
|
| 26 |
+
0.014098431587219238,
|
| 27 |
+
0.014656512260437012,
|
| 28 |
+
0.014072832107543945,
|
| 29 |
+
0.014032896041870118,
|
| 30 |
+
0.014147520065307617,
|
| 31 |
+
0.014495583534240722,
|
| 32 |
+
0.014276608467102051,
|
| 33 |
+
0.014178303718566895,
|
| 34 |
+
0.015162400245666504,
|
| 35 |
+
0.017354751586914064,
|
| 36 |
+
0.0147958402633667,
|
| 37 |
+
0.014825471878051758,
|
| 38 |
+
0.01470464038848877,
|
| 39 |
+
0.01472000026702881,
|
| 40 |
+
0.014593024253845215,
|
| 41 |
+
0.014302207946777343,
|
| 42 |
+
0.015350784301757812,
|
| 43 |
+
0.01479587173461914,
|
| 44 |
+
0.014662495613098145,
|
| 45 |
+
0.01459609603881836,
|
| 46 |
+
0.01482960033416748,
|
| 47 |
+
0.014487680435180663,
|
| 48 |
+
0.01406668758392334,
|
| 49 |
+
0.014143487930297852,
|
| 50 |
+
0.014358400344848633,
|
| 51 |
+
0.014075008392333985,
|
| 52 |
+
0.014101504325866699,
|
| 53 |
+
0.014173184394836426,
|
| 54 |
+
0.014194687843322755,
|
| 55 |
+
0.014568448066711426,
|
| 56 |
+
0.014457759857177734,
|
| 57 |
+
0.01429196834564209,
|
| 58 |
+
0.014202879905700684,
|
| 59 |
+
0.014183423995971679,
|
| 60 |
+
0.01394380760192871,
|
| 61 |
+
0.0140830717086792,
|
| 62 |
+
0.014041088104248046,
|
| 63 |
+
0.013773823738098144,
|
| 64 |
+
0.01420083236694336,
|
| 65 |
+
0.014611455917358398,
|
| 66 |
+
0.01528217601776123,
|
| 67 |
+
0.017035263061523438,
|
| 68 |
+
0.01501696014404297,
|
| 69 |
+
0.014133248329162598,
|
| 70 |
+
0.013983743667602539,
|
| 71 |
+
0.013650943756103515,
|
| 72 |
+
0.01363046360015869,
|
| 73 |
+
0.013845503807067871,
|
| 74 |
+
0.013901887893676758,
|
| 75 |
+
0.013549568176269532,
|
| 76 |
+
0.013486080169677735,
|
| 77 |
+
0.013709312438964843,
|
| 78 |
0.013684736251831055,
|
| 79 |
+
0.013677536010742187,
|
| 80 |
+
0.013662336349487304,
|
| 81 |
+
0.014077887535095215,
|
| 82 |
+
0.013945856094360352,
|
| 83 |
+
0.013668352127075196,
|
| 84 |
+
0.013616127967834473,
|
| 85 |
+
0.013607935905456543,
|
| 86 |
+
0.01358847999572754,
|
| 87 |
+
0.013616127967834473,
|
| 88 |
+
0.013620223999023438,
|
| 89 |
+
0.013693951606750488,
|
| 90 |
+
0.013613056182861329,
|
| 91 |
+
0.013642751693725585
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 92 |
]
|
| 93 |
},
|
| 94 |
"throughput": {
|
| 95 |
"unit": "samples/s",
|
| 96 |
+
"value": 69.81101872494747
|
| 97 |
},
|
| 98 |
"energy": {
|
| 99 |
"unit": "kWh",
|
| 100 |
+
"cpu": 1.6979205939504838e-07,
|
| 101 |
+
"ram": 9.24548919517422e-08,
|
| 102 |
+
"gpu": 3.294645492857122e-07,
|
| 103 |
+
"total": 5.917115006325028e-07
|
| 104 |
},
|
| 105 |
"efficiency": {
|
| 106 |
"unit": "samples/kWh",
|
| 107 |
+
"value": 1690012.7831401995
|
| 108 |
}
|
| 109 |
}
|
| 110 |
}
|