Upload cuda_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark_report.json with huggingface_hub
Browse files
cuda_inference_transformers_text-classification_FacebookAI/roberta-base/benchmark_report.json
CHANGED
|
@@ -2,152 +2,144 @@
|
|
| 2 |
"forward": {
|
| 3 |
"memory": {
|
| 4 |
"unit": "MB",
|
| 5 |
-
"max_ram":
|
| 6 |
-
"max_global_vram":
|
| 7 |
-
"max_process_vram":
|
| 8 |
"max_reserved": 555.74528,
|
| 9 |
"max_allocated": 499.443712
|
| 10 |
},
|
| 11 |
"latency": {
|
| 12 |
"unit": "s",
|
| 13 |
-
"count":
|
| 14 |
-
"total":
|
| 15 |
-
"mean": 0.
|
| 16 |
-
"stdev": 0.
|
| 17 |
-
"p50": 0.
|
| 18 |
-
"p90": 0.
|
| 19 |
-
"p95": 0.
|
| 20 |
-
"p99": 0.
|
| 21 |
"values": [
|
| 22 |
-
0.
|
| 23 |
-
0.
|
| 24 |
-
0.
|
| 25 |
-
0.
|
| 26 |
-
0.
|
| 27 |
-
0.
|
| 28 |
-
0.
|
| 29 |
-
0.
|
| 30 |
-
0.
|
| 31 |
-
0.
|
| 32 |
-
0.
|
| 33 |
-
0.
|
| 34 |
-
0.
|
| 35 |
-
0.
|
| 36 |
-
0.
|
| 37 |
-
0.
|
| 38 |
-
0.
|
| 39 |
-
0.
|
| 40 |
-
0.
|
| 41 |
-
0.
|
| 42 |
-
0.
|
| 43 |
-
0.
|
| 44 |
-
0.
|
| 45 |
-
0.
|
| 46 |
-
0.
|
| 47 |
-
0.
|
| 48 |
-
0.
|
| 49 |
-
0.
|
| 50 |
-
0.
|
| 51 |
-
0.
|
| 52 |
-
0.
|
| 53 |
-
0.
|
| 54 |
-
0.
|
| 55 |
-
0.
|
| 56 |
-
0.
|
| 57 |
-
0.
|
| 58 |
-
0.
|
| 59 |
-
0.
|
| 60 |
-
0.
|
| 61 |
-
0.
|
| 62 |
-
0.
|
| 63 |
-
0.
|
| 64 |
-
0.
|
| 65 |
-
0.
|
| 66 |
-
0.
|
| 67 |
-
0.
|
| 68 |
-
0.
|
| 69 |
-
0.
|
| 70 |
-
0.
|
| 71 |
-
0.
|
| 72 |
-
0.
|
| 73 |
-
0.
|
| 74 |
-
0.
|
| 75 |
-
0.
|
| 76 |
-
0.
|
| 77 |
-
0.
|
| 78 |
-
0.
|
| 79 |
-
0.
|
| 80 |
-
0.
|
| 81 |
-
0.
|
| 82 |
-
0.
|
| 83 |
-
0.
|
| 84 |
-
0.
|
| 85 |
-
0.
|
| 86 |
-
0.
|
| 87 |
-
0.
|
| 88 |
-
0.
|
| 89 |
-
0.
|
| 90 |
-
0.
|
| 91 |
-
0.
|
| 92 |
-
0.
|
| 93 |
-
0.
|
| 94 |
-
0.
|
| 95 |
-
0.
|
| 96 |
-
0.
|
| 97 |
-
0.
|
| 98 |
-
0.
|
| 99 |
-
0.
|
| 100 |
-
0.
|
| 101 |
-
0.
|
| 102 |
-
0.
|
| 103 |
-
0.
|
| 104 |
-
0.
|
| 105 |
-
0.
|
| 106 |
-
0.
|
| 107 |
-
0.
|
| 108 |
-
0.
|
| 109 |
-
0.
|
| 110 |
-
0.
|
| 111 |
-
0.
|
| 112 |
-
0.
|
| 113 |
-
0.
|
| 114 |
-
0.
|
| 115 |
-
0.
|
| 116 |
-
0.
|
| 117 |
-
0.
|
| 118 |
-
0.
|
| 119 |
-
0.
|
| 120 |
-
0.
|
| 121 |
-
0.
|
| 122 |
-
0.
|
| 123 |
-
0.
|
| 124 |
-
0.
|
| 125 |
-
0.
|
| 126 |
-
0.
|
| 127 |
-
0.
|
| 128 |
-
0.
|
| 129 |
-
0.
|
| 130 |
-
0.
|
| 131 |
-
0.
|
| 132 |
-
0.
|
| 133 |
-
0.
|
| 134 |
-
0.
|
| 135 |
-
0.
|
| 136 |
-
0.
|
| 137 |
-
0.
|
| 138 |
-
0.008022894859313965,
|
| 139 |
-
0.008049935340881347,
|
| 140 |
-
0.007988335132598877,
|
| 141 |
-
0.008012494087219239,
|
| 142 |
-
0.008027533531188964,
|
| 143 |
-
0.008008174896240235,
|
| 144 |
-
0.008003854751586914,
|
| 145 |
-
0.008044334411621093
|
| 146 |
]
|
| 147 |
},
|
| 148 |
"throughput": {
|
| 149 |
"unit": "samples/s",
|
| 150 |
-
"value":
|
| 151 |
},
|
| 152 |
"energy": null,
|
| 153 |
"efficiency": null
|
|
|
|
| 2 |
"forward": {
|
| 3 |
"memory": {
|
| 4 |
"unit": "MB",
|
| 5 |
+
"max_ram": 1011.109888,
|
| 6 |
+
"max_global_vram": 897.138688,
|
| 7 |
+
"max_process_vram": 190471.282688,
|
| 8 |
"max_reserved": 555.74528,
|
| 9 |
"max_allocated": 499.443712
|
| 10 |
},
|
| 11 |
"latency": {
|
| 12 |
"unit": "s",
|
| 13 |
+
"count": 116,
|
| 14 |
+
"total": 1.0012317423820498,
|
| 15 |
+
"mean": 0.008631308123983186,
|
| 16 |
+
"stdev": 0.0005121272580937299,
|
| 17 |
+
"p50": 0.00873374891281128,
|
| 18 |
+
"p90": 0.009061668395996094,
|
| 19 |
+
"p95": 0.009153548002243042,
|
| 20 |
+
"p99": 0.009307956743240357,
|
| 21 |
"values": [
|
| 22 |
+
0.008815988540649414,
|
| 23 |
+
0.009057588577270508,
|
| 24 |
+
0.009206229209899903,
|
| 25 |
+
0.009325908660888673,
|
| 26 |
+
0.009200148582458496,
|
| 27 |
+
0.009138869285583497,
|
| 28 |
+
0.009111828804016113,
|
| 29 |
+
0.008656788825988769,
|
| 30 |
+
0.008726869583129883,
|
| 31 |
+
0.008587828636169434,
|
| 32 |
+
0.008493590354919433,
|
| 33 |
+
0.008672148704528808,
|
| 34 |
+
0.008578229904174804,
|
| 35 |
+
0.008461588859558106,
|
| 36 |
+
0.008932788848876954,
|
| 37 |
+
0.008930869102478027,
|
| 38 |
+
0.008958229064941406,
|
| 39 |
+
0.008949589729309083,
|
| 40 |
+
0.008932310104370117,
|
| 41 |
+
0.009004948616027832,
|
| 42 |
+
0.00906574821472168,
|
| 43 |
+
0.00897150993347168,
|
| 44 |
+
0.008951990127563477,
|
| 45 |
+
0.008966549873352051,
|
| 46 |
+
0.00894766902923584,
|
| 47 |
+
0.00892798900604248,
|
| 48 |
+
0.008887349128723144,
|
| 49 |
+
0.008942388534545898,
|
| 50 |
+
0.00896078872680664,
|
| 51 |
+
0.00896574878692627,
|
| 52 |
+
0.008925588607788085,
|
| 53 |
+
0.00893390941619873,
|
| 54 |
+
0.008957908630371094,
|
| 55 |
+
0.00895966911315918,
|
| 56 |
+
0.008967028617858887,
|
| 57 |
+
0.008897748947143555,
|
| 58 |
+
0.00895470905303955,
|
| 59 |
+
0.00894606876373291,
|
| 60 |
+
0.008960148811340333,
|
| 61 |
+
0.008939189910888672,
|
| 62 |
+
0.008844788551330567,
|
| 63 |
+
0.009191669464111329,
|
| 64 |
+
0.009179828643798828,
|
| 65 |
+
0.00912782859802246,
|
| 66 |
+
0.009137748718261719,
|
| 67 |
+
0.009144787788391113,
|
| 68 |
+
0.00879086971282959,
|
| 69 |
+
0.011746705055236817,
|
| 70 |
+
0.008908948898315429,
|
| 71 |
+
0.00872799015045166,
|
| 72 |
+
0.008724629402160644,
|
| 73 |
+
0.008727828979492187,
|
| 74 |
+
0.008665109634399414,
|
| 75 |
+
0.008699989318847657,
|
| 76 |
+
0.008677749633789063,
|
| 77 |
+
0.008676790237426757,
|
| 78 |
+
0.008767349243164062,
|
| 79 |
+
0.00873694896697998,
|
| 80 |
+
0.008718870162963866,
|
| 81 |
+
0.008756308555603028,
|
| 82 |
+
0.008776948928833008,
|
| 83 |
+
0.008708149909973144,
|
| 84 |
+
0.00873854923248291,
|
| 85 |
+
0.008770709037780761,
|
| 86 |
+
0.00822959041595459,
|
| 87 |
+
0.00809807014465332,
|
| 88 |
+
0.008182550430297851,
|
| 89 |
+
0.007785271167755127,
|
| 90 |
+
0.00780767011642456,
|
| 91 |
+
0.008098710060119628,
|
| 92 |
+
0.007810710906982422,
|
| 93 |
+
0.008158069610595703,
|
| 94 |
+
0.007787350177764892,
|
| 95 |
+
0.007811670780181885,
|
| 96 |
+
0.007758230209350586,
|
| 97 |
+
0.007793750762939453,
|
| 98 |
+
0.007747350215911865,
|
| 99 |
+
0.008730548858642578,
|
| 100 |
+
0.008741909980773926,
|
| 101 |
+
0.008718548774719238,
|
| 102 |
+
0.008503990173339844,
|
| 103 |
+
0.00855902862548828,
|
| 104 |
+
0.008770869255065919,
|
| 105 |
+
0.00875102996826172,
|
| 106 |
+
0.008390549659729005,
|
| 107 |
+
0.008147351264953613,
|
| 108 |
+
0.008121271133422852,
|
| 109 |
+
0.00816415023803711,
|
| 110 |
+
0.00812911033630371,
|
| 111 |
+
0.008287830352783203,
|
| 112 |
+
0.008350230216979981,
|
| 113 |
+
0.007905110836029053,
|
| 114 |
+
0.008252630233764649,
|
| 115 |
+
0.00825199031829834,
|
| 116 |
+
0.008473270416259766,
|
| 117 |
+
0.008221590042114257,
|
| 118 |
+
0.008022390365600585,
|
| 119 |
+
0.00802655029296875,
|
| 120 |
+
0.00804175090789795,
|
| 121 |
+
0.007787990093231201,
|
| 122 |
+
0.007759671211242675,
|
| 123 |
+
0.00784191083908081,
|
| 124 |
+
0.007779190063476563,
|
| 125 |
+
0.0078039908409118655,
|
| 126 |
+
0.007913909912109374,
|
| 127 |
+
0.008536470413208009,
|
| 128 |
+
0.008741429328918457,
|
| 129 |
+
0.008822069168090821,
|
| 130 |
+
0.008717907905578613,
|
| 131 |
+
0.008715029716491698,
|
| 132 |
+
0.008763348579406739,
|
| 133 |
+
0.0087894287109375,
|
| 134 |
+
0.008597749710083008,
|
| 135 |
+
0.00880302906036377,
|
| 136 |
+
0.008765109062194824,
|
| 137 |
+
0.008743670463562012
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 138 |
]
|
| 139 |
},
|
| 140 |
"throughput": {
|
| 141 |
"unit": "samples/s",
|
| 142 |
+
"value": 115.85729366112803
|
| 143 |
},
|
| 144 |
"energy": null,
|
| 145 |
"efficiency": null
|