Upload cpu_training_transformers_text-generation_hf-internal-testing/tiny-random-LlamaForCausalLM/benchmark.json with huggingface_hub
Browse files
cpu_training_transformers_text-generation_hf-internal-testing/tiny-random-LlamaForCausalLM/benchmark.json
CHANGED
|
@@ -5,11 +5,11 @@
|
|
| 5 |
"name": "pytorch",
|
| 6 |
"version": "2.5.1+cpu",
|
| 7 |
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
|
|
|
|
|
|
|
| 8 |
"task": "text-generation",
|
| 9 |
"library": "transformers",
|
| 10 |
"model_type": "llama",
|
| 11 |
-
"model": "hf-internal-testing/tiny-random-LlamaForCausalLM",
|
| 12 |
-
"processor": "hf-internal-testing/tiny-random-LlamaForCausalLM",
|
| 13 |
"device": "cpu",
|
| 14 |
"device_ids": null,
|
| 15 |
"seed": 42,
|
|
@@ -86,10 +86,10 @@
|
|
| 86 |
"processor": "x86_64",
|
| 87 |
"python_version": "3.10.15",
|
| 88 |
"optimum_benchmark_version": "0.5.0.dev0",
|
| 89 |
-
"optimum_benchmark_commit": "
|
| 90 |
"transformers_version": "4.47.0",
|
| 91 |
"transformers_commit": null,
|
| 92 |
-
"accelerate_version": "1.2.
|
| 93 |
"accelerate_commit": null,
|
| 94 |
"diffusers_version": "0.31.0",
|
| 95 |
"diffusers_commit": null,
|
|
@@ -107,7 +107,7 @@
|
|
| 107 |
"overall": {
|
| 108 |
"memory": {
|
| 109 |
"unit": "MB",
|
| 110 |
-
"max_ram":
|
| 111 |
"max_global_vram": null,
|
| 112 |
"max_process_vram": null,
|
| 113 |
"max_reserved": null,
|
|
@@ -116,42 +116,42 @@
|
|
| 116 |
"latency": {
|
| 117 |
"unit": "s",
|
| 118 |
"values": [
|
| 119 |
-
0.
|
| 120 |
-
0.
|
| 121 |
-
0.
|
| 122 |
-
0.
|
| 123 |
-
0.
|
| 124 |
],
|
| 125 |
"count": 5,
|
| 126 |
-
"total": 0.
|
| 127 |
-
"mean": 0.
|
| 128 |
-
"p50": 0.
|
| 129 |
-
"p90": 0.
|
| 130 |
-
"p95": 0.
|
| 131 |
-
"p99": 0.
|
| 132 |
-
"stdev": 0.
|
| 133 |
-
"stdev_":
|
| 134 |
},
|
| 135 |
"throughput": {
|
| 136 |
"unit": "samples/s",
|
| 137 |
-
"value": 741.
|
| 138 |
},
|
| 139 |
"energy": {
|
| 140 |
"unit": "kWh",
|
| 141 |
-
"cpu": 8.
|
| 142 |
-
"ram": 3.
|
| 143 |
"gpu": 0,
|
| 144 |
-
"total":
|
| 145 |
},
|
| 146 |
"efficiency": {
|
| 147 |
"unit": "samples/kWh",
|
| 148 |
-
"value":
|
| 149 |
}
|
| 150 |
},
|
| 151 |
"warmup": {
|
| 152 |
"memory": {
|
| 153 |
"unit": "MB",
|
| 154 |
-
"max_ram":
|
| 155 |
"max_global_vram": null,
|
| 156 |
"max_process_vram": null,
|
| 157 |
"max_reserved": null,
|
|
@@ -160,22 +160,22 @@
|
|
| 160 |
"latency": {
|
| 161 |
"unit": "s",
|
| 162 |
"values": [
|
| 163 |
-
0.
|
| 164 |
-
0.
|
| 165 |
],
|
| 166 |
"count": 2,
|
| 167 |
-
"total": 0.
|
| 168 |
-
"mean": 0.
|
| 169 |
-
"p50": 0.
|
| 170 |
-
"p90": 0.
|
| 171 |
-
"p95": 0.
|
| 172 |
-
"p99": 0.
|
| 173 |
-
"stdev": 0.
|
| 174 |
-
"stdev_":
|
| 175 |
},
|
| 176 |
"throughput": {
|
| 177 |
"unit": "samples/s",
|
| 178 |
-
"value":
|
| 179 |
},
|
| 180 |
"energy": null,
|
| 181 |
"efficiency": null
|
|
@@ -183,7 +183,7 @@
|
|
| 183 |
"train": {
|
| 184 |
"memory": {
|
| 185 |
"unit": "MB",
|
| 186 |
-
"max_ram":
|
| 187 |
"max_global_vram": null,
|
| 188 |
"max_process_vram": null,
|
| 189 |
"max_reserved": null,
|
|
@@ -192,23 +192,23 @@
|
|
| 192 |
"latency": {
|
| 193 |
"unit": "s",
|
| 194 |
"values": [
|
| 195 |
-
0.
|
| 196 |
-
0.
|
| 197 |
-
0.
|
| 198 |
],
|
| 199 |
"count": 3,
|
| 200 |
-
"total": 0.
|
| 201 |
-
"mean": 0.
|
| 202 |
-
"p50": 0.
|
| 203 |
-
"p90": 0.
|
| 204 |
-
"p95": 0.
|
| 205 |
-
"p99": 0.
|
| 206 |
-
"stdev":
|
| 207 |
-
"stdev_":
|
| 208 |
},
|
| 209 |
"throughput": {
|
| 210 |
"unit": "samples/s",
|
| 211 |
-
"value":
|
| 212 |
},
|
| 213 |
"energy": null,
|
| 214 |
"efficiency": null
|
|
|
|
| 5 |
"name": "pytorch",
|
| 6 |
"version": "2.5.1+cpu",
|
| 7 |
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
|
| 8 |
+
"model": "hf-internal-testing/tiny-random-LlamaForCausalLM",
|
| 9 |
+
"processor": "hf-internal-testing/tiny-random-LlamaForCausalLM",
|
| 10 |
"task": "text-generation",
|
| 11 |
"library": "transformers",
|
| 12 |
"model_type": "llama",
|
|
|
|
|
|
|
| 13 |
"device": "cpu",
|
| 14 |
"device_ids": null,
|
| 15 |
"seed": 42,
|
|
|
|
| 86 |
"processor": "x86_64",
|
| 87 |
"python_version": "3.10.15",
|
| 88 |
"optimum_benchmark_version": "0.5.0.dev0",
|
| 89 |
+
"optimum_benchmark_commit": "7cec62e016d76fe612308e4c2c074fc7f09289fd",
|
| 90 |
"transformers_version": "4.47.0",
|
| 91 |
"transformers_commit": null,
|
| 92 |
+
"accelerate_version": "1.2.1",
|
| 93 |
"accelerate_commit": null,
|
| 94 |
"diffusers_version": "0.31.0",
|
| 95 |
"diffusers_commit": null,
|
|
|
|
| 107 |
"overall": {
|
| 108 |
"memory": {
|
| 109 |
"unit": "MB",
|
| 110 |
+
"max_ram": 522.866688,
|
| 111 |
"max_global_vram": null,
|
| 112 |
"max_process_vram": null,
|
| 113 |
"max_reserved": null,
|
|
|
|
| 116 |
"latency": {
|
| 117 |
"unit": "s",
|
| 118 |
"values": [
|
| 119 |
+
0.019453390999984776,
|
| 120 |
+
0.012075559000010117,
|
| 121 |
+
0.011965802999981179,
|
| 122 |
+
0.011934053999993921,
|
| 123 |
+
0.011987785000002305
|
| 124 |
],
|
| 125 |
"count": 5,
|
| 126 |
+
"total": 0.0674165919999723,
|
| 127 |
+
"mean": 0.013483318399994459,
|
| 128 |
+
"p50": 0.011987785000002305,
|
| 129 |
+
"p90": 0.016502258199994912,
|
| 130 |
+
"p95": 0.017977824599989842,
|
| 131 |
+
"p99": 0.01915827771998579,
|
| 132 |
+
"stdev": 0.0029854060594256856,
|
| 133 |
+
"stdev_": 22.141478609797662
|
| 134 |
},
|
| 135 |
"throughput": {
|
| 136 |
"unit": "samples/s",
|
| 137 |
+
"value": 741.6571872992415
|
| 138 |
},
|
| 139 |
"energy": {
|
| 140 |
"unit": "kWh",
|
| 141 |
+
"cpu": 8.703254777777798e-06,
|
| 142 |
+
"ram": 3.635053251718738e-07,
|
| 143 |
"gpu": 0,
|
| 144 |
+
"total": 9.066760102949672e-06
|
| 145 |
},
|
| 146 |
"efficiency": {
|
| 147 |
"unit": "samples/kWh",
|
| 148 |
+
"value": 1102929.8102578802
|
| 149 |
}
|
| 150 |
},
|
| 151 |
"warmup": {
|
| 152 |
"memory": {
|
| 153 |
"unit": "MB",
|
| 154 |
+
"max_ram": 522.866688,
|
| 155 |
"max_global_vram": null,
|
| 156 |
"max_process_vram": null,
|
| 157 |
"max_reserved": null,
|
|
|
|
| 160 |
"latency": {
|
| 161 |
"unit": "s",
|
| 162 |
"values": [
|
| 163 |
+
0.019453390999984776,
|
| 164 |
+
0.012075559000010117
|
| 165 |
],
|
| 166 |
"count": 2,
|
| 167 |
+
"total": 0.03152894999999489,
|
| 168 |
+
"mean": 0.015764474999997447,
|
| 169 |
+
"p50": 0.015764474999997447,
|
| 170 |
+
"p90": 0.01871560779998731,
|
| 171 |
+
"p95": 0.019084499399986042,
|
| 172 |
+
"p99": 0.01937961267998503,
|
| 173 |
+
"stdev": 0.00368891599998733,
|
| 174 |
+
"stdev_": 23.400183006334988
|
| 175 |
},
|
| 176 |
"throughput": {
|
| 177 |
"unit": "samples/s",
|
| 178 |
+
"value": 253.73505936611576
|
| 179 |
},
|
| 180 |
"energy": null,
|
| 181 |
"efficiency": null
|
|
|
|
| 183 |
"train": {
|
| 184 |
"memory": {
|
| 185 |
"unit": "MB",
|
| 186 |
+
"max_ram": 522.866688,
|
| 187 |
"max_global_vram": null,
|
| 188 |
"max_process_vram": null,
|
| 189 |
"max_reserved": null,
|
|
|
|
| 192 |
"latency": {
|
| 193 |
"unit": "s",
|
| 194 |
"values": [
|
| 195 |
+
0.011965802999981179,
|
| 196 |
+
0.011934053999993921,
|
| 197 |
+
0.011987785000002305
|
| 198 |
],
|
| 199 |
"count": 3,
|
| 200 |
+
"total": 0.035887641999977404,
|
| 201 |
+
"mean": 0.011962547333325801,
|
| 202 |
+
"p50": 0.011965802999981179,
|
| 203 |
+
"p90": 0.01198338859999808,
|
| 204 |
+
"p95": 0.011985586800000192,
|
| 205 |
+
"p99": 0.011987345360001882,
|
| 206 |
+
"stdev": 2.2056059099537384e-05,
|
| 207 |
+
"stdev_": 0.184375940048259
|
| 208 |
},
|
| 209 |
"throughput": {
|
| 210 |
"unit": "samples/s",
|
| 211 |
+
"value": 501.56541352065796
|
| 212 |
},
|
| 213 |
"energy": null,
|
| 214 |
"efficiency": null
|