IlyasMoutawwakil HF Staff commited on
Commit
1fd9a7f
·
verified ·
1 Parent(s): a3f6af2

Upload cpu_inference_timm_image-classification_timm/tiny_vit_21m_224.in1k/benchmark.json with huggingface_hub

Browse files
cpu_inference_timm_image-classification_timm/tiny_vit_21m_224.in1k/benchmark.json CHANGED
@@ -3,7 +3,7 @@
3
  "name": "cpu_inference_timm_image-classification_timm/tiny_vit_21m_224.in1k",
4
  "backend": {
5
  "name": "pytorch",
6
- "version": "2.6.0+cpu",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "model": "timm/tiny_vit_21m_224.in1k",
9
  "processor": "timm/tiny_vit_21m_224.in1k",
@@ -74,14 +74,14 @@
74
  "environment": {
75
  "cpu": " AMD EPYC 7763 64-Core Processor",
76
  "cpu_count": 4,
77
- "cpu_ram_mb": 16766.767104,
78
  "system": "Linux",
79
  "machine": "x86_64",
80
- "platform": "Linux-6.8.0-1021-azure-x86_64-with-glibc2.39",
81
  "processor": "x86_64",
82
  "python_version": "3.10.17",
83
  "optimum_benchmark_version": "0.6.0.dev0",
84
- "optimum_benchmark_commit": "bbca4c0833dddd655a709821a3b4cc93af36a0c7",
85
  "transformers_version": "4.51.3",
86
  "transformers_commit": null,
87
  "accelerate_version": "1.6.0",
@@ -102,7 +102,7 @@
102
  "load_model": {
103
  "memory": {
104
  "unit": "MB",
105
- "max_ram": 556.740608,
106
  "max_global_vram": null,
107
  "max_process_vram": null,
108
  "max_reserved": null,
@@ -111,32 +111,32 @@
111
  "latency": {
112
  "unit": "s",
113
  "values": [
114
- 0.34889640500000496
115
  ],
116
  "count": 1,
117
- "total": 0.34889640500000496,
118
- "mean": 0.34889640500000496,
119
- "p50": 0.34889640500000496,
120
- "p90": 0.34889640500000496,
121
- "p95": 0.34889640500000496,
122
- "p99": 0.34889640500000496,
123
  "stdev": 0,
124
  "stdev_": 0
125
  },
126
  "throughput": null,
127
  "energy": {
128
  "unit": "kWh",
129
- "cpu": 9.904781866264484e-05,
130
- "ram": 1.2615052791666731e-05,
131
  "gpu": 0,
132
- "total": 0.00011166287145431157
133
  },
134
  "efficiency": null
135
  },
136
  "first_forward": {
137
  "memory": {
138
  "unit": "MB",
139
- "max_ram": 621.142016,
140
  "max_global_vram": null,
141
  "max_process_vram": null,
142
  "max_reserved": null,
@@ -145,32 +145,32 @@
145
  "latency": {
146
  "unit": "s",
147
  "values": [
148
- 0.2032233099999985
149
  ],
150
  "count": 1,
151
- "total": 0.2032233099999985,
152
- "mean": 0.2032233099999985,
153
- "p50": 0.2032233099999985,
154
- "p90": 0.2032233099999985,
155
- "p95": 0.2032233099999985,
156
- "p99": 0.2032233099999985,
157
  "stdev": 0,
158
  "stdev_": 0
159
  },
160
  "throughput": null,
161
  "energy": {
162
  "unit": "kWh",
163
- "cpu": 0.00010435700463346643,
164
- "ram": 1.224325018333335e-05,
165
  "gpu": 0,
166
- "total": 0.00011660025481679981
167
  },
168
  "efficiency": null
169
  },
170
  "forward": {
171
  "memory": {
172
  "unit": "MB",
173
- "max_ram": 618.917888,
174
  "max_global_vram": null,
175
  "max_process_vram": null,
176
  "max_reserved": null,
@@ -179,37 +179,37 @@
179
  "latency": {
180
  "unit": "s",
181
  "values": [
182
- 0.1837770529999716,
183
- 0.18208510100004105,
184
- 0.18971868299996686,
185
- 0.18047952199998463,
186
- 0.18084043999999722,
187
- 0.18329008699998894
188
  ],
189
  "count": 6,
190
- "total": 1.1001908859999503,
191
- "mean": 0.18336514766665837,
192
- "p50": 0.182687594000015,
193
- "p90": 0.18674786799996923,
194
- "p95": 0.18823327549996804,
195
- "p99": 0.1894216014999671,
196
- "stdev": 0.0030790208156131435,
197
- "stdev_": 1.6791745076935398
198
  },
199
  "throughput": {
200
  "unit": "samples/s",
201
- "value": 10.907198153248965
202
  },
203
  "energy": {
204
  "unit": "kWh",
205
- "cpu": 3.443586855242688e-06,
206
- "ram": 1.0289179768518701e-06,
207
  "gpu": 0.0,
208
- "total": 4.472504832094561e-06
209
  },
210
  "efficiency": {
211
  "unit": "samples/kWh",
212
- "value": 447176.7108328335
213
  }
214
  }
215
  }
 
3
  "name": "cpu_inference_timm_image-classification_timm/tiny_vit_21m_224.in1k",
4
  "backend": {
5
  "name": "pytorch",
6
+ "version": "2.7.0+cpu",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "model": "timm/tiny_vit_21m_224.in1k",
9
  "processor": "timm/tiny_vit_21m_224.in1k",
 
74
  "environment": {
75
  "cpu": " AMD EPYC 7763 64-Core Processor",
76
  "cpu_count": 4,
77
+ "cpu_ram_mb": 16766.410752,
78
  "system": "Linux",
79
  "machine": "x86_64",
80
+ "platform": "Linux-6.11.0-1012-azure-x86_64-with-glibc2.39",
81
  "processor": "x86_64",
82
  "python_version": "3.10.17",
83
  "optimum_benchmark_version": "0.6.0.dev0",
84
+ "optimum_benchmark_commit": "78d6fd7c05e4903fc03580f46bd45ef32863e57b",
85
  "transformers_version": "4.51.3",
86
  "transformers_commit": null,
87
  "accelerate_version": "1.6.0",
 
102
  "load_model": {
103
  "memory": {
104
  "unit": "MB",
105
+ "max_ram": 567.64416,
106
  "max_global_vram": null,
107
  "max_process_vram": null,
108
  "max_reserved": null,
 
111
  "latency": {
112
  "unit": "s",
113
  "values": [
114
+ 0.3302246190000062
115
  ],
116
  "count": 1,
117
+ "total": 0.3302246190000062,
118
+ "mean": 0.3302246190000062,
119
+ "p50": 0.3302246190000062,
120
+ "p90": 0.3302246190000062,
121
+ "p95": 0.3302246190000062,
122
+ "p99": 0.3302246190000062,
123
  "stdev": 0,
124
  "stdev_": 0
125
  },
126
  "throughput": null,
127
  "energy": {
128
  "unit": "kWh",
129
+ "cpu": 0.00010332769150371126,
130
+ "ram": 1.3536714641666711e-05,
131
  "gpu": 0,
132
+ "total": 0.00011686440614537797
133
  },
134
  "efficiency": null
135
  },
136
  "first_forward": {
137
  "memory": {
138
  "unit": "MB",
139
+ "max_ram": 625.385472,
140
  "max_global_vram": null,
141
  "max_process_vram": null,
142
  "max_reserved": null,
 
145
  "latency": {
146
  "unit": "s",
147
  "values": [
148
+ 0.20437252300001774
149
  ],
150
  "count": 1,
151
+ "total": 0.20437252300001774,
152
+ "mean": 0.20437252300001774,
153
+ "p50": 0.20437252300001774,
154
+ "p90": 0.20437252300001774,
155
+ "p95": 0.20437252300001774,
156
+ "p99": 0.20437252300001774,
157
  "stdev": 0,
158
  "stdev_": 0
159
  },
160
  "throughput": null,
161
  "energy": {
162
  "unit": "kWh",
163
+ "cpu": 0.00010321843247781316,
164
+ "ram": 1.320275437222228e-05,
165
  "gpu": 0,
166
+ "total": 0.00011642118685003545
167
  },
168
  "efficiency": null
169
  },
170
  "forward": {
171
  "memory": {
172
  "unit": "MB",
173
+ "max_ram": 649.17504,
174
  "max_global_vram": null,
175
  "max_process_vram": null,
176
  "max_reserved": null,
 
179
  "latency": {
180
  "unit": "s",
181
  "values": [
182
+ 0.19776902399996743,
183
+ 0.1938801699999999,
184
+ 0.19240756299996065,
185
+ 0.1965839929999902,
186
+ 0.19184013500000674,
187
+ 0.19152314299998352
188
  ],
189
  "count": 6,
190
+ "total": 1.1640040279999084,
191
+ "mean": 0.19400067133331808,
192
+ "p50": 0.19314386649998028,
193
+ "p90": 0.1971765084999788,
194
+ "p95": 0.19747276624997312,
195
+ "p99": 0.19770977244996857,
196
+ "stdev": 0.0023886052861073243,
197
+ "stdev_": 1.2312355775323032
198
  },
199
  "throughput": {
200
  "unit": "samples/s",
201
+ "value": 10.309242675576844
202
  },
203
  "energy": {
204
  "unit": "kWh",
205
+ "cpu": 4.0583520669998465e-06,
206
+ "ram": 1.0734528592592524e-06,
207
  "gpu": 0.0,
208
+ "total": 5.131804926259089e-06
209
  },
210
  "efficiency": {
211
  "unit": "samples/kWh",
212
+ "value": 389726.42739519174
213
  }
214
  }
215
  }