IlyasMoutawwakil HF Staff commited on
Commit
bf60ffa
·
verified ·
1 Parent(s): 83f5034

Upload test_api_push_to_hub_mixin/benchmark.json with huggingface_hub

Browse files
test_api_push_to_hub_mixin/benchmark.json CHANGED
@@ -3,7 +3,7 @@
3
  "name": "test_api_push_to_hub_mixin",
4
  "backend": {
5
  "name": "pytorch",
6
- "version": "2.4.1+cpu",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "fill-mask",
9
  "library": "transformers",
@@ -66,16 +66,16 @@
66
  "start_method": "spawn"
67
  },
68
  "environment": {
69
- "cpu": " AMD EPYC 7763 64-Core Processor",
70
- "cpu_count": 4,
71
- "cpu_ram_mb": 16766.7712,
72
- "system": "Linux",
73
- "machine": "x86_64",
74
- "platform": "Linux-6.8.0-1014-azure-x86_64-with-glibc2.35",
75
- "processor": "x86_64",
76
- "python_version": "3.10.15",
77
  "optimum_benchmark_version": "0.4.0",
78
- "optimum_benchmark_commit": "4a884fa4ee31f882ca8d7e63513b8fff8137b30d",
79
  "transformers_version": "4.44.2",
80
  "transformers_commit": null,
81
  "accelerate_version": "0.34.2",
@@ -94,7 +94,7 @@
94
  "load": {
95
  "memory": {
96
  "unit": "MB",
97
- "max_ram": 551.555072,
98
  "max_global_vram": null,
99
  "max_process_vram": null,
100
  "max_reserved": null,
@@ -103,15 +103,15 @@
103
  "latency": {
104
  "unit": "s",
105
  "count": 1,
106
- "total": 4.3745375089999925,
107
- "mean": 4.3745375089999925,
108
  "stdev": 0.0,
109
- "p50": 4.3745375089999925,
110
- "p90": 4.3745375089999925,
111
- "p95": 4.3745375089999925,
112
- "p99": 4.3745375089999925,
113
  "values": [
114
- 4.3745375089999925
115
  ]
116
  },
117
  "throughput": null,
@@ -121,7 +121,7 @@
121
  "forward": {
122
  "memory": {
123
  "unit": "MB",
124
- "max_ram": 912.785408,
125
  "max_global_vram": null,
126
  "max_process_vram": null,
127
  "max_reserved": null,
@@ -129,32 +129,38 @@
129
  },
130
  "latency": {
131
  "unit": "s",
132
- "count": 12,
133
- "total": 1.0716282269999908,
134
- "mean": 0.08930235224999923,
135
- "stdev": 0.006814398695282422,
136
- "p50": 0.08921616799999299,
137
- "p90": 0.09688972920000652,
138
- "p95": 0.10085634700001264,
139
- "p99": 0.10415374220001923,
140
  "values": [
141
- 0.09748401100000592,
142
- 0.10497809100002087,
143
- 0.08763912700001697,
144
- 0.08931440599999974,
145
- 0.09154119300001184,
146
- 0.08911792999998625,
147
- 0.0906474459999913,
148
- 0.08968946299998493,
149
- 0.08802499600000147,
150
- 0.08550054599999157,
151
- 0.07893355299998461,
152
- 0.07875746499999536
 
 
 
 
 
 
153
  ]
154
  },
155
  "throughput": {
156
  "unit": "samples/s",
157
- "value": 22.395826645204828
158
  },
159
  "energy": null,
160
  "efficiency": null
 
3
  "name": "test_api_push_to_hub_mixin",
4
  "backend": {
5
  "name": "pytorch",
6
+ "version": "2.4.1",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "fill-mask",
9
  "library": "transformers",
 
66
  "start_method": "spawn"
67
  },
68
  "environment": {
69
+ "cpu": "Apple M1 (Virtual)",
70
+ "cpu_count": 3,
71
+ "cpu_ram_mb": 7516.192768,
72
+ "system": "Darwin",
73
+ "machine": "arm64",
74
+ "platform": "macOS-14.6.1-arm64-arm-64bit",
75
+ "processor": "arm",
76
+ "python_version": "3.12.6",
77
  "optimum_benchmark_version": "0.4.0",
78
+ "optimum_benchmark_commit": "96dfb7e6bb8c002c5fbc444701732aa424c1a4c3",
79
  "transformers_version": "4.44.2",
80
  "transformers_commit": null,
81
  "accelerate_version": "0.34.2",
 
94
  "load": {
95
  "memory": {
96
  "unit": "MB",
97
+ "max_ram": 527.532032,
98
  "max_global_vram": null,
99
  "max_process_vram": null,
100
  "max_reserved": null,
 
103
  "latency": {
104
  "unit": "s",
105
  "count": 1,
106
+ "total": 6.816807458999847,
107
+ "mean": 6.816807458999847,
108
  "stdev": 0.0,
109
+ "p50": 6.816807458999847,
110
+ "p90": 6.816807458999847,
111
+ "p95": 6.816807458999847,
112
+ "p99": 6.816807458999847,
113
  "values": [
114
+ 6.816807458999847
115
  ]
116
  },
117
  "throughput": null,
 
121
  "forward": {
122
  "memory": {
123
  "unit": "MB",
124
+ "max_ram": 933.052416,
125
  "max_global_vram": null,
126
  "max_process_vram": null,
127
  "max_reserved": null,
 
129
  },
130
  "latency": {
131
  "unit": "s",
132
+ "count": 18,
133
+ "total": 1.002908333999585,
134
+ "mean": 0.05571712966664361,
135
+ "stdev": 0.014691444976260123,
136
+ "p50": 0.04669187500007865,
137
+ "p90": 0.07825311710007554,
138
+ "p95": 0.08072692135028774,
139
+ "p99": 0.08407745067001086,
140
  "values": [
141
+ 0.08491508299994166,
142
+ 0.0574071669998375,
143
+ 0.07750966699995843,
144
+ 0.07462058300006902,
145
+ 0.07998783400034881,
146
+ 0.07558708300030048,
147
+ 0.054365916999813635,
148
+ 0.04563666700005342,
149
+ 0.0445503330001884,
150
+ 0.044178666999869165,
151
+ 0.0443248329997914,
152
+ 0.045784958000240294,
153
+ 0.0439992080000593,
154
+ 0.044452541999817186,
155
+ 0.04451904100005777,
156
+ 0.04759879199991701,
157
+ 0.04860704199973043,
158
+ 0.04486291699959111
159
  ]
160
  },
161
  "throughput": {
162
  "unit": "samples/s",
163
+ "value": 35.89560359562721
164
  },
165
  "energy": null,
166
  "efficiency": null