IlyasMoutawwakil HF Staff commited on
Commit
c2258cb
·
verified ·
1 Parent(s): eaf7bc2

Upload test_api_push_to_hub_mixin/benchmark.json with huggingface_hub

Browse files
test_api_push_to_hub_mixin/benchmark.json CHANGED
@@ -6,19 +6,17 @@
6
  "version": "2.3.0+cpu",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "fill-mask",
9
- "model": "google-bert/bert-base-uncased",
10
  "library": "transformers",
 
 
11
  "device": "cpu",
12
  "device_ids": null,
13
  "seed": 42,
14
  "inter_op_num_threads": null,
15
  "intra_op_num_threads": null,
16
- "hub_kwargs": {
17
- "revision": "main",
18
- "force_download": false,
19
- "local_files_only": false,
20
- "trust_remote_code": false
21
- },
22
  "no_weights": false,
23
  "device_map": null,
24
  "torch_dtype": null,
@@ -75,7 +73,7 @@
75
  "processor": "x86_64",
76
  "python_version": "3.10.14",
77
  "optimum_benchmark_version": "0.2.0",
78
- "optimum_benchmark_commit": "3e2eebdc0f80ae6deeb2e1faad3e889ed5a4df2d",
79
  "transformers_version": "4.40.2",
80
  "transformers_commit": null,
81
  "accelerate_version": "0.30.1",
@@ -94,7 +92,7 @@
94
  "forward": {
95
  "memory": {
96
  "unit": "MB",
97
- "max_ram": 896.34816,
98
  "max_global_vram": null,
99
  "max_process_vram": null,
100
  "max_reserved": null,
@@ -102,31 +100,32 @@
102
  },
103
  "latency": {
104
  "unit": "s",
105
- "count": 11,
106
- "total": 1.0083977129999653,
107
- "mean": 0.0916725193636332,
108
- "stdev": 0.0008547829088302858,
109
- "p50": 0.09192783699998586,
110
- "p90": 0.09260690100001057,
111
- "p95": 0.09275233200000343,
112
- "p99": 0.09286867679999773,
113
  "values": [
114
- 0.09204590599998141,
115
- 0.09135427900000082,
116
- 0.09148013199998672,
117
- 0.09022596699998076,
118
- 0.08998157299998866,
119
- 0.09192783699998586,
120
- 0.09169161700000927,
121
- 0.0928977629999963,
122
- 0.09260690100001057,
123
- 0.09200789600001258,
124
- 0.09217784200001233
 
125
  ]
126
  },
127
  "throughput": {
128
  "unit": "samples/s",
129
- "value": 21.81678886850149
130
  },
131
  "energy": null,
132
  "efficiency": null
 
6
  "version": "2.3.0+cpu",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "fill-mask",
 
9
  "library": "transformers",
10
+ "model": "google-bert/bert-base-uncased",
11
+ "processor": "google-bert/bert-base-uncased",
12
  "device": "cpu",
13
  "device_ids": null,
14
  "seed": 42,
15
  "inter_op_num_threads": null,
16
  "intra_op_num_threads": null,
17
+ "model_kwargs": {},
18
+ "processor_kwargs": {},
19
+ "hub_kwargs": {},
 
 
 
20
  "no_weights": false,
21
  "device_map": null,
22
  "torch_dtype": null,
 
73
  "processor": "x86_64",
74
  "python_version": "3.10.14",
75
  "optimum_benchmark_version": "0.2.0",
76
+ "optimum_benchmark_commit": "6fd377459e287bb09e9383ba2516b1b2a271a562",
77
  "transformers_version": "4.40.2",
78
  "transformers_commit": null,
79
  "accelerate_version": "0.30.1",
 
92
  "forward": {
93
  "memory": {
94
  "unit": "MB",
95
+ "max_ram": 895.823872,
96
  "max_global_vram": null,
97
  "max_process_vram": null,
98
  "max_reserved": null,
 
100
  },
101
  "latency": {
102
  "unit": "s",
103
+ "count": 12,
104
+ "total": 1.0651276760000314,
105
+ "mean": 0.08876063966666929,
106
+ "stdev": 0.0012597758822470333,
107
+ "p50": 0.08915893999999014,
108
+ "p90": 0.08984839350000584,
109
+ "p95": 0.09005040515000218,
110
+ "p99": 0.09024240662999858,
111
  "values": [
112
+ 0.09029040699999769,
113
+ 0.08917816599998218,
114
+ 0.08979757500000574,
115
+ 0.0883613969999999,
116
+ 0.08694348500000615,
117
+ 0.08975327200002425,
118
+ 0.08944578700001671,
119
+ 0.08985404000000585,
120
+ 0.0891397139999981,
121
+ 0.08769171399998754,
122
+ 0.08874988400000916,
123
+ 0.08592223499999818
124
  ]
125
  },
126
  "throughput": {
127
  "unit": "samples/s",
128
+ "value": 22.532509989909688
129
  },
130
  "energy": null,
131
  "efficiency": null