IlyasMoutawwakil HF Staff commited on
Commit
6bb3e6b
·
verified ·
1 Parent(s): 929af98

Upload test_api_push_to_hub_mixin/benchmark.json with huggingface_hub

Browse files
test_api_push_to_hub_mixin/benchmark.json CHANGED
@@ -6,17 +6,19 @@
6
  "version": "2.3.0+cpu",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "fill-mask",
9
- "library": "transformers",
10
  "model": "google-bert/bert-base-uncased",
11
- "processor": "google-bert/bert-base-uncased",
12
  "device": "cpu",
13
  "device_ids": null,
14
  "seed": 42,
15
  "inter_op_num_threads": null,
16
  "intra_op_num_threads": null,
17
- "model_kwargs": {},
18
- "processor_kwargs": {},
19
- "hub_kwargs": {},
 
 
 
20
  "no_weights": false,
21
  "device_map": null,
22
  "torch_dtype": null,
@@ -73,10 +75,10 @@
73
  "processor": "x86_64",
74
  "python_version": "3.10.14",
75
  "optimum_benchmark_version": "0.2.0",
76
- "optimum_benchmark_commit": "a8ccb91469272f9f38883fbf3e14bffcd5e95f2d",
77
  "transformers_version": "4.40.2",
78
  "transformers_commit": null,
79
- "accelerate_version": "0.30.0",
80
  "accelerate_commit": null,
81
  "diffusers_version": "0.27.2",
82
  "diffusers_commit": null,
@@ -92,7 +94,7 @@
92
  "forward": {
93
  "memory": {
94
  "unit": "MB",
95
- "max_ram": 897.122304,
96
  "max_global_vram": null,
97
  "max_process_vram": null,
98
  "max_reserved": null,
@@ -100,32 +102,28 @@
100
  },
101
  "latency": {
102
  "unit": "s",
103
- "count": 12,
104
- "total": 1.0749065910000581,
105
- "mean": 0.08957554925000484,
106
- "stdev": 0.0015155725450235792,
107
- "p50": 0.08915296599997191,
108
- "p90": 0.09120770909996737,
109
- "p95": 0.09193051570001672,
110
- "p99": 0.09262576234006929,
111
  "values": [
112
- 0.09279957400008243,
113
- 0.08876638599997477,
114
- 0.09066671699997642,
115
- 0.0891072910000048,
116
- 0.08736478200000874,
117
- 0.08919864099993902,
118
- 0.09121946799996294,
119
- 0.0911018790000071,
120
- 0.08905745699996714,
121
- 0.0883817880001061,
122
- 0.0894954240000061,
123
- 0.08774718400002257
124
  ]
125
  },
126
  "throughput": {
127
  "unit": "samples/s",
128
- "value": 22.327521480421087
129
  },
130
  "energy": null,
131
  "efficiency": null
 
6
  "version": "2.3.0+cpu",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "fill-mask",
 
9
  "model": "google-bert/bert-base-uncased",
10
+ "library": "transformers",
11
  "device": "cpu",
12
  "device_ids": null,
13
  "seed": 42,
14
  "inter_op_num_threads": null,
15
  "intra_op_num_threads": null,
16
+ "hub_kwargs": {
17
+ "revision": "main",
18
+ "force_download": false,
19
+ "local_files_only": false,
20
+ "trust_remote_code": false
21
+ },
22
  "no_weights": false,
23
  "device_map": null,
24
  "torch_dtype": null,
 
75
  "processor": "x86_64",
76
  "python_version": "3.10.14",
77
  "optimum_benchmark_version": "0.2.0",
78
+ "optimum_benchmark_commit": "b880a42960acf7600cf87a7b5756d96d29e2fb24",
79
  "transformers_version": "4.40.2",
80
  "transformers_commit": null,
81
+ "accelerate_version": "0.30.1",
82
  "accelerate_commit": null,
83
  "diffusers_version": "0.27.2",
84
  "diffusers_commit": null,
 
94
  "forward": {
95
  "memory": {
96
  "unit": "MB",
97
+ "max_ram": 895.71328,
98
  "max_global_vram": null,
99
  "max_process_vram": null,
100
  "max_reserved": null,
 
102
  },
103
  "latency": {
104
  "unit": "s",
105
+ "count": 8,
106
+ "total": 1.0463647430000265,
107
+ "mean": 0.1307955928750033,
108
+ "stdev": 0.03156207544168526,
109
+ "p50": 0.14421038000000408,
110
+ "p90": 0.15657725569999456,
111
+ "p95": 0.16415413234998935,
112
+ "p99": 0.1702156336699852,
113
  "values": [
114
+ 0.1255648250000263,
115
+ 0.17173100899998417,
116
+ 0.15008278999999902,
117
+ 0.15000067500000114,
118
+ 0.14976606400000492,
119
+ 0.13865469600000324,
120
+ 0.08082785400000603,
121
+ 0.07973683000000165
 
 
 
 
122
  ]
123
  },
124
  "throughput": {
125
  "unit": "samples/s",
126
+ "value": 15.29103508794313
127
  },
128
  "energy": null,
129
  "efficiency": null