IlyasMoutawwakil HF Staff commited on
Commit
1167313
·
verified ·
1 Parent(s): 7ed35c3

Upload test_api_push_to_hub_mixin/benchmark.json with huggingface_hub

Browse files
test_api_push_to_hub_mixin/benchmark.json CHANGED
@@ -6,19 +6,17 @@
6
  "version": "2.3.0+cpu",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "fill-mask",
9
- "model": "google-bert/bert-base-uncased",
10
  "library": "transformers",
 
 
11
  "device": "cpu",
12
  "device_ids": null,
13
  "seed": 42,
14
  "inter_op_num_threads": null,
15
  "intra_op_num_threads": null,
16
- "hub_kwargs": {
17
- "revision": "main",
18
- "force_download": false,
19
- "local_files_only": false,
20
- "trust_remote_code": false
21
- },
22
  "no_weights": false,
23
  "device_map": null,
24
  "torch_dtype": null,
@@ -75,7 +73,7 @@
75
  "processor": "x86_64",
76
  "python_version": "3.10.14",
77
  "optimum_benchmark_version": "0.2.0",
78
- "optimum_benchmark_commit": "d35829e539df8480b726c647eeabf91e41eae047",
79
  "transformers_version": "4.40.2",
80
  "transformers_commit": null,
81
  "accelerate_version": "0.30.0",
@@ -94,7 +92,7 @@
94
  "forward": {
95
  "memory": {
96
  "unit": "MB",
97
- "max_ram": 896.79872,
98
  "max_global_vram": null,
99
  "max_process_vram": null,
100
  "max_reserved": null,
@@ -103,31 +101,31 @@
103
  "latency": {
104
  "unit": "s",
105
  "count": 12,
106
- "total": 1.069926570999968,
107
- "mean": 0.08916054758333065,
108
- "stdev": 0.0015660041696446366,
109
- "p50": 0.08896936699999003,
110
- "p90": 0.09100829570000428,
111
- "p95": 0.09166590930000496,
112
- "p99": 0.09226576746000603,
113
  "values": [
114
- 0.09061119500000814,
115
- 0.0924157320000063,
116
- 0.09007313099999692,
117
- 0.08780973899999367,
118
- 0.08691816399999652,
119
- 0.08893616999998244,
120
- 0.08818674100001545,
121
- 0.08950312799998983,
122
- 0.08900256399999762,
123
- 0.08793980999999462,
124
- 0.09105241800000385,
125
- 0.08747777899998255
126
  ]
127
  },
128
  "throughput": {
129
  "unit": "samples/s",
130
- "value": 22.43144590527299
131
  },
132
  "energy": null,
133
  "efficiency": null
 
6
  "version": "2.3.0+cpu",
7
  "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
  "task": "fill-mask",
 
9
  "library": "transformers",
10
+ "model": "google-bert/bert-base-uncased",
11
+ "processor": "google-bert/bert-base-uncased",
12
  "device": "cpu",
13
  "device_ids": null,
14
  "seed": 42,
15
  "inter_op_num_threads": null,
16
  "intra_op_num_threads": null,
17
+ "model_kwargs": {},
18
+ "processor_kwargs": {},
19
+ "hub_kwargs": {},
 
 
 
20
  "no_weights": false,
21
  "device_map": null,
22
  "torch_dtype": null,
 
73
  "processor": "x86_64",
74
  "python_version": "3.10.14",
75
  "optimum_benchmark_version": "0.2.0",
76
+ "optimum_benchmark_commit": "a8ccb91469272f9f38883fbf3e14bffcd5e95f2d",
77
  "transformers_version": "4.40.2",
78
  "transformers_commit": null,
79
  "accelerate_version": "0.30.0",
 
92
  "forward": {
93
  "memory": {
94
  "unit": "MB",
95
+ "max_ram": 897.122304,
96
  "max_global_vram": null,
97
  "max_process_vram": null,
98
  "max_reserved": null,
 
101
  "latency": {
102
  "unit": "s",
103
  "count": 12,
104
+ "total": 1.0749065910000581,
105
+ "mean": 0.08957554925000484,
106
+ "stdev": 0.0015155725450235792,
107
+ "p50": 0.08915296599997191,
108
+ "p90": 0.09120770909996737,
109
+ "p95": 0.09193051570001672,
110
+ "p99": 0.09262576234006929,
111
  "values": [
112
+ 0.09279957400008243,
113
+ 0.08876638599997477,
114
+ 0.09066671699997642,
115
+ 0.0891072910000048,
116
+ 0.08736478200000874,
117
+ 0.08919864099993902,
118
+ 0.09121946799996294,
119
+ 0.0911018790000071,
120
+ 0.08905745699996714,
121
+ 0.0883817880001061,
122
+ 0.0894954240000061,
123
+ 0.08774718400002257
124
  ]
125
  },
126
  "throughput": {
127
  "unit": "samples/s",
128
+ "value": 22.327521480421087
129
  },
130
  "energy": null,
131
  "efficiency": null