IlyasMoutawwakil HF Staff commited on
Commit
05b8617
·
verified ·
1 Parent(s): b187c10

Upload cuda_inference_transformers_fill-mask_hf-internal-testing/tiny-random-BertModel/benchmark.json with huggingface_hub

Browse files
cuda_inference_transformers_fill-mask_hf-internal-testing/tiny-random-BertModel/benchmark.json ADDED
@@ -0,0 +1,537 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "config": {
3
+ "name": "cuda_inference_transformers_fill-mask_hf-internal-testing/tiny-random-BertModel",
4
+ "backend": {
5
+ "name": "pytorch",
6
+ "version": "2.5.1+cu124",
7
+ "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
8
+ "task": "fill-mask",
9
+ "library": "transformers",
10
+ "model_type": "bert",
11
+ "model": "hf-internal-testing/tiny-random-BertModel",
12
+ "processor": "hf-internal-testing/tiny-random-BertModel",
13
+ "device": "cuda",
14
+ "device_ids": "0",
15
+ "seed": 42,
16
+ "inter_op_num_threads": null,
17
+ "intra_op_num_threads": null,
18
+ "model_kwargs": {},
19
+ "processor_kwargs": {},
20
+ "no_weights": true,
21
+ "device_map": null,
22
+ "torch_dtype": null,
23
+ "eval_mode": true,
24
+ "to_bettertransformer": false,
25
+ "low_cpu_mem_usage": null,
26
+ "attn_implementation": null,
27
+ "cache_implementation": null,
28
+ "autocast_enabled": false,
29
+ "autocast_dtype": null,
30
+ "torch_compile": false,
31
+ "torch_compile_target": "forward",
32
+ "torch_compile_config": {},
33
+ "quantization_scheme": null,
34
+ "quantization_config": {},
35
+ "deepspeed_inference": false,
36
+ "deepspeed_inference_config": {},
37
+ "peft_type": null,
38
+ "peft_config": {}
39
+ },
40
+ "scenario": {
41
+ "name": "inference",
42
+ "_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario",
43
+ "iterations": 1,
44
+ "duration": 1,
45
+ "warmup_runs": 1,
46
+ "input_shapes": {
47
+ "batch_size": 2,
48
+ "sequence_length": 16,
49
+ "num_choices": 2
50
+ },
51
+ "new_tokens": null,
52
+ "memory": true,
53
+ "latency": true,
54
+ "energy": true,
55
+ "forward_kwargs": {},
56
+ "generate_kwargs": {
57
+ "max_new_tokens": 2,
58
+ "min_new_tokens": 2
59
+ },
60
+ "call_kwargs": {
61
+ "num_inference_steps": 2
62
+ }
63
+ },
64
+ "launcher": {
65
+ "name": "process",
66
+ "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
67
+ "device_isolation": true,
68
+ "device_isolation_action": "error",
69
+ "numactl": false,
70
+ "numactl_kwargs": {},
71
+ "start_method": "spawn"
72
+ },
73
+ "environment": {
74
+ "cpu": " AMD EPYC 7R32",
75
+ "cpu_count": 16,
76
+ "cpu_ram_mb": 66697.248768,
77
+ "system": "Linux",
78
+ "machine": "x86_64",
79
+ "platform": "Linux-5.10.228-219.884.amzn2.x86_64-x86_64-with-glibc2.35",
80
+ "processor": "x86_64",
81
+ "python_version": "3.10.12",
82
+ "gpu": [
83
+ "NVIDIA A10G"
84
+ ],
85
+ "gpu_count": 1,
86
+ "gpu_vram_mb": 24146608128,
87
+ "optimum_benchmark_version": "0.5.0.dev0",
88
+ "optimum_benchmark_commit": null,
89
+ "transformers_version": "4.47.0",
90
+ "transformers_commit": null,
91
+ "accelerate_version": "1.2.0",
92
+ "accelerate_commit": null,
93
+ "diffusers_version": "0.31.0",
94
+ "diffusers_commit": null,
95
+ "optimum_version": null,
96
+ "optimum_commit": null,
97
+ "timm_version": "1.0.12",
98
+ "timm_commit": null,
99
+ "peft_version": "0.14.0",
100
+ "peft_commit": null
101
+ },
102
+ "print_report": true,
103
+ "log_report": true
104
+ },
105
+ "report": {
106
+ "load_model": {
107
+ "memory": {
108
+ "unit": "MB",
109
+ "max_ram": 799.014912,
110
+ "max_global_vram": 639.107072,
111
+ "max_process_vram": 0.0,
112
+ "max_reserved": 2.097152,
113
+ "max_allocated": 0.534528
114
+ },
115
+ "latency": {
116
+ "unit": "s",
117
+ "values": [
118
+ 0.05666099166870117
119
+ ],
120
+ "count": 1,
121
+ "total": 0.05666099166870117,
122
+ "mean": 0.05666099166870117,
123
+ "p50": 0.05666099166870117,
124
+ "p90": 0.05666099166870117,
125
+ "p95": 0.05666099166870117,
126
+ "p99": 0.05666099166870117,
127
+ "stdev": 0,
128
+ "stdev_": 0
129
+ },
130
+ "throughput": null,
131
+ "energy": {
132
+ "unit": "kWh",
133
+ "cpu": 9.589206865416694e-05,
134
+ "ram": 5.253550660922821e-05,
135
+ "gpu": 0.00013195871667800313,
136
+ "total": 0.0002803862919413983
137
+ },
138
+ "efficiency": null
139
+ },
140
+ "forward": {
141
+ "memory": {
142
+ "unit": "MB",
143
+ "max_ram": 1038.31552,
144
+ "max_global_vram": 670.564352,
145
+ "max_process_vram": 0.0,
146
+ "max_reserved": 23.068672,
147
+ "max_allocated": 10.112
148
+ },
149
+ "latency": {
150
+ "unit": "s",
151
+ "values": [
152
+ 0.0028774399757385254,
153
+ 0.0026746881008148193,
154
+ 0.0027032959461212158,
155
+ 0.002702336072921753,
156
+ 0.0026603519916534426,
157
+ 0.0027934720516204835,
158
+ 0.0028108799457550048,
159
+ 0.0027740159034729006,
160
+ 0.002913280010223389,
161
+ 0.002944000005722046,
162
+ 0.0029378559589385987,
163
+ 0.002846719980239868,
164
+ 0.0029521920680999755,
165
+ 0.002916352033615112,
166
+ 0.0029020159244537353,
167
+ 0.002845695972442627,
168
+ 0.002942975997924805,
169
+ 0.0029061119556427002,
170
+ 0.002916352033615112,
171
+ 0.002826240062713623,
172
+ 0.002836479902267456,
173
+ 0.0027985920906066896,
174
+ 0.002775039911270142,
175
+ 0.0029716479778289797,
176
+ 0.002876415967941284,
177
+ 0.002862080097198486,
178
+ 0.002838464021682739,
179
+ 0.0028538880348205568,
180
+ 0.0028938241004943847,
181
+ 0.002820096015930176,
182
+ 0.002891776084899902,
183
+ 0.002942975997924805,
184
+ 0.002799616098403931,
185
+ 0.002615295886993408,
186
+ 0.0026387839317321775,
187
+ 0.002643968105316162,
188
+ 0.0026767361164093018,
189
+ 0.00275763201713562,
190
+ 0.0027535679340362548,
191
+ 0.002763808012008667,
192
+ 0.0028221440315246583,
193
+ 0.0027811839580535887,
194
+ 0.0027944960594177247,
195
+ 0.0027893760204315186,
196
+ 0.0028446719646453857,
197
+ 0.0028098559379577635,
198
+ 0.0027299840450286865,
199
+ 0.002810784101486206,
200
+ 0.0027893760204315186,
201
+ 0.002758591890335083,
202
+ 0.002845695972442627,
203
+ 0.002815999984741211,
204
+ 0.002778111934661865,
205
+ 0.002762752056121826,
206
+ 0.0028088319301605223,
207
+ 0.002826175928115845,
208
+ 0.002830336093902588,
209
+ 0.002944000005722046,
210
+ 0.0029470720291137694,
211
+ 0.0029992959499359133,
212
+ 0.002905087947845459,
213
+ 0.002942975997924805,
214
+ 0.002924544095993042,
215
+ 0.0030464000701904297,
216
+ 0.002905087947845459,
217
+ 0.0029480960369110106,
218
+ 0.002928704023361206,
219
+ 0.0029184000492095945,
220
+ 0.0028190720081329346,
221
+ 0.002850816011428833,
222
+ 0.002820096015930176,
223
+ 0.002748415946960449,
224
+ 0.0027146239280700685,
225
+ 0.002766848087310791,
226
+ 0.0029061119556427002,
227
+ 0.0028641281127929686,
228
+ 0.0029594240188598633,
229
+ 0.0028979198932647704,
230
+ 0.002894848108291626,
231
+ 0.0027545599937438964,
232
+ 0.002870271921157837,
233
+ 0.002853951930999756,
234
+ 0.002862080097198486,
235
+ 0.0029061119556427002,
236
+ 0.002871295928955078,
237
+ 0.0028241920471191407,
238
+ 0.002723839998245239,
239
+ 0.002748415946960449,
240
+ 0.0027596800327301025,
241
+ 0.0027381761074066163,
242
+ 0.002929663896560669,
243
+ 0.0028569600582122805,
244
+ 0.002860095977783203,
245
+ 0.002762752056121826,
246
+ 0.002819999933242798,
247
+ 0.0027944960594177247,
248
+ 0.0027351040840148926,
249
+ 0.0027822721004486083,
250
+ 0.0028180480003356934,
251
+ 0.0028784000873565674,
252
+ 0.002934783935546875,
253
+ 0.0029859519004821777,
254
+ 0.00301363205909729,
255
+ 0.002831360101699829,
256
+ 0.002706432104110718,
257
+ 0.002734015941619873,
258
+ 0.002791424036026001,
259
+ 0.0027217919826507567,
260
+ 0.0027607040405273437,
261
+ 0.0027351040840148926,
262
+ 0.0027442560195922853,
263
+ 0.002831360101699829,
264
+ 0.0029521920680999755,
265
+ 0.002939903974533081,
266
+ 0.002753535985946655,
267
+ 0.0028190720081329346,
268
+ 0.002787328004837036,
269
+ 0.0027432959079742433,
270
+ 0.002775039911270142,
271
+ 0.0027729918956756593,
272
+ 0.0030310399532318113,
273
+ 0.0028129279613494872,
274
+ 0.0027975680828094484,
275
+ 0.0027667839527130127,
276
+ 0.0027299840450286865,
277
+ 0.0028026878833770754,
278
+ 0.0027607040405273437,
279
+ 0.002707456111907959,
280
+ 0.002766848087310791,
281
+ 0.0028057599067687986,
282
+ 0.0028344318866729734,
283
+ 0.0028559360504150392,
284
+ 0.0029808640480041503,
285
+ 0.003023871898651123,
286
+ 0.0029777920246124266,
287
+ 0.0028897280693054197,
288
+ 0.002933759927749634,
289
+ 0.0029767680168151854,
290
+ 0.002958336114883423,
291
+ 0.0029081599712371827,
292
+ 0.002972671985626221,
293
+ 0.002861056089401245,
294
+ 0.0027402238845825196,
295
+ 0.0028231680393218995,
296
+ 0.0028897280693054197,
297
+ 0.002877408027648926,
298
+ 0.0028108799457550048,
299
+ 0.0028344318866729734,
300
+ 0.0027688961029052736,
301
+ 0.0026890239715576174,
302
+ 0.002746367931365967,
303
+ 0.0028518400192260743,
304
+ 0.0029061119556427002,
305
+ 0.0027135999202728273,
306
+ 0.0027985920906066896,
307
+ 0.0027381761074066163,
308
+ 0.002736128091812134,
309
+ 0.002770944118499756,
310
+ 0.00279967999458313,
311
+ 0.002714688062667847,
312
+ 0.0027555840015411376,
313
+ 0.002711551904678345,
314
+ 0.0027136640548706056,
315
+ 0.002778111934661865,
316
+ 0.002960383892059326,
317
+ 0.0028037118911743166,
318
+ 0.0026419200897216796,
319
+ 0.0029378559589385987,
320
+ 0.0028569600582122805,
321
+ 0.0027156479358673097,
322
+ 0.0026705920696258544,
323
+ 0.0027012479305267333,
324
+ 0.002741247892379761,
325
+ 0.002668544054031372,
326
+ 0.0026920959949493406,
327
+ 0.002732032060623169,
328
+ 0.002826240062713623,
329
+ 0.0027248640060424803,
330
+ 0.0026910719871521,
331
+ 0.002669568061828613,
332
+ 0.0027299840450286865,
333
+ 0.0028334081172943113,
334
+ 0.002954240083694458,
335
+ 0.002742271900177002,
336
+ 0.00276582407951355,
337
+ 0.0027392001152038575,
338
+ 0.0028098559379577635,
339
+ 0.0027893760204315186,
340
+ 0.003202048063278198,
341
+ 0.0034693119525909424,
342
+ 0.0029112319946289064,
343
+ 0.0029224960803985595,
344
+ 0.002936863899230957,
345
+ 0.002858047962188721,
346
+ 0.002894848108291626,
347
+ 0.0027504639625549315,
348
+ 0.003357696056365967,
349
+ 0.0027453439235687258,
350
+ 0.002663424015045166,
351
+ 0.002708479881286621,
352
+ 0.002698240041732788,
353
+ 0.002678783893585205,
354
+ 0.002712575912475586,
355
+ 0.0026972479820251464,
356
+ 0.002673664093017578,
357
+ 0.0027095038890838623,
358
+ 0.002698240041732788,
359
+ 0.0026920959949493406,
360
+ 0.002777087926864624,
361
+ 0.0027105278968811036,
362
+ 0.00268287992477417,
363
+ 0.0027586560249328613,
364
+ 0.002699264049530029,
365
+ 0.0026951680183410643,
366
+ 0.002775039911270142,
367
+ 0.002702336072921753,
368
+ 0.002753535985946655,
369
+ 0.0026634559631347654,
370
+ 0.0027033278942108154,
371
+ 0.002706432104110718,
372
+ 0.002643968105316162,
373
+ 0.0027095038890838623,
374
+ 0.002708479881286621,
375
+ 0.002669568061828613,
376
+ 0.0027279040813446044,
377
+ 0.0027545599937438964,
378
+ 0.0026470398902893066,
379
+ 0.0027187199592590334,
380
+ 0.002706432104110718,
381
+ 0.0026685121059417723,
382
+ 0.0027054080963134766,
383
+ 0.002699264049530029,
384
+ 0.0027043840885162354,
385
+ 0.0026603519916534426,
386
+ 0.002786303997039795,
387
+ 0.002708479881286621,
388
+ 0.0027135999202728273,
389
+ 0.0026961920261383055,
390
+ 0.0027105278968811036,
391
+ 0.0026736319065093993,
392
+ 0.0027105278968811036,
393
+ 0.002716671943664551,
394
+ 0.002708479881286621,
395
+ 0.0027904000282287598,
396
+ 0.002712575912475586,
397
+ 0.0026552319526672364,
398
+ 0.002722815990447998,
399
+ 0.002769920110702515,
400
+ 0.002696160078048706,
401
+ 0.002756608009338379,
402
+ 0.0026757760047912598,
403
+ 0.0027238719463348387,
404
+ 0.0027043840885162354,
405
+ 0.002678783893585205,
406
+ 0.0026818881034851075,
407
+ 0.0026746881008148193,
408
+ 0.0027637760639190674,
409
+ 0.0026746881008148193,
410
+ 0.0026593279838562013,
411
+ 0.002698240041732788,
412
+ 0.0026839039325714112,
413
+ 0.0026552319526672364,
414
+ 0.0027043840885162354,
415
+ 0.002673664093017578,
416
+ 0.0026818559169769288,
417
+ 0.002707456111907959,
418
+ 0.0026890239715576174,
419
+ 0.002662400007247925,
420
+ 0.0026910719871521,
421
+ 0.0026808319091796875,
422
+ 0.0026675200462341307,
423
+ 0.002687999963760376,
424
+ 0.0026808319091796875,
425
+ 0.002669568061828613,
426
+ 0.002678719997406006,
427
+ 0.0026716160774230956,
428
+ 0.0026920320987701414,
429
+ 0.002677759885787964,
430
+ 0.002717695951461792,
431
+ 0.002694144010543823,
432
+ 0.0026859519481658937,
433
+ 0.0026808319091796875,
434
+ 0.0026798079013824463,
435
+ 0.0026603519916534426,
436
+ 0.002748415946960449,
437
+ 0.0027648000717163087,
438
+ 0.0026961920261383055,
439
+ 0.0027156479358673097,
440
+ 0.0026818559169769288,
441
+ 0.002687999963760376,
442
+ 0.002717695951461792,
443
+ 0.0026951680183410643,
444
+ 0.0026849279403686525,
445
+ 0.0027002880573272704,
446
+ 0.0026900479793548586,
447
+ 0.0026715519428253173,
448
+ 0.0026961920261383055,
449
+ 0.0027207679748535154,
450
+ 0.0027156479358673097,
451
+ 0.002695199966430664,
452
+ 0.0026972160339355468,
453
+ 0.0027146239280700685,
454
+ 0.0026715519428253173,
455
+ 0.002678783893585205,
456
+ 0.002699264049530029,
457
+ 0.0027207679748535154,
458
+ 0.0027085120677948,
459
+ 0.0027648000717163087,
460
+ 0.0026757121086120605,
461
+ 0.002697184085845947,
462
+ 0.0026828160285949708,
463
+ 0.002663424015045166,
464
+ 0.002706432104110718,
465
+ 0.002686975955963135,
466
+ 0.002672640085220337,
467
+ 0.002703360080718994,
468
+ 0.0026839039325714112,
469
+ 0.0026849279403686525,
470
+ 0.0026859519481658937,
471
+ 0.0026859519481658937,
472
+ 0.0026705920696258544,
473
+ 0.0026818559169769288,
474
+ 0.00268287992477417,
475
+ 0.002698240041732788,
476
+ 0.0026818559169769288,
477
+ 0.0026900479793548586,
478
+ 0.0026859519481658937,
479
+ 0.0026798079013824463,
480
+ 0.0026859519481658937,
481
+ 0.002707456111907959,
482
+ 0.002673664093017578,
483
+ 0.002717695951461792,
484
+ 0.002698240041732788,
485
+ 0.0026859519481658937,
486
+ 0.0027136640548706056,
487
+ 0.0026798079013824463,
488
+ 0.0026808319091796875,
489
+ 0.0027248640060424803,
490
+ 0.00268287992477417,
491
+ 0.002671583890914917,
492
+ 0.002699264049530029,
493
+ 0.0027187199592590334,
494
+ 0.0026706559658050536,
495
+ 0.0026798079013824463,
496
+ 0.0026859519481658937,
497
+ 0.002668544054031372,
498
+ 0.0026839039325714112,
499
+ 0.0026859519481658937,
500
+ 0.002712575912475586,
501
+ 0.002663424015045166,
502
+ 0.0027075200080871583,
503
+ 0.0027074880599975587,
504
+ 0.0026798079013824463,
505
+ 0.0026818559169769288,
506
+ 0.002698240041732788,
507
+ 0.0026859519481658937,
508
+ 0.0026839039325714112
509
+ ],
510
+ "count": 357,
511
+ "total": 0.9891591353416446,
512
+ "mean": 0.0027707538805088073,
513
+ "p50": 0.0027351040840148926,
514
+ "p90": 0.0029233152866363524,
515
+ "p95": 0.002952601671218872,
516
+ "p99": 0.0030377984046936033,
517
+ "stdev": 0.00010735937395649541,
518
+ "stdev_": 3.8747351293713783
519
+ },
520
+ "throughput": {
521
+ "unit": "samples/s",
522
+ "value": 721.8252094021177
523
+ },
524
+ "energy": {
525
+ "unit": "kWh",
526
+ "cpu": 3.150399591089434e-08,
527
+ "ram": 1.7215031121280853e-08,
528
+ "gpu": 4.470304994680101e-08,
529
+ "total": 9.342207697897618e-08
530
+ },
531
+ "efficiency": {
532
+ "unit": "samples/kWh",
533
+ "value": 21408215.966447443
534
+ }
535
+ }
536
+ }
537
+ }