rocm / cuda_inference_transformers_fill-mask_hf-internal-testing /tiny-random-BertModel /benchmark.json
IlyasMoutawwakil HF Staff
Upload cuda_inference_transformers_fill-mask_hf-internal-testing/tiny-random-BertModel/benchmark.json with huggingface_hub
ac090b0 verified | { | |
| "config": { | |
| "name": "cuda_inference_transformers_fill-mask_hf-internal-testing/tiny-random-BertModel", | |
| "backend": { | |
| "name": "pytorch", | |
| "version": "2.7.0+rocm6.3", | |
| "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", | |
| "model": "hf-internal-testing/tiny-random-BertModel", | |
| "processor": "hf-internal-testing/tiny-random-BertModel", | |
| "task": "fill-mask", | |
| "library": "transformers", | |
| "model_type": "bert", | |
| "device": "cuda", | |
| "device_ids": "5", | |
| "seed": 42, | |
| "inter_op_num_threads": null, | |
| "intra_op_num_threads": null, | |
| "model_kwargs": {}, | |
| "processor_kwargs": {}, | |
| "no_weights": true, | |
| "tp_plan": null, | |
| "device_map": null, | |
| "torch_dtype": null, | |
| "eval_mode": true, | |
| "to_bettertransformer": false, | |
| "low_cpu_mem_usage": null, | |
| "attn_implementation": null, | |
| "cache_implementation": null, | |
| "allow_tf32": false, | |
| "autocast_enabled": false, | |
| "autocast_dtype": null, | |
| "torch_compile": false, | |
| "torch_compile_target": "forward", | |
| "torch_compile_config": {}, | |
| "quantization_scheme": null, | |
| "quantization_config": {}, | |
| "deepspeed_inference": false, | |
| "deepspeed_inference_config": {}, | |
| "peft_type": null, | |
| "peft_config": {} | |
| }, | |
| "scenario": { | |
| "name": "inference", | |
| "_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario", | |
| "iterations": 1, | |
| "duration": 1, | |
| "warmup_runs": 1, | |
| "input_shapes": { | |
| "batch_size": 2, | |
| "sequence_length": 16, | |
| "num_choices": 2 | |
| }, | |
| "new_tokens": null, | |
| "memory": true, | |
| "latency": true, | |
| "energy": false, | |
| "forward_kwargs": {}, | |
| "generate_kwargs": { | |
| "max_new_tokens": 2, | |
| "min_new_tokens": 2 | |
| }, | |
| "call_kwargs": { | |
| "num_inference_steps": 2 | |
| } | |
| }, | |
| "launcher": { | |
| "name": "process", | |
| "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", | |
| "device_isolation": true, | |
| "device_isolation_action": "warn", | |
| "numactl": false, | |
| "numactl_kwargs": {}, | |
| "start_method": "spawn" | |
| }, | |
| "environment": { | |
| "cpu": " AMD EPYC 7763 64-Core Processor", | |
| "cpu_count": 128, | |
| "cpu_ram_mb": 1082014.486528, | |
| "system": "Linux", | |
| "machine": "x86_64", | |
| "platform": "Linux-5.15.0-131-generic-x86_64-with-glibc2.35", | |
| "processor": "x86_64", | |
| "python_version": "3.10.12", | |
| "gpu": [ | |
| "Advanced Micro Devices, Inc. [AMD/ATI]", | |
| "Advanced Micro Devices, Inc. [AMD/ATI]", | |
| "Advanced Micro Devices, Inc. [AMD/ATI]", | |
| "Advanced Micro Devices, Inc. [AMD/ATI]", | |
| "Advanced Micro Devices, Inc. [AMD/ATI]", | |
| "Advanced Micro Devices, Inc. [AMD/ATI]", | |
| "Advanced Micro Devices, Inc. [AMD/ATI]", | |
| "Advanced Micro Devices, Inc. [AMD/ATI]" | |
| ], | |
| "gpu_count": 8, | |
| "gpu_vram_mb": 549621596160, | |
| "optimum_benchmark_version": "0.6.0.dev0", | |
| "optimum_benchmark_commit": null, | |
| "transformers_version": "4.51.3", | |
| "transformers_commit": null, | |
| "accelerate_version": "1.6.0", | |
| "accelerate_commit": null, | |
| "diffusers_version": "0.33.1", | |
| "diffusers_commit": null, | |
| "optimum_version": null, | |
| "optimum_commit": null, | |
| "timm_version": "1.0.15", | |
| "timm_commit": null, | |
| "peft_version": null, | |
| "peft_commit": null | |
| }, | |
| "print_report": true, | |
| "log_report": true | |
| }, | |
| "report": { | |
| "load_model": { | |
| "memory": { | |
| "unit": "MB", | |
| "max_ram": 1936.900096, | |
| "max_global_vram": 68702.69952, | |
| "max_process_vram": 0.0, | |
| "max_reserved": 2.097152, | |
| "max_allocated": 0.389632 | |
| }, | |
| "latency": { | |
| "unit": "s", | |
| "values": [ | |
| 0.3582965393066406 | |
| ], | |
| "count": 1, | |
| "total": 0.3582965393066406, | |
| "mean": 0.3582965393066406, | |
| "p50": 0.3582965393066406, | |
| "p90": 0.3582965393066406, | |
| "p95": 0.3582965393066406, | |
| "p99": 0.3582965393066406, | |
| "stdev": 0, | |
| "stdev_": 0 | |
| }, | |
| "throughput": null, | |
| "energy": null, | |
| "efficiency": null | |
| }, | |
| "first_forward": { | |
| "memory": { | |
| "unit": "MB", | |
| "max_ram": 2148.06528, | |
| "max_global_vram": 68702.69952, | |
| "max_process_vram": 0.0, | |
| "max_reserved": 81.788928, | |
| "max_allocated": 80.23552 | |
| }, | |
| "latency": { | |
| "unit": "s", | |
| "values": [ | |
| 0.5060724792480469 | |
| ], | |
| "count": 1, | |
| "total": 0.5060724792480469, | |
| "mean": 0.5060724792480469, | |
| "p50": 0.5060724792480469, | |
| "p90": 0.5060724792480469, | |
| "p95": 0.5060724792480469, | |
| "p99": 0.5060724792480469, | |
| "stdev": 0, | |
| "stdev_": 0 | |
| }, | |
| "throughput": null, | |
| "energy": null, | |
| "efficiency": null | |
| }, | |
| "forward": { | |
| "memory": { | |
| "unit": "MB", | |
| "max_ram": 2148.33152, | |
| "max_global_vram": 68702.69952, | |
| "max_process_vram": 0.0, | |
| "max_reserved": 81.788928, | |
| "max_allocated": 80.23552 | |
| }, | |
| "latency": { | |
| "unit": "s", | |
| "values": [ | |
| 0.0035619139671325684, | |
| 0.0032092750072479246, | |
| 0.0033617539405822756, | |
| 0.0032183949947357177, | |
| 0.003224955081939697, | |
| 0.0032247951030731203, | |
| 0.0032268741130828857, | |
| 0.0032043149471282957, | |
| 0.0031886351108551024, | |
| 0.003199515104293823, | |
| 0.003427354097366333, | |
| 0.003284795045852661, | |
| 0.0032150349617004396, | |
| 0.003422394037246704, | |
| 0.0032201550006866453, | |
| 0.003520633935928345, | |
| 0.0032639949321746826, | |
| 0.0033323140144348143, | |
| 0.0032423949241638184, | |
| 0.003186875104904175, | |
| 0.003223035097122192, | |
| 0.003278074026107788, | |
| 0.003448154926300049, | |
| 0.003337913990020752, | |
| 0.0032087950706481934, | |
| 0.0031846349239349364, | |
| 0.003364794969558716, | |
| 0.003325434923171997, | |
| 0.0032049551010131837, | |
| 0.003334233999252319, | |
| 0.0033807950019836427, | |
| 0.0034135940074920654, | |
| 0.003416475057601929, | |
| 0.003416954040527344, | |
| 0.0034113550186157227, | |
| 0.0033630340099334716, | |
| 0.0032739150524139404, | |
| 0.00344991397857666, | |
| 0.0034884750843048097, | |
| 0.0032327940464019775, | |
| 0.0032243149280548095, | |
| 0.0033804750442504882, | |
| 0.0032236740589141845, | |
| 0.003223675012588501, | |
| 0.0032097549438476563, | |
| 0.003219515085220337, | |
| 0.0032073540687561036, | |
| 0.0032273550033569335, | |
| 0.0032140750885009768, | |
| 0.003376955032348633, | |
| 0.0032019150257110596, | |
| 0.0032078349590301514, | |
| 0.0032171149253845217, | |
| 0.0032254350185394287, | |
| 0.003228473901748657, | |
| 0.003209594964981079, | |
| 0.003225754976272583, | |
| 0.0032343950271606445, | |
| 0.0032457540035247803, | |
| 0.0032303950786590576, | |
| 0.0032167949676513673, | |
| 0.0032153549194335936, | |
| 0.003223994016647339, | |
| 0.0032630350589752197, | |
| 0.003225754976272583, | |
| 0.0032275149822235108, | |
| 0.0032043139934539795, | |
| 0.003592794895172119, | |
| 0.0036587150096893313, | |
| 0.003645914077758789, | |
| 0.0036484739780426024, | |
| 0.0036463940143585203, | |
| 0.0036767940521240236, | |
| 0.0036412739753723146, | |
| 0.0036385540962219237, | |
| 0.0036625540256500244, | |
| 0.003642714023590088, | |
| 0.003639833927154541, | |
| 0.003625754117965698, | |
| 0.0036523139476776125, | |
| 0.00363279390335083, | |
| 0.003612795114517212, | |
| 0.0036235148906707763, | |
| 0.0036331140995025635, | |
| 0.0036452739238739015, | |
| 0.003636473894119263, | |
| 0.0036529541015625, | |
| 0.003633913993835449, | |
| 0.0036435139179229734, | |
| 0.0036231939792633057, | |
| 0.0036148738861083982, | |
| 0.003630393981933594, | |
| 0.0036345551013946533, | |
| 0.003644474983215332, | |
| 0.00366335391998291, | |
| 0.003651834011077881, | |
| 0.0036401538848876954, | |
| 0.0036355140209197996, | |
| 0.0036332740783691407, | |
| 0.003653274059295654, | |
| 0.0036371140480041505, | |
| 0.0036326339244842527, | |
| 0.003639833927154541, | |
| 0.003647994041442871, | |
| 0.0036390340328216553, | |
| 0.0036596739292144775, | |
| 0.0036563150882720948, | |
| 0.0036455950736999513, | |
| 0.0036644740104675292, | |
| 0.00365679407119751, | |
| 0.0036452739238739015, | |
| 0.0036542339324951174, | |
| 0.0036547141075134277, | |
| 0.0036812739372253417, | |
| 0.0036513540744781492, | |
| 0.0036425540447235106, | |
| 0.003625593900680542, | |
| 0.0036363139152526857, | |
| 0.0036465539932250975, | |
| 0.0036433539390563967, | |
| 0.003661113977432251, | |
| 0.003647834062576294, | |
| 0.003650394916534424, | |
| 0.003672795057296753, | |
| 0.0036406340599060057, | |
| 0.003655993938446045, | |
| 0.0036473538875579836, | |
| 0.0036583940982818602, | |
| 0.003629434108734131, | |
| 0.0036302340030670167, | |
| 0.0036350340843200684, | |
| 0.0036363139152526857, | |
| 0.0036289539337158202, | |
| 0.003637913942337036, | |
| 0.003650873899459839, | |
| 0.0036438350677490236, | |
| 0.0036478350162506105, | |
| 0.0036375939846038817, | |
| 0.0036382339000701905, | |
| 0.00362735390663147, | |
| 0.0036439940929412842, | |
| 0.003643834114074707, | |
| 0.0036259140968322753, | |
| 0.003644953966140747, | |
| 0.003626713991165161, | |
| 0.003653913974761963, | |
| 0.003633434057235718, | |
| 0.0036399950981140135, | |
| 0.003626075029373169, | |
| 0.0036471939086914064, | |
| 0.0036463940143585203, | |
| 0.0036435139179229734, | |
| 0.0036300740242004395, | |
| 0.0036332740783691407, | |
| 0.0037127940654754638, | |
| 0.0036492741107940676, | |
| 0.0036473538875579836, | |
| 0.0036571140289306642, | |
| 0.00366223406791687, | |
| 0.00367151403427124, | |
| 0.0036435139179229734, | |
| 0.0036526339054107665, | |
| 0.003642394065856934, | |
| 0.0036532750129699707, | |
| 0.0036622350215911865, | |
| 0.003660794973373413, | |
| 0.0036582338809967043, | |
| 0.0036431939601898194, | |
| 0.003650873899459839, | |
| 0.003643834114074707, | |
| 0.0036495940685272216, | |
| 0.0036382339000701905, | |
| 0.003647994041442871, | |
| 0.0036444740295410154, | |
| 0.0036521539688110353, | |
| 0.003637434005737305, | |
| 0.0036355140209197996, | |
| 0.0036617538928985596, | |
| 0.003635354995727539, | |
| 0.003634394884109497, | |
| 0.003664154052734375, | |
| 0.0036444740295410154, | |
| 0.003653274059295654, | |
| 0.0036371140480041505, | |
| 0.0036473538875579836, | |
| 0.0036787140369415284, | |
| 0.0036598339080810547, | |
| 0.0036529541015625, | |
| 0.0036393539905548097, | |
| 0.0036372740268707277, | |
| 0.003653274059295654, | |
| 0.0036585540771484374, | |
| 0.003634553909301758, | |
| 0.003642394065856934, | |
| 0.003644315004348755, | |
| 0.0036449549198150637, | |
| 0.0036446340084075927, | |
| 0.0036551940441131594, | |
| 0.0036745541095733643, | |
| 0.003649914026260376, | |
| 0.003647834062576294, | |
| 0.003645914077758789, | |
| 0.0036321539878845215, | |
| 0.0036467139720916747, | |
| 0.003630393981933594, | |
| 0.003648953914642334, | |
| 0.0036230340003967285, | |
| 0.0036601541042327883, | |
| 0.0036428749561309816, | |
| 0.0036761550903320313, | |
| 0.0036555140018463134, | |
| 0.0036492741107940676, | |
| 0.0036460740566253663, | |
| 0.0036691141128540037, | |
| 0.0036385540962219237, | |
| 0.0036452739238739015, | |
| 0.003641913890838623, | |
| 0.0036076738834381105, | |
| 0.0036215939521789552, | |
| 0.003629434108734131, | |
| 0.003616473913192749, | |
| 0.0036387150287628175, | |
| 0.003634874105453491, | |
| 0.00364095401763916, | |
| 0.003630873918533325, | |
| 0.003641753911972046, | |
| 0.003619833946228027, | |
| 0.00362735390663147, | |
| 0.00362351393699646, | |
| 0.0036137540340423586, | |
| 0.003628633975982666, | |
| 0.0036177549362182617, | |
| 0.0036395139694213865, | |
| 0.0036393539905548097, | |
| 0.0036131141185760498, | |
| 0.0036119940280914305, | |
| 0.0036420741081237794, | |
| 0.003631514072418213, | |
| 0.003619513988494873, | |
| 0.0036105539798736573, | |
| 0.0036409549713134768, | |
| 0.003623353958129883, | |
| 0.0035972740650177, | |
| 0.0036185541152954103, | |
| 0.003625593900680542, | |
| 0.003507034063339233, | |
| 0.003594715118408203, | |
| 0.0036257550716400147, | |
| 0.003547194004058838, | |
| 0.003558233976364136, | |
| 0.0035489540100097656, | |
| 0.0034569549560546876, | |
| 0.003447033882141113, | |
| 0.0034689550399780275, | |
| 0.0035065550804138186, | |
| 0.003565593957901001, | |
| 0.003527833938598633, | |
| 0.003447355031967163, | |
| 0.003613274097442627, | |
| 0.003535994052886963, | |
| 0.0035235140323638917, | |
| 0.0036118340492248533, | |
| 0.0034833550453186035, | |
| 0.0035742340087890627, | |
| 0.003471513986587524, | |
| 0.003582715034484863, | |
| 0.0034659130573272707, | |
| 0.0036220738887786865, | |
| 0.0034380741119384764, | |
| 0.003482875108718872, | |
| 0.0035457539558410643, | |
| 0.003481113910675049, | |
| 0.0035433549880981444, | |
| 0.0034788739681243896, | |
| 0.00341823410987854, | |
| 0.003597434997558594, | |
| 0.0036441540718078614, | |
| 0.003470233917236328, | |
| 0.003653913974761963, | |
| 0.0034443149566650393 | |
| ], | |
| "count": 280, | |
| "total": 0.9916500844955454, | |
| "mean": 0.003541607444626944, | |
| "p50": 0.0036327139139175416, | |
| "p90": 0.00365602605342865, | |
| "p95": 0.003662234115600586, | |
| "p99": 0.0036771972489356994, | |
| "stdev": 0.0001598697844641318, | |
| "stdev_": 4.514045866564743 | |
| }, | |
| "throughput": { | |
| "unit": "samples/s", | |
| "value": 564.7153252499079 | |
| }, | |
| "energy": null, | |
| "efficiency": null | |
| } | |
| } | |
| } |