local / cuda_inference_transformers_fill-mask_hf-internal-testing /tiny-random-BertModel /benchmark.json
IlyasMoutawwakil HF Staff
Upload cuda_inference_transformers_fill-mask_hf-internal-testing/tiny-random-BertModel/benchmark.json with huggingface_hub
2d44cd7 verified | { | |
| "config": { | |
| "name": "cuda_inference_transformers_fill-mask_hf-internal-testing/tiny-random-BertModel", | |
| "backend": { | |
| "name": "pytorch", | |
| "version": "2.5.1", | |
| "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", | |
| "model": "hf-internal-testing/tiny-random-BertModel", | |
| "processor": "hf-internal-testing/tiny-random-BertModel", | |
| "task": "fill-mask", | |
| "library": "transformers", | |
| "model_type": "bert", | |
| "device": "cuda", | |
| "device_ids": "0", | |
| "seed": 42, | |
| "inter_op_num_threads": null, | |
| "intra_op_num_threads": null, | |
| "model_kwargs": {}, | |
| "processor_kwargs": {}, | |
| "no_weights": true, | |
| "tp_plan": null, | |
| "device_map": null, | |
| "torch_dtype": null, | |
| "eval_mode": true, | |
| "to_bettertransformer": false, | |
| "low_cpu_mem_usage": null, | |
| "attn_implementation": null, | |
| "cache_implementation": null, | |
| "allow_tf32": false, | |
| "autocast_enabled": false, | |
| "autocast_dtype": null, | |
| "torch_compile": false, | |
| "torch_compile_target": "forward", | |
| "torch_compile_config": {}, | |
| "quantization_scheme": null, | |
| "quantization_config": {}, | |
| "deepspeed_inference": false, | |
| "deepspeed_inference_config": {}, | |
| "peft_type": null, | |
| "peft_config": {} | |
| }, | |
| "scenario": { | |
| "name": "inference", | |
| "_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario", | |
| "iterations": 1, | |
| "duration": 1, | |
| "warmup_runs": 1, | |
| "input_shapes": { | |
| "batch_size": 2, | |
| "sequence_length": 16, | |
| "num_choices": 2 | |
| }, | |
| "new_tokens": null, | |
| "memory": true, | |
| "latency": true, | |
| "energy": true, | |
| "forward_kwargs": {}, | |
| "generate_kwargs": { | |
| "max_new_tokens": 2, | |
| "min_new_tokens": 2 | |
| }, | |
| "call_kwargs": { | |
| "num_inference_steps": 2 | |
| } | |
| }, | |
| "launcher": { | |
| "name": "process", | |
| "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", | |
| "device_isolation": true, | |
| "device_isolation_action": "error", | |
| "numactl": false, | |
| "numactl_kwargs": {}, | |
| "start_method": "spawn" | |
| }, | |
| "environment": { | |
| "cpu": " AMD EPYC 7742 64-Core Processor", | |
| "cpu_count": 128, | |
| "cpu_ram_mb": 540671.643648, | |
| "system": "Linux", | |
| "machine": "x86_64", | |
| "platform": "Linux-5.4.0-166-generic-x86_64-with-glibc2.31", | |
| "processor": "x86_64", | |
| "python_version": "3.10.18", | |
| "gpu": [ | |
| "NVIDIA A100-SXM4-80GB", | |
| "NVIDIA A100-SXM4-80GB", | |
| "NVIDIA A100-SXM4-80GB", | |
| "NVIDIA DGX Display", | |
| "NVIDIA A100-SXM4-80GB" | |
| ], | |
| "gpu_count": 5, | |
| "gpu_vram_mb": 347892350976, | |
| "optimum_benchmark_version": "0.6.0.dev0", | |
| "optimum_benchmark_commit": "61a08086def388b3e78bbf6b42ed20ab4af3f8db", | |
| "transformers_version": "4.45.1", | |
| "transformers_commit": "61a08086def388b3e78bbf6b42ed20ab4af3f8db", | |
| "accelerate_version": "1.10.0", | |
| "accelerate_commit": "61a08086def388b3e78bbf6b42ed20ab4af3f8db", | |
| "diffusers_version": "0.34.0", | |
| "diffusers_commit": "61a08086def388b3e78bbf6b42ed20ab4af3f8db", | |
| "optimum_version": null, | |
| "optimum_commit": null, | |
| "timm_version": "1.0.19", | |
| "timm_commit": "61a08086def388b3e78bbf6b42ed20ab4af3f8db", | |
| "peft_version": "0.17.0", | |
| "peft_commit": "61a08086def388b3e78bbf6b42ed20ab4af3f8db" | |
| }, | |
| "print_report": true, | |
| "log_report": true | |
| }, | |
| "report": { | |
| "load_model": { | |
| "memory": { | |
| "unit": "MB", | |
| "max_ram": 811.282432, | |
| "max_global_vram": 1363.083264, | |
| "max_process_vram": 436.207616, | |
| "max_reserved": 2.097152, | |
| "max_allocated": 0.389632 | |
| }, | |
| "latency": { | |
| "unit": "s", | |
| "values": [ | |
| 0.1634396209716797 | |
| ], | |
| "count": 1, | |
| "total": 0.1634396209716797, | |
| "mean": 0.1634396209716797, | |
| "p50": 0.1634396209716797, | |
| "p90": 0.1634396209716797, | |
| "p95": 0.1634396209716797, | |
| "p99": 0.1634396209716797, | |
| "stdev": 0, | |
| "stdev_": 0 | |
| }, | |
| "throughput": null, | |
| "energy": { | |
| "unit": "kWh", | |
| "cpu": 0.00030914322155877014, | |
| "ram": 0.0005188749783373466, | |
| "gpu": 0.00021641100645553024, | |
| "total": 0.001044429206351647 | |
| }, | |
| "efficiency": null | |
| }, | |
| "first_forward": { | |
| "memory": { | |
| "unit": "MB", | |
| "max_ram": 1006.05952, | |
| "max_global_vram": 1474.23232, | |
| "max_process_vram": 547.356672, | |
| "max_reserved": 23.068672, | |
| "max_allocated": 10.112 | |
| }, | |
| "latency": { | |
| "unit": "s", | |
| "values": [ | |
| 0.5644544067382813 | |
| ], | |
| "count": 1, | |
| "total": 0.5644544067382813, | |
| "mean": 0.5644544067382813, | |
| "p50": 0.5644544067382813, | |
| "p90": 0.5644544067382813, | |
| "p95": 0.5644544067382813, | |
| "p99": 0.5644544067382813, | |
| "stdev": 0, | |
| "stdev_": 0 | |
| }, | |
| "throughput": null, | |
| "energy": { | |
| "unit": "kWh", | |
| "cpu": 0.0003217162053479114, | |
| "ram": 0.0005399784620465254, | |
| "gpu": 0.0002210579546257918, | |
| "total": 0.001082752622020229 | |
| }, | |
| "efficiency": null | |
| }, | |
| "forward": { | |
| "memory": { | |
| "unit": "MB", | |
| "max_ram": 1007.47264, | |
| "max_global_vram": 1474.23232, | |
| "max_process_vram": 547.356672, | |
| "max_reserved": 23.068672, | |
| "max_allocated": 10.112 | |
| }, | |
| "latency": { | |
| "unit": "s", | |
| "values": [ | |
| 0.002642944097518921, | |
| 0.0025815041065216063, | |
| 0.0027904000282287598, | |
| 0.0026204159259796142, | |
| 0.0026224639415740967, | |
| 0.0025784320831298826, | |
| 0.002577408075332642, | |
| 0.002693120002746582, | |
| 0.0026521599292755127, | |
| 0.002702336072921753, | |
| 0.0026265599727630616, | |
| 0.0026603519916534426, | |
| 0.0025753600597381594, | |
| 0.0026316800117492677, | |
| 0.0026603519916534426, | |
| 0.002638848066329956, | |
| 0.0026204159259796142, | |
| 0.0028887040615081785, | |
| 0.0027207679748535154, | |
| 0.00265830397605896, | |
| 0.0026112000942230223, | |
| 0.0026368000507354735, | |
| 0.00263372802734375, | |
| 0.0026675200462341307, | |
| 0.0027688961029052736, | |
| 0.0026449921131134033, | |
| 0.0025681920051574708, | |
| 0.0025036799907684325, | |
| 0.00253439998626709, | |
| 0.0025231359004974367, | |
| 0.0025528318881988523, | |
| 0.002540544033050537, | |
| 0.0028354558944702147, | |
| 0.002687999963760376, | |
| 0.0026316800117492677, | |
| 0.0027002880573272704, | |
| 0.002663424015045166, | |
| 0.002664448022842407, | |
| 0.0026296319961547853, | |
| 0.002698240041732788, | |
| 0.00263372802734375, | |
| 0.0026214399337768555, | |
| 0.0025518081188201903, | |
| 0.0026112000942230223, | |
| 0.002632704019546509, | |
| 0.002548736095428467, | |
| 0.002549760103225708, | |
| 0.002698240041732788, | |
| 0.0026675200462341307, | |
| 0.002550784111022949, | |
| 0.0026849279403686525, | |
| 0.00253439998626709, | |
| 0.002540544033050537, | |
| 0.002548736095428467, | |
| 0.002657279968261719, | |
| 0.0028129279613494872, | |
| 0.002654207944869995, | |
| 0.002584575891494751, | |
| 0.0025671679973602295, | |
| 0.002643968105316162, | |
| 0.0026552319526672364, | |
| 0.0026122241020202636, | |
| 0.0028282880783081056, | |
| 0.0027555840015411376, | |
| 0.0026951680183410643, | |
| 0.002664448022842407, | |
| 0.002569216012954712, | |
| 0.0026112000942230223, | |
| 0.0025681920051574708, | |
| 0.0025241599082946775, | |
| 0.002769920110702515, | |
| 0.002597887992858887, | |
| 0.002579456090927124, | |
| 0.0024872961044311525, | |
| 0.002530303955078125, | |
| 0.0025333759784698486, | |
| 0.0026920959949493406, | |
| 0.0027432959079742433, | |
| 0.0027146239280700685, | |
| 0.002711551904678345, | |
| 0.002678783893585205, | |
| 0.0026122241020202636, | |
| 0.0026808319091796875, | |
| 0.0027156479358673097, | |
| 0.002753535985946655, | |
| 0.002811903953552246, | |
| 0.002687999963760376, | |
| 0.0028569600582122805, | |
| 0.0026951680183410643, | |
| 0.0027197439670562746, | |
| 0.002654207944869995, | |
| 0.0025712640285491944, | |
| 0.0026757121086120605, | |
| 0.002732032060623169, | |
| 0.002657279968261719, | |
| 0.0026460158824920654, | |
| 0.002672640085220337, | |
| 0.0026419200897216796, | |
| 0.0026122241020202636, | |
| 0.0026460158824920654, | |
| 0.00276582407951355, | |
| 0.0027217919826507567, | |
| 0.002686975955963135, | |
| 0.002661375999450684, | |
| 0.002604032039642334, | |
| 0.002603008031845093, | |
| 0.002535423994064331, | |
| 0.002632704019546509, | |
| 0.002668544054031372, | |
| 0.0026408960819244384, | |
| 0.0026081280708312986, | |
| 0.002673664093017578, | |
| 0.002672640085220337, | |
| 0.002639872074127197, | |
| 0.002632704019546509, | |
| 0.0026972160339355468, | |
| 0.002648063898086548, | |
| 0.0025681920051574708, | |
| 0.0026009600162506105, | |
| 0.0025866239070892333, | |
| 0.002642944097518921, | |
| 0.0026122241020202636, | |
| 0.002770944118499756, | |
| 0.0026460158824920654, | |
| 0.002543616056442261, | |
| 0.002554879903793335, | |
| 0.0025784320831298826, | |
| 0.002573312044143677, | |
| 0.0025579519271850585, | |
| 0.0026122241020202636, | |
| 0.0026859519481658937, | |
| 0.002712575912475586, | |
| 0.00268287992477417, | |
| 0.00263372802734375, | |
| 0.0027678720951080323, | |
| 0.002737152099609375, | |
| 0.0026849279403686525, | |
| 0.002756608009338379, | |
| 0.0027443199157714845, | |
| 0.0027985920906066896, | |
| 0.0027310080528259277, | |
| 0.002693120002746582, | |
| 0.00268287992477417, | |
| 0.002597887992858887, | |
| 0.0026019840240478517, | |
| 0.002634752035140991, | |
| 0.0025384960174560547, | |
| 0.0025528318881988523, | |
| 0.0025159680843353273, | |
| 0.002594815969467163, | |
| 0.002489343881607056, | |
| 0.0024453120231628417, | |
| 0.0025159680843353273, | |
| 0.0025518081188201903, | |
| 0.002465791940689087, | |
| 0.002460671901702881, | |
| 0.002476032018661499, | |
| 0.0024401919841766356, | |
| 0.0025159680843353273, | |
| 0.0024494080543518066, | |
| 0.0025620479583740234, | |
| 0.002574336051940918, | |
| 0.002494463920593262, | |
| 0.0024688639640808104, | |
| 0.002467839956283569, | |
| 0.0024995839595794676, | |
| 0.0024924159049987795, | |
| 0.002514944076538086, | |
| 0.0025251839160919188, | |
| 0.0025763840675354006, | |
| 0.002530303955078125, | |
| 0.0025169920921325685, | |
| 0.002539520025253296, | |
| 0.0025180160999298097, | |
| 0.0025815041065216063, | |
| 0.0024954879283905027, | |
| 0.0025681920051574708, | |
| 0.002590719938278198, | |
| 0.0025128960609436036, | |
| 0.002480128049850464, | |
| 0.0025088000297546387, | |
| 0.0024954879283905027, | |
| 0.0025333759784698486, | |
| 0.002548736095428467, | |
| 0.0025384960174560547, | |
| 0.002544640064239502, | |
| 0.0025231359004974367, | |
| 0.002500607967376709, | |
| 0.002514944076538086, | |
| 0.0025088000297546387, | |
| 0.0025036799907684325, | |
| 0.0025118720531463624, | |
| 0.002569216012954712, | |
| 0.002548736095428467, | |
| 0.0025722880363464357, | |
| 0.0024872961044311525, | |
| 0.002505728006362915, | |
| 0.0025569279193878173, | |
| 0.0025323519706726074, | |
| 0.002543616056442261, | |
| 0.002597887992858887, | |
| 0.002577408075332642, | |
| 0.0025425920486450196, | |
| 0.0025088000297546387, | |
| 0.0024985599517822264, | |
| 0.002513920068740845, | |
| 0.0025067520141601562, | |
| 0.002573312044143677, | |
| 0.002590719938278198, | |
| 0.0025466880798339846, | |
| 0.0025231359004974367, | |
| 0.0025118720531463624, | |
| 0.002543616056442261, | |
| 0.002561023950576782, | |
| 0.0024791040420532227, | |
| 0.002573312044143677, | |
| 0.002594815969467163, | |
| 0.002584575891494751, | |
| 0.0025231359004974367, | |
| 0.0025118720531463624, | |
| 0.0025159680843353273, | |
| 0.002604032039642334, | |
| 0.00263372802734375, | |
| 0.002549760103225708, | |
| 0.002653183937072754, | |
| 0.002555903911590576, | |
| 0.002594815969467163, | |
| 0.002500607967376709, | |
| 0.0024862720966339113, | |
| 0.0025927679538726807, | |
| 0.002543616056442261, | |
| 0.0025999360084533693, | |
| 0.0025630719661712646, | |
| 0.002573312044143677, | |
| 0.002632704019546509, | |
| 0.0025077760219573975, | |
| 0.0025968639850616456, | |
| 0.0025292799472808837, | |
| 0.0025528318881988523, | |
| 0.002613248109817505, | |
| 0.002561023950576782, | |
| 0.002540544033050537, | |
| 0.002548736095428467, | |
| 0.0025569279193878173, | |
| 0.0024924159049987795, | |
| 0.0025835518836975096, | |
| 0.00253439998626709, | |
| 0.0025712640285491944, | |
| 0.0025128960609436036, | |
| 0.0025589759349822997, | |
| 0.0025876479148864745, | |
| 0.002514944076538086, | |
| 0.002539520025253296, | |
| 0.0025825281143188475, | |
| 0.002519040107727051, | |
| 0.002574336051940918, | |
| 0.0024821760654449463, | |
| 0.002514944076538086, | |
| 0.00248524808883667, | |
| 0.0024729599952697753, | |
| 0.0024913918972015382, | |
| 0.00250982403755188, | |
| 0.0024832000732421876, | |
| 0.002548736095428467, | |
| 0.0025538558959960936, | |
| 0.002510848045349121, | |
| 0.002470911979675293, | |
| 0.002489343881607056, | |
| 0.00250163197517395, | |
| 0.0025364480018615723, | |
| 0.002471935987472534, | |
| 0.002580480098724365, | |
| 0.002513920068740845, | |
| 0.002481152057647705, | |
| 0.002531327962875366, | |
| 0.0025128960609436036, | |
| 0.0024821760654449463, | |
| 0.0025374720096588135, | |
| 0.002497535943984985, | |
| 0.0025784320831298826, | |
| 0.002569216012954712, | |
| 0.002470911979675293, | |
| 0.002481152057647705, | |
| 0.0025047039985656738, | |
| 0.0025681920051574708, | |
| 0.0025968639850616456, | |
| 0.0025456640720367433, | |
| 0.0025538558959960936, | |
| 0.002603008031845093, | |
| 0.002527231931686401, | |
| 0.0025047039985656738, | |
| 0.0024985599517822264, | |
| 0.002554879903793335, | |
| 0.002510848045349121, | |
| 0.0025589759349822997, | |
| 0.002496511936187744, | |
| 0.002513920068740845, | |
| 0.0025815041065216063, | |
| 0.0025210878849029543, | |
| 0.002573312044143677, | |
| 0.0025671679973602295, | |
| 0.0025374720096588135, | |
| 0.0025815041065216063, | |
| 0.002573312044143677, | |
| 0.00250163197517395, | |
| 0.0025958399772644043, | |
| 0.0025374720096588135, | |
| 0.002497535943984985, | |
| 0.002540544033050537, | |
| 0.002466815948486328, | |
| 0.002573312044143677, | |
| 0.002634752035140991, | |
| 0.0024821760654449463, | |
| 0.0025036799907684325, | |
| 0.0025221118927001955, | |
| 0.0024627199172973634, | |
| 0.002471935987472534, | |
| 0.0025210878849029543, | |
| 0.002642944097518921, | |
| 0.0024862720966339113, | |
| 0.002471935987472534, | |
| 0.002510848045349121, | |
| 0.0025169920921325685, | |
| 0.002451456069946289, | |
| 0.002451456069946289, | |
| 0.0025323519706726074, | |
| 0.002548736095428467, | |
| 0.0025047039985656738, | |
| 0.0024750080108642577, | |
| 0.002514944076538086, | |
| 0.0025528318881988523, | |
| 0.0024442880153656005, | |
| 0.0024576001167297364, | |
| 0.0024729599952697753, | |
| 0.0025681920051574708, | |
| 0.0024576001167297364, | |
| 0.0024739840030670165, | |
| 0.0025128960609436036, | |
| 0.002519040107727051, | |
| 0.0024688639640808104, | |
| 0.0025241599082946775, | |
| 0.0024627199172973634, | |
| 0.0025671679973602295, | |
| 0.0024995839595794676, | |
| 0.002484224081039429, | |
| 0.0025088000297546387, | |
| 0.002455552101135254, | |
| 0.0024637439250946047, | |
| 0.00253439998626709, | |
| 0.0024729599952697753, | |
| 0.0025917439460754394, | |
| 0.002514944076538086, | |
| 0.0024371199607849123, | |
| 0.002500607967376709, | |
| 0.002550784111022949, | |
| 0.0024750080108642577, | |
| 0.0025231359004974367, | |
| 0.0024821760654449463, | |
| 0.002555903911590576, | |
| 0.0024954879283905027, | |
| 0.0025159680843353273, | |
| 0.002461695909500122, | |
| 0.0025077760219573975, | |
| 0.0024279038906097413, | |
| 0.0024115200042724608, | |
| 0.002484224081039429, | |
| 0.002550784111022949, | |
| 0.0024995839595794676, | |
| 0.0025047039985656738, | |
| 0.0024524800777435303, | |
| 0.0024883201122283937, | |
| 0.002421760082244873, | |
| 0.002406399965286255, | |
| 0.002480128049850464, | |
| 0.0025210878849029543, | |
| 0.002494463920593262, | |
| 0.002421760082244873, | |
| 0.002441215991973877, | |
| 0.002470911979675293, | |
| 0.002484224081039429, | |
| 0.00252620792388916, | |
| 0.002471935987472534, | |
| 0.0025661439895629883, | |
| 0.0024780800342559814, | |
| 0.002496511936187744 | |
| ], | |
| "count": 385, | |
| "total": 0.9895669796466826, | |
| "mean": 0.002570303843238137, | |
| "p50": 0.0025528318881988523, | |
| "p90": 0.002687999963760376, | |
| "p95": 0.0027318272590637206, | |
| "p99": 0.0028153855800628655, | |
| "stdev": 8.564019716831984e-05, | |
| "stdev_": 3.3319094702993572 | |
| }, | |
| "throughput": { | |
| "unit": "samples/s", | |
| "value": 778.1181222062629 | |
| }, | |
| "energy": { | |
| "unit": "kWh", | |
| "cpu": 7.926489950800206e-08, | |
| "ram": 1.3301133581781163e-07, | |
| "gpu": 5.085165812373287e-08, | |
| "total": 2.631278934495463e-07 | |
| }, | |
| "efficiency": { | |
| "unit": "samples/kWh", | |
| "value": 7600866.535966442 | |
| } | |
| } | |
| } | |
| } |