local
/
cpu_inference_transformers_fill-mask_hf-internal-testing
/tiny-random-BertModel
/benchmark.json
IlyasMoutawwakil
HF Staff
Upload cpu_inference_transformers_fill-mask_hf-internal-testing/tiny-random-BertModel/benchmark.json with huggingface_hub
1c3d23d
verified
| { | |
| "config": { | |
| "name": "cpu_inference_transformers_fill-mask_hf-internal-testing/tiny-random-BertModel", | |
| "backend": { | |
| "name": "pytorch", | |
| "version": "2.5.1", | |
| "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", | |
| "model": "hf-internal-testing/tiny-random-BertModel", | |
| "processor": "hf-internal-testing/tiny-random-BertModel", | |
| "task": "fill-mask", | |
| "library": "transformers", | |
| "model_type": "bert", | |
| "device": "cpu", | |
| "device_ids": null, | |
| "seed": 42, | |
| "inter_op_num_threads": null, | |
| "intra_op_num_threads": null, | |
| "model_kwargs": {}, | |
| "processor_kwargs": {}, | |
| "no_weights": true, | |
| "tp_plan": null, | |
| "device_map": null, | |
| "torch_dtype": null, | |
| "eval_mode": true, | |
| "to_bettertransformer": false, | |
| "low_cpu_mem_usage": null, | |
| "attn_implementation": null, | |
| "cache_implementation": null, | |
| "allow_tf32": false, | |
| "autocast_enabled": false, | |
| "autocast_dtype": null, | |
| "torch_compile": false, | |
| "torch_compile_target": "forward", | |
| "torch_compile_config": {}, | |
| "quantization_scheme": null, | |
| "quantization_config": {}, | |
| "deepspeed_inference": false, | |
| "deepspeed_inference_config": {}, | |
| "peft_type": null, | |
| "peft_config": {} | |
| }, | |
| "scenario": { | |
| "name": "inference", | |
| "_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario", | |
| "iterations": 1, | |
| "duration": 1, | |
| "warmup_runs": 1, | |
| "input_shapes": { | |
| "batch_size": 2, | |
| "sequence_length": 16, | |
| "num_choices": 2 | |
| }, | |
| "new_tokens": null, | |
| "memory": true, | |
| "latency": true, | |
| "energy": true, | |
| "forward_kwargs": {}, | |
| "generate_kwargs": { | |
| "max_new_tokens": 2, | |
| "min_new_tokens": 2 | |
| }, | |
| "call_kwargs": { | |
| "num_inference_steps": 2 | |
| } | |
| }, | |
| "launcher": { | |
| "name": "process", | |
| "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", | |
| "device_isolation": false, | |
| "device_isolation_action": null, | |
| "numactl": false, | |
| "numactl_kwargs": {}, | |
| "start_method": "spawn" | |
| }, | |
| "environment": { | |
| "cpu": " AMD EPYC 7742 64-Core Processor", | |
| "cpu_count": 128, | |
| "cpu_ram_mb": 540671.643648, | |
| "system": "Linux", | |
| "machine": "x86_64", | |
| "platform": "Linux-5.4.0-166-generic-x86_64-with-glibc2.31", | |
| "processor": "x86_64", | |
| "python_version": "3.10.18", | |
| "gpu": [ | |
| "NVIDIA A100-SXM4-80GB", | |
| "NVIDIA A100-SXM4-80GB", | |
| "NVIDIA A100-SXM4-80GB", | |
| "NVIDIA DGX Display", | |
| "NVIDIA A100-SXM4-80GB" | |
| ], | |
| "gpu_count": 5, | |
| "gpu_vram_mb": 347892350976, | |
| "optimum_benchmark_version": "0.6.0.dev0", | |
| "optimum_benchmark_commit": "61a08086def388b3e78bbf6b42ed20ab4af3f8db", | |
| "transformers_version": "4.45.1", | |
| "transformers_commit": "61a08086def388b3e78bbf6b42ed20ab4af3f8db", | |
| "accelerate_version": "1.10.0", | |
| "accelerate_commit": "61a08086def388b3e78bbf6b42ed20ab4af3f8db", | |
| "diffusers_version": "0.34.0", | |
| "diffusers_commit": "61a08086def388b3e78bbf6b42ed20ab4af3f8db", | |
| "optimum_version": null, | |
| "optimum_commit": null, | |
| "timm_version": "1.0.19", | |
| "timm_commit": "61a08086def388b3e78bbf6b42ed20ab4af3f8db", | |
| "peft_version": "0.17.0", | |
| "peft_commit": "61a08086def388b3e78bbf6b42ed20ab4af3f8db" | |
| }, | |
| "print_report": true, | |
| "log_report": true | |
| }, | |
| "report": { | |
| "load_model": { | |
| "memory": { | |
| "unit": "MB", | |
| "max_ram": 679.559168, | |
| "max_global_vram": null, | |
| "max_process_vram": null, | |
| "max_reserved": null, | |
| "max_allocated": null | |
| }, | |
| "latency": { | |
| "unit": "s", | |
| "values": [ | |
| 0.14881402999162674 | |
| ], | |
| "count": 1, | |
| "total": 0.14881402999162674, | |
| "mean": 0.14881402999162674, | |
| "p50": 0.14881402999162674, | |
| "p90": 0.14881402999162674, | |
| "p95": 0.14881402999162674, | |
| "p99": 0.14881402999162674, | |
| "stdev": 0, | |
| "stdev_": 0 | |
| }, | |
| "throughput": null, | |
| "energy": { | |
| "unit": "kWh", | |
| "cpu": 0.00015839945019979496, | |
| "ram": 0.0002658581267750873, | |
| "gpu": 0.00040478087937856344, | |
| "total": 0.0008290384563534457 | |
| }, | |
| "efficiency": null | |
| }, | |
| "first_forward": { | |
| "memory": { | |
| "unit": "MB", | |
| "max_ram": 680.267776, | |
| "max_global_vram": null, | |
| "max_process_vram": null, | |
| "max_reserved": null, | |
| "max_allocated": null | |
| }, | |
| "latency": { | |
| "unit": "s", | |
| "values": [ | |
| 0.008422935847193003 | |
| ], | |
| "count": 1, | |
| "total": 0.008422935847193003, | |
| "mean": 0.008422935847193003, | |
| "p50": 0.008422935847193003, | |
| "p90": 0.008422935847193003, | |
| "p95": 0.008422935847193003, | |
| "p99": 0.008422935847193003, | |
| "stdev": 0, | |
| "stdev_": 0 | |
| }, | |
| "throughput": null, | |
| "energy": { | |
| "unit": "kWh", | |
| "cpu": 0.0001545904000668088, | |
| "ram": 0.0002594645603005972, | |
| "gpu": 0.00038416502955840315, | |
| "total": 0.000798219989925809 | |
| }, | |
| "efficiency": null | |
| }, | |
| "forward": { | |
| "memory": { | |
| "unit": "MB", | |
| "max_ram": 687.833088, | |
| "max_global_vram": null, | |
| "max_process_vram": null, | |
| "max_reserved": null, | |
| "max_allocated": null | |
| }, | |
| "latency": { | |
| "unit": "s", | |
| "values": [ | |
| 0.002831811085343361, | |
| 0.0027529229409992695, | |
| 0.0027246088720858097, | |
| 0.0028367098420858383, | |
| 0.0027337572537362576, | |
| 0.002710293047130108, | |
| 0.0027137091383337975, | |
| 0.002639930695295334, | |
| 0.0026258742436766624, | |
| 0.002674704883247614, | |
| 0.002695303875952959, | |
| 0.0027544950135052204, | |
| 0.0027153408154845238, | |
| 0.002661501057446003, | |
| 0.0026737446896731853, | |
| 0.002680666744709015, | |
| 0.0026717609725892544, | |
| 0.0026349108666181564, | |
| 0.0027218740433454514, | |
| 0.0026273271068930626, | |
| 0.0025601410306990147, | |
| 0.0025830939412117004, | |
| 0.0026146029122173786, | |
| 0.0025855777785182, | |
| 0.0025835949927568436, | |
| 0.0025653610937297344, | |
| 0.0026709591038525105, | |
| 0.0026035429909825325, | |
| 0.002610635943710804, | |
| 0.002631494775414467, | |
| 0.002649869304150343, | |
| 0.002667392138391733, | |
| 0.0026115477085113525, | |
| 0.002584747038781643, | |
| 0.0026388089172542095, | |
| 0.00255892938002944, | |
| 0.0025786347687244415, | |
| 0.0025820820592343807, | |
| 0.002565390896052122, | |
| 0.002562745939940214, | |
| 0.0026137609966099262, | |
| 0.0026076999492943287, | |
| 0.002572102937847376, | |
| 0.002562555018812418, | |
| 0.002585679292678833, | |
| 0.0026052147150039673, | |
| 0.0025855586864054203, | |
| 0.0025789663195610046, | |
| 0.0026012081652879715, | |
| 0.00263608293607831, | |
| 0.002570010256022215, | |
| 0.002641713712364435, | |
| 0.0026368550024926662, | |
| 0.0026613399386405945, | |
| 0.0026696957647800446, | |
| 0.002653926145285368, | |
| 0.0026658792048692703, | |
| 0.0027573509141802788, | |
| 0.002686257939785719, | |
| 0.002671870868653059, | |
| 0.002650290261954069, | |
| 0.0025801178999245167, | |
| 0.0025781039148569107, | |
| 0.0025948258116841316, | |
| 0.0026422650553286076, | |
| 0.002637235913425684, | |
| 0.0026484872214496136, | |
| 0.002593553625047207, | |
| 0.0025963387452065945, | |
| 0.0025863200426101685, | |
| 0.0025483090430498123, | |
| 0.002589415293186903, | |
| 0.0027642627246677876, | |
| 0.002826390787959099, | |
| 0.002728687133640051, | |
| 0.0027176570147275925, | |
| 0.0026821200735867023, | |
| 0.002717094961553812, | |
| 0.0027332347817718983, | |
| 0.002709541004151106, | |
| 0.0028211409226059914, | |
| 0.0027777189388871193, | |
| 0.0027347891591489315, | |
| 0.0027413908392190933, | |
| 0.002716194372624159, | |
| 0.0026784627698361874, | |
| 0.0026074089109897614, | |
| 0.0027065258473157883, | |
| 0.002648227382451296, | |
| 0.002652904950082302, | |
| 0.002605196088552475, | |
| 0.0025926423259079456, | |
| 0.002583986148238182, | |
| 0.0026267259381711483, | |
| 0.0025806291960179806, | |
| 0.002675727941095829, | |
| 0.0026479260995984077, | |
| 0.00265714293345809, | |
| 0.0026852251030504704, | |
| 0.0026663709431886673, | |
| 0.0026177302934229374, | |
| 0.002601148094981909, | |
| 0.002701906953006983, | |
| 0.0025688670575618744, | |
| 0.0025847270153462887, | |
| 0.0026042540557682514, | |
| 0.002581731416285038, | |
| 0.0025981119833886623, | |
| 0.0025748591870069504, | |
| 0.002553307916969061, | |
| 0.0026549389585852623, | |
| 0.002589976880699396, | |
| 0.0025946958921849728, | |
| 0.0025792871601879597, | |
| 0.002557766158133745, | |
| 0.002560011111199856, | |
| 0.0026106061413884163, | |
| 0.00258822413161397, | |
| 0.0026662801392376423, | |
| 0.0026261350139975548, | |
| 0.0026178103871643543, | |
| 0.0025977017357945442, | |
| 0.00260597700253129, | |
| 0.002601419109851122, | |
| 0.002598542720079422, | |
| 0.0026544490829110146, | |
| 0.00261505413800478, | |
| 0.002621215768158436, | |
| 0.002593092620372772, | |
| 0.00259593827649951, | |
| 0.0026677530258893967, | |
| 0.002680005971342325, | |
| 0.0026747160591185093, | |
| 0.0028274422511458397, | |
| 0.002778911031782627, | |
| 0.0027624801732599735, | |
| 0.002755246590822935, | |
| 0.0027657868340611458, | |
| 0.002757551148533821, | |
| 0.002725029829889536, | |
| 0.00284017575904727, | |
| 0.0027232267893850803, | |
| 0.00271938880905509, | |
| 0.0026777819730341434, | |
| 0.002665778622031212, | |
| 0.0026338500902056694, | |
| 0.002633708994835615, | |
| 0.0026066978462040424, | |
| 0.002645730972290039, | |
| 0.002599785104393959, | |
| 0.002630883827805519, | |
| 0.0026519829407334328, | |
| 0.00270979106426239, | |
| 0.002751339226961136, | |
| 0.0026912870816886425, | |
| 0.0027004238218069077, | |
| 0.0026159859262406826, | |
| 0.0025703287683427334, | |
| 0.0025942251086235046, | |
| 0.002606278285384178, | |
| 0.002604664769023657, | |
| 0.0025924108922481537, | |
| 0.00259727006778121, | |
| 0.0026735030114650726, | |
| 0.002579587046056986, | |
| 0.0025506732054054737, | |
| 0.002545243129134178, | |
| 0.0025887549854815006, | |
| 0.0025591892190277576, | |
| 0.0025789160281419754, | |
| 0.0025969198904931545, | |
| 0.0026767291128635406, | |
| 0.0026505510322749615, | |
| 0.0026231491938233376, | |
| 0.0025825733318924904, | |
| 0.002604193054139614, | |
| 0.0026407232508063316, | |
| 0.002597931306809187, | |
| 0.0026321657933294773, | |
| 0.002742283046245575, | |
| 0.0027189780957996845, | |
| 0.0027602771297097206, | |
| 0.0027607372030615807, | |
| 0.0026779319159686565, | |
| 0.0026023900136351585, | |
| 0.002625253051519394, | |
| 0.0026576551608741283, | |
| 0.002606980036944151, | |
| 0.002618159167468548, | |
| 0.0025986325927078724, | |
| 0.00259243231266737, | |
| 0.002606297843158245, | |
| 0.002581820823252201, | |
| 0.0026193922385573387, | |
| 0.002694672904908657, | |
| 0.002622256986796856, | |
| 0.0026164259761571884, | |
| 0.0026205149479210377, | |
| 0.0026259650476276875, | |
| 0.002641574013978243, | |
| 0.0026219268329441547, | |
| 0.002665828913450241, | |
| 0.0026539969258010387, | |
| 0.0026247319765388966, | |
| 0.002590416930615902, | |
| 0.0026114871725440025, | |
| 0.002616936806589365, | |
| 0.002601799089461565, | |
| 0.0026082410477101803, | |
| 0.002683452796190977, | |
| 0.0026596570387482643, | |
| 0.002697909716516733, | |
| 0.0027176872827112675, | |
| 0.0027074976824223995, | |
| 0.0027302498929202557, | |
| 0.0027610273100435734, | |
| 0.002785805147141218, | |
| 0.00276709022000432, | |
| 0.0027224458754062653, | |
| 0.0026900148950517178, | |
| 0.002702157013118267, | |
| 0.002749084960669279, | |
| 0.0027403589338064194, | |
| 0.0027441363781690598, | |
| 0.0027553769759833813, | |
| 0.002630012109875679, | |
| 0.0026188911870121956, | |
| 0.0026341299526393414, | |
| 0.002647495362907648, | |
| 0.002632095944136381, | |
| 0.002622507978230715, | |
| 0.002701836172491312, | |
| 0.0026323669590055943, | |
| 0.0026261149905622005, | |
| 0.0025915498845279217, | |
| 0.002607117872685194, | |
| 0.0026581836864352226, | |
| 0.002740319352596998, | |
| 0.002685976680368185, | |
| 0.0026982896961271763, | |
| 0.002615605015307665, | |
| 0.002637697383761406, | |
| 0.00261252885684371, | |
| 0.002633138094097376, | |
| 0.002654068171977997, | |
| 0.0026060566306114197, | |
| 0.002590127754956484, | |
| 0.0026873089373111725, | |
| 0.002620665356516838, | |
| 0.0026073800399899483, | |
| 0.002676489297300577, | |
| 0.0026698061265051365, | |
| 0.0026501002721488476, | |
| 0.0026447600685060024, | |
| 0.0026764790527522564, | |
| 0.00260948296636343, | |
| 0.0026153340004384518, | |
| 0.0026255641132593155, | |
| 0.0026244018226861954, | |
| 0.0026540779508650303, | |
| 0.0026348107494413853, | |
| 0.0026372959837317467, | |
| 0.002701575867831707, | |
| 0.0026408019475638866, | |
| 0.0026236302219331264, | |
| 0.002644078340381384, | |
| 0.0026295706629753113, | |
| 0.002660319209098816, | |
| 0.0026665409095585346, | |
| 0.0026920479722321033, | |
| 0.0026878509670495987, | |
| 0.002672250848263502, | |
| 0.002675988245755434, | |
| 0.0026738038286566734, | |
| 0.002720171120017767, | |
| 0.0027068760246038437, | |
| 0.0027110143564641476, | |
| 0.00283443508669734, | |
| 0.0027134385891258717, | |
| 0.0027259723283350468, | |
| 0.0026616910472512245, | |
| 0.00266280397772789, | |
| 0.002625504042953253, | |
| 0.0026272074319422245, | |
| 0.002691878005862236, | |
| 0.002691968809813261, | |
| 0.0026845536194741726, | |
| 0.0026857461780309677, | |
| 0.0026493482291698456, | |
| 0.0026848549023270607, | |
| 0.0026484858244657516, | |
| 0.0026492690667510033, | |
| 0.0027198009192943573, | |
| 0.002644097898155451, | |
| 0.002560941968113184, | |
| 0.0025781849399209023, | |
| 0.0025786557234823704, | |
| 0.002567645162343979, | |
| 0.0025629960000514984, | |
| 0.0025580767542123795, | |
| 0.002608812879770994, | |
| 0.002560561988502741, | |
| 0.0025816713459789753, | |
| 0.0025675445795059204, | |
| 0.0025757201947271824, | |
| 0.0025894762948155403, | |
| 0.002610113937407732, | |
| 0.0026615308597683907, | |
| 0.0026047746650874615, | |
| 0.002584496047347784, | |
| 0.0025961389765143394, | |
| 0.00257821474224329, | |
| 0.0026288102380931377, | |
| 0.0026331888511776924, | |
| 0.002625604160130024, | |
| 0.0026751761324703693, | |
| 0.0025922409258782864, | |
| 0.002580889966338873, | |
| 0.002572744619101286, | |
| 0.002562996931374073, | |
| 0.0025603421963751316, | |
| 0.002570711076259613, | |
| 0.002588694915175438, | |
| 0.002619482111185789, | |
| 0.0025880038738250732, | |
| 0.002567254938185215, | |
| 0.0025802291929721832, | |
| 0.0025844271294772625, | |
| 0.002586841117590666, | |
| 0.0025745779275894165, | |
| 0.0025813300162553787, | |
| 0.002612078096717596, | |
| 0.0025806589983403683, | |
| 0.0025669229216873646, | |
| 0.0025774440728127956, | |
| 0.0025810301303863525, | |
| 0.0025612828321754932, | |
| 0.002566984388977289, | |
| 0.0025728149339556694, | |
| 0.0026128599420189857, | |
| 0.0025754603557288647, | |
| 0.002572544850409031, | |
| 0.002579357009381056, | |
| 0.002562706358730793, | |
| 0.0025800177827477455, | |
| 0.0025769220665097237, | |
| 0.0026219473220407963, | |
| 0.0025798678398132324, | |
| 0.0025990940630435944, | |
| 0.002597521059215069, | |
| 0.002593433950096369, | |
| 0.00259942514821887, | |
| 0.002599504776299, | |
| 0.0025898870080709457, | |
| 0.002639380283653736, | |
| 0.002604834735393524, | |
| 0.0025976509787142277, | |
| 0.002606858965009451, | |
| 0.002601419109851122, | |
| 0.0026102247647941113, | |
| 0.0026018288917839527, | |
| 0.002616487443447113, | |
| 0.0026577343232929707, | |
| 0.0026196730323135853, | |
| 0.0026202131994068623, | |
| 0.002624782267957926, | |
| 0.0026039928197860718, | |
| 0.002626976929605007, | |
| 0.0026326170191168785, | |
| 0.0026807766407728195, | |
| 0.0026374259032309055, | |
| 0.002629511058330536, | |
| 0.002639159094542265, | |
| 0.0026420247741043568, | |
| 0.002641173079609871, | |
| 0.0026397807523608208, | |
| 0.002645079977810383, | |
| 0.002682761289179325 | |
| ], | |
| "count": 378, | |
| "total": 0.998018033336848, | |
| "mean": 0.002640259347451979, | |
| "p50": 0.0026264304760843515, | |
| "p90": 0.0027236414141952993, | |
| "p95": 0.0027556730667129157, | |
| "p99": 0.0028284470830112697, | |
| "stdev": 5.993870029621387e-05, | |
| "stdev_": 2.270182296828476 | |
| }, | |
| "throughput": { | |
| "unit": "samples/s", | |
| "value": 757.5013424079455 | |
| }, | |
| "energy": { | |
| "unit": "kWh", | |
| "cpu": 8.709281649729789e-08, | |
| "ram": 1.461470135525111e-07, | |
| "gpu": 2.1126912600919842e-07, | |
| "total": 4.4450895605900725e-07 | |
| }, | |
| "efficiency": { | |
| "unit": "samples/kWh", | |
| "value": 4499346.914698623 | |
| } | |
| } | |
| } | |
| } |