{ "config": { "name": "cpu_inference_transformers_text-classification_FacebookAI/roberta-base", "backend": { "name": "pytorch", "version": "2.3.1+cpu", "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", "task": "text-classification", "library": "transformers", "model_type": "roberta", "model": "FacebookAI/roberta-base", "processor": "FacebookAI/roberta-base", "device": "cpu", "device_ids": null, "seed": 42, "inter_op_num_threads": null, "intra_op_num_threads": null, "model_kwargs": {}, "processor_kwargs": {}, "no_weights": true, "device_map": null, "torch_dtype": null, "eval_mode": true, "to_bettertransformer": false, "low_cpu_mem_usage": null, "attn_implementation": null, "cache_implementation": null, "autocast_enabled": false, "autocast_dtype": null, "torch_compile": false, "torch_compile_target": "forward", "torch_compile_config": {}, "quantization_scheme": null, "quantization_config": {}, "deepspeed_inference": false, "deepspeed_inference_config": {}, "peft_type": null, "peft_config": {} }, "scenario": { "name": "inference", "_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario", "iterations": 1, "duration": 1, "warmup_runs": 1, "input_shapes": { "batch_size": 1, "num_choices": 2, "sequence_length": 2 }, "new_tokens": null, "memory": true, "latency": true, "energy": true, "forward_kwargs": {}, "generate_kwargs": { "max_new_tokens": 2, "min_new_tokens": 2 }, "call_kwargs": { "num_inference_steps": 2 } }, "launcher": { "name": "process", "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", "device_isolation": false, "device_isolation_action": "error", "numactl": false, "numactl_kwargs": {}, "start_method": "spawn" }, "environment": { "cpu": " AMD EPYC 7742 64-Core Processor", "cpu_count": 128, "cpu_ram_mb": 540671.660032, "system": "Linux", "machine": "x86_64", "platform": "Linux-5.4.0-166-generic-x86_64-with-glibc2.35", "processor": "x86_64", "python_version": "3.10.12", "optimum_benchmark_version": "0.4.0", "optimum_benchmark_commit": null, "transformers_version": "4.44.0", "transformers_commit": null, "accelerate_version": "0.33.0", "accelerate_commit": null, "diffusers_version": "0.30.0", "diffusers_commit": null, "optimum_version": null, "optimum_commit": null, "timm_version": "1.0.8", "timm_commit": null, "peft_version": "0.12.0", "peft_commit": null } }, "report": { "load": { "memory": { "unit": "MB", "max_ram": 932.970496, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "count": 1, "total": 4.826313242316246, "mean": 4.826313242316246, "stdev": 0.0, "p50": 4.826313242316246, "p90": 4.826313242316246, "p95": 4.826313242316246, "p99": 4.826313242316246, "values": [ 4.826313242316246 ] }, "throughput": null, "energy": { "unit": "kWh", "cpu": 4.93125157954637e-05, "ram": 8.275760335821636e-05, "gpu": 0, "total": 0.00013207011915368006 }, "efficiency": null }, "forward": { "memory": { "unit": "MB", "max_ram": 939.401216, "max_global_vram": null, "max_process_vram": null, "max_reserved": null, "max_allocated": null }, "latency": { "unit": "s", "count": 73, "total": 1.004445269703865, "mean": 0.0137595242425187, "stdev": 0.002681957134060987, "p50": 0.013237454928457737, "p90": 0.013591494970023632, "p95": 0.016829280741512742, "p99": 0.026973111443221575, "values": [ 0.013589794747531414, 0.01350796315819025, 0.01350696012377739, 0.01345173642039299, 0.013591920025646687, 0.013205344788730145, 0.01327508594840765, 0.013286015950143337, 0.013271508738398552, 0.013275175355374813, 0.01340556051582098, 0.013237454928457737, 0.013269043527543545, 0.013266097754240036, 0.013387937098741531, 0.013316133059561253, 0.0135231614112854, 0.01330757699906826, 0.01518827211111784, 0.01498947013169527, 0.014381790533661842, 0.012849448248744011, 0.012833258137106895, 0.01277841441333294, 0.012757246382534504, 0.01277922559529543, 0.012846463359892368, 0.012949345633387566, 0.012779676355421543, 0.01276503037661314, 0.012823337689042091, 0.01280023530125618, 0.01280030608177185, 0.012851080857217312, 0.01281034480780363, 0.012781369499862194, 0.012837535701692104, 0.012843586504459381, 0.012832175008952618, 0.012785819359123707, 0.030748971737921238, 0.019308296032249928, 0.01929079368710518, 0.013026909902691841, 0.012960856780409813, 0.012909481301903725, 0.01283207442611456, 0.012735874392092228, 0.0128360316157341, 0.01281804870814085, 0.012826554477214813, 0.01280226930975914, 0.012883090414106846, 0.012832175008952618, 0.012827487662434578, 0.013161363080143929, 0.013353902846574783, 0.013474230654537678, 0.013304339721798897, 0.025504721328616142, 0.013331371359527111, 0.013482905924320221, 0.013226453214883804, 0.013216877356171608, 0.013500909321010113, 0.013174005784094334, 0.013559439219534397, 0.013263463042676449, 0.013299130834639072, 0.013256601057946682, 0.0132904639467597, 0.013352160342037678, 0.013316112570464611 ] }, "throughput": { "unit": "samples/s", "value": 72.67693143850653 }, "energy": { "unit": "kWh", "cpu": 4.59740020206495e-07, "ram": 7.714110161284184e-07, "gpu": 0.0, "total": 1.2311510363349134e-06 }, "efficiency": { "unit": "samples/kWh", "value": 812248.0268358944 } } } }