IlyasMoutawwakil
HF Staff
Upload cuda_inference_diffusers_text-to-image_hf-internal-testing/tiny-stable-diffusion-torch/benchmark.json with huggingface_hub
ca154a6
verified
| { | |
| "config": { | |
| "name": "cuda_inference_diffusers_text-to-image_hf-internal-testing/tiny-stable-diffusion-torch", | |
| "backend": { | |
| "name": "pytorch", | |
| "version": "2.7.0+rocm6.3", | |
| "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", | |
| "model": "hf-internal-testing/tiny-stable-diffusion-torch", | |
| "processor": "hf-internal-testing/tiny-stable-diffusion-torch", | |
| "task": "text-to-image", | |
| "library": "diffusers", | |
| "model_type": "stable-diffusion", | |
| "device": "cuda", | |
| "device_ids": "5", | |
| "seed": 42, | |
| "inter_op_num_threads": null, | |
| "intra_op_num_threads": null, | |
| "model_kwargs": {}, | |
| "processor_kwargs": {}, | |
| "no_weights": false, | |
| "tp_plan": null, | |
| "device_map": null, | |
| "torch_dtype": null, | |
| "eval_mode": true, | |
| "to_bettertransformer": false, | |
| "low_cpu_mem_usage": null, | |
| "attn_implementation": null, | |
| "cache_implementation": null, | |
| "allow_tf32": false, | |
| "autocast_enabled": false, | |
| "autocast_dtype": null, | |
| "torch_compile": false, | |
| "torch_compile_target": "forward", | |
| "torch_compile_config": {}, | |
| "quantization_scheme": null, | |
| "quantization_config": {}, | |
| "deepspeed_inference": false, | |
| "deepspeed_inference_config": {}, | |
| "peft_type": null, | |
| "peft_config": {} | |
| }, | |
| "scenario": { | |
| "name": "inference", | |
| "_target_": "optimum_benchmark.scenarios.inference.scenario.InferenceScenario", | |
| "iterations": 1, | |
| "duration": 1, | |
| "warmup_runs": 1, | |
| "input_shapes": { | |
| "batch_size": 2, | |
| "sequence_length": 16, | |
| "num_choices": 2 | |
| }, | |
| "new_tokens": null, | |
| "memory": true, | |
| "latency": true, | |
| "energy": false, | |
| "forward_kwargs": {}, | |
| "generate_kwargs": { | |
| "max_new_tokens": 2, | |
| "min_new_tokens": 2 | |
| }, | |
| "call_kwargs": { | |
| "num_inference_steps": 2 | |
| } | |
| }, | |
| "launcher": { | |
| "name": "process", | |
| "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", | |
| "device_isolation": true, | |
| "device_isolation_action": "warn", | |
| "numactl": false, | |
| "numactl_kwargs": {}, | |
| "start_method": "spawn" | |
| }, | |
| "environment": { | |
| "cpu": " AMD EPYC 7763 64-Core Processor", | |
| "cpu_count": 128, | |
| "cpu_ram_mb": 1082014.486528, | |
| "system": "Linux", | |
| "machine": "x86_64", | |
| "platform": "Linux-5.15.0-131-generic-x86_64-with-glibc2.35", | |
| "processor": "x86_64", | |
| "python_version": "3.10.12", | |
| "gpu": [ | |
| "Advanced Micro Devices, Inc. [AMD/ATI]", | |
| "Advanced Micro Devices, Inc. [AMD/ATI]", | |
| "Advanced Micro Devices, Inc. [AMD/ATI]", | |
| "Advanced Micro Devices, Inc. [AMD/ATI]", | |
| "Advanced Micro Devices, Inc. [AMD/ATI]", | |
| "Advanced Micro Devices, Inc. [AMD/ATI]", | |
| "Advanced Micro Devices, Inc. [AMD/ATI]", | |
| "Advanced Micro Devices, Inc. [AMD/ATI]" | |
| ], | |
| "gpu_count": 8, | |
| "gpu_vram_mb": 549621596160, | |
| "optimum_benchmark_version": "0.6.0.dev0", | |
| "optimum_benchmark_commit": null, | |
| "transformers_version": "4.51.3", | |
| "transformers_commit": null, | |
| "accelerate_version": "1.6.0", | |
| "accelerate_commit": null, | |
| "diffusers_version": "0.33.1", | |
| "diffusers_commit": null, | |
| "optimum_version": null, | |
| "optimum_commit": null, | |
| "timm_version": "1.0.15", | |
| "timm_commit": null, | |
| "peft_version": null, | |
| "peft_commit": null | |
| }, | |
| "print_report": true, | |
| "log_report": true | |
| }, | |
| "report": { | |
| "load_model": { | |
| "memory": { | |
| "unit": "MB", | |
| "max_ram": 1939.08736, | |
| "max_global_vram": 68702.69952, | |
| "max_process_vram": 0.0, | |
| "max_reserved": 10.48576, | |
| "max_allocated": 8.715776 | |
| }, | |
| "latency": { | |
| "unit": "s", | |
| "values": [ | |
| 0.6740250244140625 | |
| ], | |
| "count": 1, | |
| "total": 0.6740250244140625, | |
| "mean": 0.6740250244140625, | |
| "p50": 0.6740250244140625, | |
| "p90": 0.6740250244140625, | |
| "p95": 0.6740250244140625, | |
| "p99": 0.6740250244140625, | |
| "stdev": 0, | |
| "stdev_": 0 | |
| }, | |
| "throughput": null, | |
| "energy": null, | |
| "efficiency": null | |
| }, | |
| "first_call": { | |
| "memory": { | |
| "unit": "MB", | |
| "max_ram": 2373.77536, | |
| "max_global_vram": 68702.69952, | |
| "max_process_vram": 0.0, | |
| "max_reserved": 182.452224, | |
| "max_allocated": 113.242624 | |
| }, | |
| "latency": { | |
| "unit": "s", | |
| "values": [ | |
| 1.7953411865234374 | |
| ], | |
| "count": 1, | |
| "total": 1.7953411865234374, | |
| "mean": 1.7953411865234374, | |
| "p50": 1.7953411865234374, | |
| "p90": 1.7953411865234374, | |
| "p95": 1.7953411865234374, | |
| "p99": 1.7953411865234374, | |
| "stdev": 0, | |
| "stdev_": 0 | |
| }, | |
| "throughput": null, | |
| "energy": null, | |
| "efficiency": null | |
| }, | |
| "call": { | |
| "memory": { | |
| "unit": "MB", | |
| "max_ram": 2387.341312, | |
| "max_global_vram": 68702.69952, | |
| "max_process_vram": 0.0, | |
| "max_reserved": 182.452224, | |
| "max_allocated": 113.242624 | |
| }, | |
| "latency": { | |
| "unit": "s", | |
| "values": [ | |
| 0.06926068115234375, | |
| 0.06829940032958984, | |
| 0.06837396240234375, | |
| 0.07087155151367187, | |
| 0.06858867645263672, | |
| 0.067856201171875, | |
| 0.06710516357421875, | |
| 0.0662763671875, | |
| 0.06395268630981445, | |
| 0.06378420639038086, | |
| 0.06152741241455078, | |
| 0.059811576843261716, | |
| 0.060093814849853516, | |
| 0.06645668029785157, | |
| 0.06131397247314453, | |
| 0.062130290985107424 | |
| ], | |
| "count": 16, | |
| "total": 1.0457026443481445, | |
| "mean": 0.06535641527175903, | |
| "p50": 0.06636652374267579, | |
| "p90": 0.06892467880249023, | |
| "p95": 0.06966339874267578, | |
| "p99": 0.07062992095947265, | |
| "stdev": 0.003441758321290376, | |
| "stdev_": 5.266136930826412 | |
| }, | |
| "throughput": { | |
| "unit": "images/s", | |
| "value": 30.601433565225143 | |
| }, | |
| "energy": null, | |
| "efficiency": null | |
| }, | |
| "per_step": { | |
| "memory": null, | |
| "latency": { | |
| "unit": "s", | |
| "values": [ | |
| 0.016453731536865236, | |
| 0.01706077003479004, | |
| 0.016453731536865236, | |
| 0.01706077003479004, | |
| 0.03484714126586914, | |
| 0.016763969421386717, | |
| 0.016877731323242186, | |
| 0.016453731536865236, | |
| 0.01706077003479004, | |
| 0.03484714126586914, | |
| 0.016763969421386717, | |
| 0.016877731323242186, | |
| 0.034414821624755856, | |
| 0.016995010375976564, | |
| 0.016857889175415038, | |
| 0.016453731536865236, | |
| 0.01706077003479004, | |
| 0.03484714126586914, | |
| 0.016763969421386717, | |
| 0.016877731323242186, | |
| 0.034414821624755856, | |
| 0.016995010375976564, | |
| 0.016857889175415038, | |
| 0.03678169631958008, | |
| 0.01740572738647461, | |
| 0.016774051666259764, | |
| 0.016453731536865236, | |
| 0.01706077003479004, | |
| 0.03484714126586914, | |
| 0.016763969421386717, | |
| 0.016877731323242186, | |
| 0.034414821624755856, | |
| 0.016995010375976564, | |
| 0.016857889175415038, | |
| 0.03678169631958008, | |
| 0.01740572738647461, | |
| 0.016774051666259764, | |
| 0.03524313735961914, | |
| 0.016712289810180664, | |
| 0.016640451431274413, | |
| 0.016453731536865236, | |
| 0.01706077003479004, | |
| 0.03484714126586914, | |
| 0.016763969421386717, | |
| 0.016877731323242186, | |
| 0.034414821624755856, | |
| 0.016995010375976564, | |
| 0.016857889175415038, | |
| 0.03678169631958008, | |
| 0.01740572738647461, | |
| 0.016774051666259764, | |
| 0.03524313735961914, | |
| 0.016712289810180664, | |
| 0.016640451431274413, | |
| 0.0347386589050293, | |
| 0.016635650634765624, | |
| 0.016548290252685546, | |
| 0.016453731536865236, | |
| 0.01706077003479004, | |
| 0.03484714126586914, | |
| 0.016763969421386717, | |
| 0.016877731323242186, | |
| 0.034414821624755856, | |
| 0.016995010375976564, | |
| 0.016857889175415038, | |
| 0.03678169631958008, | |
| 0.01740572738647461, | |
| 0.016774051666259764, | |
| 0.03524313735961914, | |
| 0.016712289810180664, | |
| 0.016640451431274413, | |
| 0.0347386589050293, | |
| 0.016635650634765624, | |
| 0.016548290252685546, | |
| 0.03451721954345703, | |
| 0.016591489791870116, | |
| 0.01598829174041748, | |
| 0.016453731536865236, | |
| 0.01706077003479004, | |
| 0.03484714126586914, | |
| 0.016763969421386717, | |
| 0.016877731323242186, | |
| 0.034414821624755856, | |
| 0.016995010375976564, | |
| 0.016857889175415038, | |
| 0.03678169631958008, | |
| 0.01740572738647461, | |
| 0.016774051666259764, | |
| 0.03524313735961914, | |
| 0.016712289810180664, | |
| 0.016640451431274413, | |
| 0.0347386589050293, | |
| 0.016635650634765624, | |
| 0.016548290252685546, | |
| 0.03451721954345703, | |
| 0.016591489791870116, | |
| 0.01598829174041748, | |
| 0.033906978607177736, | |
| 0.016644609451293944, | |
| 0.015609413146972657, | |
| 0.016453731536865236, | |
| 0.01706077003479004, | |
| 0.03484714126586914, | |
| 0.016763969421386717, | |
| 0.016877731323242186, | |
| 0.034414821624755856, | |
| 0.016995010375976564, | |
| 0.016857889175415038, | |
| 0.03678169631958008, | |
| 0.01740572738647461, | |
| 0.016774051666259764, | |
| 0.03524313735961914, | |
| 0.016712289810180664, | |
| 0.016640451431274413, | |
| 0.0347386589050293, | |
| 0.016635650634765624, | |
| 0.016548290252685546, | |
| 0.03451721954345703, | |
| 0.016591489791870116, | |
| 0.01598829174041748, | |
| 0.033906978607177736, | |
| 0.016644609451293944, | |
| 0.015609413146972657, | |
| 0.03292794036865234, | |
| 0.01585565185546875, | |
| 0.015533573150634765, | |
| 0.016453731536865236, | |
| 0.01706077003479004, | |
| 0.03484714126586914, | |
| 0.016763969421386717, | |
| 0.016877731323242186, | |
| 0.034414821624755856, | |
| 0.016995010375976564, | |
| 0.016857889175415038, | |
| 0.03678169631958008, | |
| 0.01740572738647461, | |
| 0.016774051666259764, | |
| 0.03524313735961914, | |
| 0.016712289810180664, | |
| 0.016640451431274413, | |
| 0.0347386589050293, | |
| 0.016635650634765624, | |
| 0.016548290252685546, | |
| 0.03451721954345703, | |
| 0.016591489791870116, | |
| 0.01598829174041748, | |
| 0.033906978607177736, | |
| 0.016644609451293944, | |
| 0.015609413146972657, | |
| 0.03292794036865234, | |
| 0.01585565185546875, | |
| 0.015533573150634765, | |
| 0.03263946151733398, | |
| 0.015603652954101562, | |
| 0.015543172836303711, | |
| 0.016453731536865236, | |
| 0.01706077003479004, | |
| 0.03484714126586914, | |
| 0.016763969421386717, | |
| 0.016877731323242186, | |
| 0.034414821624755856, | |
| 0.016995010375976564, | |
| 0.016857889175415038, | |
| 0.03678169631958008, | |
| 0.01740572738647461, | |
| 0.016774051666259764, | |
| 0.03524313735961914, | |
| 0.016712289810180664, | |
| 0.016640451431274413, | |
| 0.0347386589050293, | |
| 0.016635650634765624, | |
| 0.016548290252685546, | |
| 0.03451721954345703, | |
| 0.016591489791870116, | |
| 0.01598829174041748, | |
| 0.033906978607177736, | |
| 0.016644609451293944, | |
| 0.015609413146972657, | |
| 0.03292794036865234, | |
| 0.01585565185546875, | |
| 0.015533573150634765, | |
| 0.03263946151733398, | |
| 0.015603652954101562, | |
| 0.015543172836303711, | |
| 0.032565380096435545, | |
| 0.01464157485961914, | |
| 0.014603174209594727, | |
| 0.016453731536865236, | |
| 0.01706077003479004, | |
| 0.03484714126586914, | |
| 0.016763969421386717, | |
| 0.016877731323242186, | |
| 0.034414821624755856, | |
| 0.016995010375976564, | |
| 0.016857889175415038, | |
| 0.03678169631958008, | |
| 0.01740572738647461, | |
| 0.016774051666259764, | |
| 0.03524313735961914, | |
| 0.016712289810180664, | |
| 0.016640451431274413, | |
| 0.0347386589050293, | |
| 0.016635650634765624, | |
| 0.016548290252685546, | |
| 0.03451721954345703, | |
| 0.016591489791870116, | |
| 0.01598829174041748, | |
| 0.033906978607177736, | |
| 0.016644609451293944, | |
| 0.015609413146972657, | |
| 0.03292794036865234, | |
| 0.01585565185546875, | |
| 0.015533573150634765, | |
| 0.03263946151733398, | |
| 0.015603652954101562, | |
| 0.015543172836303711, | |
| 0.032565380096435545, | |
| 0.01464157485961914, | |
| 0.014603174209594727, | |
| 0.03065882682800293, | |
| 0.014890853881835937, | |
| 0.01459965419769287, | |
| 0.016453731536865236, | |
| 0.01706077003479004, | |
| 0.03484714126586914, | |
| 0.016763969421386717, | |
| 0.016877731323242186, | |
| 0.034414821624755856, | |
| 0.016995010375976564, | |
| 0.016857889175415038, | |
| 0.03678169631958008, | |
| 0.01740572738647461, | |
| 0.016774051666259764, | |
| 0.03524313735961914, | |
| 0.016712289810180664, | |
| 0.016640451431274413, | |
| 0.0347386589050293, | |
| 0.016635650634765624, | |
| 0.016548290252685546, | |
| 0.03451721954345703, | |
| 0.016591489791870116, | |
| 0.01598829174041748, | |
| 0.033906978607177736, | |
| 0.016644609451293944, | |
| 0.015609413146972657, | |
| 0.03292794036865234, | |
| 0.01585565185546875, | |
| 0.015533573150634765, | |
| 0.03263946151733398, | |
| 0.015603652954101562, | |
| 0.015543172836303711, | |
| 0.032565380096435545, | |
| 0.01464157485961914, | |
| 0.014603174209594727, | |
| 0.03065882682800293, | |
| 0.014890853881835937, | |
| 0.01459965419769287, | |
| 0.030262027740478516, | |
| 0.014575493812561036, | |
| 0.014565094947814942, | |
| 0.016453731536865236, | |
| 0.01706077003479004, | |
| 0.03484714126586914, | |
| 0.016763969421386717, | |
| 0.016877731323242186, | |
| 0.034414821624755856, | |
| 0.016995010375976564, | |
| 0.016857889175415038, | |
| 0.03678169631958008, | |
| 0.01740572738647461, | |
| 0.016774051666259764, | |
| 0.03524313735961914, | |
| 0.016712289810180664, | |
| 0.016640451431274413, | |
| 0.0347386589050293, | |
| 0.016635650634765624, | |
| 0.016548290252685546, | |
| 0.03451721954345703, | |
| 0.016591489791870116, | |
| 0.01598829174041748, | |
| 0.033906978607177736, | |
| 0.016644609451293944, | |
| 0.015609413146972657, | |
| 0.03292794036865234, | |
| 0.01585565185546875, | |
| 0.015533573150634765, | |
| 0.03263946151733398, | |
| 0.015603652954101562, | |
| 0.015543172836303711, | |
| 0.032565380096435545, | |
| 0.01464157485961914, | |
| 0.014603174209594727, | |
| 0.03065882682800293, | |
| 0.014890853881835937, | |
| 0.01459965419769287, | |
| 0.030262027740478516, | |
| 0.014575493812561036, | |
| 0.014565094947814942, | |
| 0.0349791374206543, | |
| 0.01609805107116699, | |
| 0.015659811973571778, | |
| 0.016453731536865236, | |
| 0.01706077003479004, | |
| 0.03484714126586914, | |
| 0.016763969421386717, | |
| 0.016877731323242186, | |
| 0.034414821624755856, | |
| 0.016995010375976564, | |
| 0.016857889175415038, | |
| 0.03678169631958008, | |
| 0.01740572738647461, | |
| 0.016774051666259764, | |
| 0.03524313735961914, | |
| 0.016712289810180664, | |
| 0.016640451431274413, | |
| 0.0347386589050293, | |
| 0.016635650634765624, | |
| 0.016548290252685546, | |
| 0.03451721954345703, | |
| 0.016591489791870116, | |
| 0.01598829174041748, | |
| 0.033906978607177736, | |
| 0.016644609451293944, | |
| 0.015609413146972657, | |
| 0.03292794036865234, | |
| 0.01585565185546875, | |
| 0.015533573150634765, | |
| 0.03263946151733398, | |
| 0.015603652954101562, | |
| 0.015543172836303711, | |
| 0.032565380096435545, | |
| 0.01464157485961914, | |
| 0.014603174209594727, | |
| 0.03065882682800293, | |
| 0.014890853881835937, | |
| 0.01459965419769287, | |
| 0.030262027740478516, | |
| 0.014575493812561036, | |
| 0.014565094947814942, | |
| 0.0349791374206543, | |
| 0.01609805107116699, | |
| 0.015659811973571778, | |
| 0.031633224487304684, | |
| 0.015035813331604003, | |
| 0.01505549430847168, | |
| 0.016453731536865236, | |
| 0.01706077003479004, | |
| 0.03484714126586914, | |
| 0.016763969421386717, | |
| 0.016877731323242186, | |
| 0.034414821624755856, | |
| 0.016995010375976564, | |
| 0.016857889175415038, | |
| 0.03678169631958008, | |
| 0.01740572738647461, | |
| 0.016774051666259764, | |
| 0.03524313735961914, | |
| 0.016712289810180664, | |
| 0.016640451431274413, | |
| 0.0347386589050293, | |
| 0.016635650634765624, | |
| 0.016548290252685546, | |
| 0.03451721954345703, | |
| 0.016591489791870116, | |
| 0.01598829174041748, | |
| 0.033906978607177736, | |
| 0.016644609451293944, | |
| 0.015609413146972657, | |
| 0.03292794036865234, | |
| 0.01585565185546875, | |
| 0.015533573150634765, | |
| 0.03263946151733398, | |
| 0.015603652954101562, | |
| 0.015543172836303711, | |
| 0.032565380096435545, | |
| 0.01464157485961914, | |
| 0.014603174209594727, | |
| 0.03065882682800293, | |
| 0.014890853881835937, | |
| 0.01459965419769287, | |
| 0.030262027740478516, | |
| 0.014575493812561036, | |
| 0.014565094947814942, | |
| 0.0349791374206543, | |
| 0.01609805107116699, | |
| 0.015659811973571778, | |
| 0.031633224487304684, | |
| 0.015035813331604003, | |
| 0.01505549430847168, | |
| 0.03139082336425781, | |
| 0.015356773376464844, | |
| 0.015158373832702637 | |
| ], | |
| "count": 392, | |
| "total": 8.534355380058287, | |
| "mean": 0.021771314745046653, | |
| "p50": 0.016774051666259764, | |
| "p90": 0.03484714126586914, | |
| "p95": 0.03524313735961914, | |
| "p99": 0.03678169631958008, | |
| "stdev": 0.008301850685450163, | |
| "stdev_": 38.13205946755685 | |
| }, | |
| "throughput": null, | |
| "energy": null, | |
| "efficiency": null | |
| } | |
| } | |
| } |