diff --git a/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-02-39/.hydra/config.yaml b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-02-39/.hydra/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..ca9c7622983c00dc182a28ef5197b99186dbcca4 --- /dev/null +++ b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-02-39/.hydra/config.yaml @@ -0,0 +1,94 @@ +backend: + name: pytorch + version: 2.4.0 + _target_: optimum_benchmark.backends.pytorch.backend.PyTorchBackend + task: text-classification + model: lvwerra/distilbert-imdb + processor: lvwerra/distilbert-imdb + library: null + device: cuda + device_ids: '0' + seed: 42 + inter_op_num_threads: null + intra_op_num_threads: null + hub_kwargs: {} + no_weights: true + device_map: null + torch_dtype: null + amp_autocast: false + amp_dtype: null + eval_mode: true + to_bettertransformer: false + low_cpu_mem_usage: null + attn_implementation: null + cache_implementation: null + torch_compile: false + torch_compile_config: {} + quantization_scheme: null + quantization_config: {} + deepspeed_inference: false + deepspeed_inference_config: {} + peft_type: null + peft_config: {} +launcher: + name: process + _target_: optimum_benchmark.launchers.process.launcher.ProcessLauncher + device_isolation: true + device_isolation_action: warn + start_method: spawn +benchmark: + name: energy_star + _target_: optimum_benchmark.benchmarks.energy_star.benchmark.EnergyStarBenchmark + dataset_name: EnergyStarAI/text_classification + dataset_config: '' + dataset_split: train + num_samples: 1000 + input_shapes: + batch_size: 1 + text_column_name: text + truncation: true + max_length: -1 + dataset_prefix1: '' + dataset_prefix2: '' + t5_task: '' + image_column_name: image + resize: false + question_column_name: question + context_column_name: context + sentence1_column_name: sentence1 + sentence2_column_name: sentence2 + audio_column_name: audio + iterations: 10 + warmup_runs: 10 + energy: true + forward_kwargs: {} + generate_kwargs: {} + call_kwargs: {} +experiment_name: text_classification +environment: + cpu: ' AMD EPYC 7R32' + cpu_count: 48 + cpu_ram_mb: 200472.73984 + system: Linux + machine: x86_64 + platform: Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35 + processor: x86_64 + python_version: 3.9.20 + gpu: + - NVIDIA A10G + gpu_count: 1 + gpu_vram_mb: 24146608128 + optimum_benchmark_version: 0.2.0 + optimum_benchmark_commit: null + transformers_version: 4.44.0 + transformers_commit: null + accelerate_version: 0.33.0 + accelerate_commit: null + diffusers_version: 0.30.0 + diffusers_commit: null + optimum_version: null + optimum_commit: null + timm_version: null + timm_commit: null + peft_version: null + peft_commit: null diff --git a/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-02-39/.hydra/hydra.yaml b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-02-39/.hydra/hydra.yaml new file mode 100644 index 0000000000000000000000000000000000000000..a9b387d70e4b577da55623bb0b4dbcd654f799be --- /dev/null +++ b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-02-39/.hydra/hydra.yaml @@ -0,0 +1,173 @@ +hydra: + run: + dir: /data/runs/${experiment_name}/${backend.model}/${now:%Y-%m-%d-%H-%M-%S} + sweep: + dir: sweeps/${experiment_name}/${now:%Y-%m-%d-%H-%M-%S} + subdir: ${hydra.job.num} + launcher: + _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher + sweeper: + _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper + max_batch_size: null + params: null + help: + app_name: ${hydra.job.name} + header: '${hydra.help.app_name} is powered by Hydra. + + ' + footer: 'Powered by Hydra (https://hydra.cc) + + Use --hydra-help to view Hydra specific help + + ' + template: '${hydra.help.header} + + == Configuration groups == + + Compose your configuration from those groups (group=option) + + + $APP_CONFIG_GROUPS + + + == Config == + + Override anything in the config (foo.bar=value) + + + $CONFIG + + + ${hydra.help.footer} + + ' + hydra_help: + template: 'Hydra (${hydra.runtime.version}) + + See https://hydra.cc for more info. + + + == Flags == + + $FLAGS_HELP + + + == Configuration groups == + + Compose your configuration from those groups (For example, append hydra/job_logging=disabled + to command line) + + + $HYDRA_CONFIG_GROUPS + + + Use ''--cfg hydra'' to Show the Hydra config. + + ' + hydra_help: ??? + hydra_logging: + version: 1 + formatters: + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s' + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + root: + level: INFO + handlers: + - console + disable_existing_loggers: false + job_logging: + version: 1 + formatters: + simple: + format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s' + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s] + - %(message)s' + log_colors: + DEBUG: purple + INFO: green + WARNING: yellow + ERROR: red + CRITICAL: red + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + file: + class: logging.FileHandler + formatter: simple + filename: ${hydra.job.name}.log + root: + level: INFO + handlers: + - console + - file + disable_existing_loggers: false + env: {} + mode: RUN + searchpath: [] + callbacks: {} + output_subdir: .hydra + overrides: + hydra: + - hydra.run.dir=/data/runs/${experiment_name}/${backend.model}/${now:%Y-%m-%d-%H-%M-%S} + - hydra.mode=RUN + task: [] + job: + name: cli + chdir: true + override_dirname: '' + id: ??? + num: ??? + config_name: text_classification + env_set: + OVERRIDE_BENCHMARKS: '1' + env_copy: [] + config: + override_dirname: + kv_sep: '=' + item_sep: ',' + exclude_keys: [] + runtime: + version: 1.3.2 + version_base: '1.3' + cwd: / + config_sources: + - path: hydra.conf + schema: pkg + provider: hydra + - path: optimum_benchmark + schema: pkg + provider: main + - path: hydra_plugins.hydra_colorlog.conf + schema: pkg + provider: hydra-colorlog + - path: /optimum-benchmark/examples/energy_star + schema: file + provider: command-line + - path: '' + schema: structured + provider: schema + output_dir: /data/runs/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-02-39 + choices: + benchmark: energy_star + launcher: process + backend: pytorch + hydra/env: default + hydra/callbacks: null + hydra/job_logging: colorlog + hydra/hydra_logging: colorlog + hydra/hydra_help: default + hydra/help: default + hydra/sweeper: basic + hydra/launcher: basic + hydra/output: default + verbose: false diff --git a/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-02-39/.hydra/overrides.yaml b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-02-39/.hydra/overrides.yaml new file mode 100644 index 0000000000000000000000000000000000000000..fe51488c7066f6687ef680d6bfaa4f7768ef205c --- /dev/null +++ b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-02-39/.hydra/overrides.yaml @@ -0,0 +1 @@ +[] diff --git a/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-02-39/benchmark_report.json b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-02-39/benchmark_report.json new file mode 100644 index 0000000000000000000000000000000000000000..da8c3fb06247799a5049da419bc4972959938eef --- /dev/null +++ b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-02-39/benchmark_report.json @@ -0,0 +1,107 @@ +{ + "forward": { + "memory": null, + "latency": null, + "throughput": null, + "energy": { + "unit": "kWh", + "cpu": 4.3033950658298426e-05, + "ram": 3.491482350135866e-07, + "gpu": 0.00016699630026373313, + "total": 0.00021037939915704516 + }, + "efficiency": { + "unit": "samples/kWh", + "value": 4753317.12138561 + }, + "measures": [ + { + "unit": "kWh", + "cpu": 4.7981154024859693e-05, + "ram": 3.8929762519253783e-07, + "gpu": 0.000184299314106795, + "total": 0.00023266976575684725 + }, + { + "unit": "kWh", + "cpu": 4.7782428311378075e-05, + "ram": 3.877802845256958e-07, + "gpu": 0.00018138903400100048, + "total": 0.00022955924259690426 + }, + { + "unit": "kWh", + "cpu": 4.7605167399807316e-05, + "ram": 3.8636395221827183e-07, + "gpu": 0.000186995705149684, + "total": 0.00023498723650170952 + }, + { + "unit": "kWh", + "cpu": 4.855238827456359e-05, + "ram": 3.935672548850149e-07, + "gpu": 0.0001856968152260663, + "total": 0.00023464277075551483 + }, + { + "unit": "kWh", + "cpu": 4.857113074960783e-05, + "ram": 3.9379859018887915e-07, + "gpu": 0.00018688737173278014, + "total": 0.00023585230107257675 + }, + { + "unit": "kWh", + "cpu": 4.7622314353551096e-05, + "ram": 3.8650354212378535e-07, + "gpu": 0.00018486348122337404, + "total": 0.00023287229911904904 + }, + { + "unit": "kWh", + "cpu": 0.0, + "ram": 0.0, + "gpu": 0.0, + "total": 0.0 + }, + { + "unit": "kWh", + "cpu": 4.769037301723274e-05, + "ram": 3.870572522958343e-07, + "gpu": 0.00018588431537125416, + "total": 0.00023396174564078282 + }, + { + "unit": "kWh", + "cpu": 4.7270123168224615e-05, + "ram": 3.835115881599748e-07, + "gpu": 0.00018772459462468305, + "total": 0.0002353782293810676 + }, + { + "unit": "kWh", + "cpu": 4.726442728375936e-05, + "ram": 3.836022605458718e-07, + "gpu": 0.00018622237120169416, + "total": 0.0002338704007459995 + } + ] + }, + "preprocess": { + "memory": null, + "latency": null, + "throughput": null, + "energy": { + "unit": "kWh", + "cpu": 5.090969423731116e-06, + "ram": 3.264302126335614e-08, + "gpu": 7.466117082799428e-06, + "total": 1.2589729527793902e-05 + }, + "efficiency": { + "unit": "samples/kWh", + "value": 79429823.95232044 + }, + "measures": null + } +} \ No newline at end of file diff --git a/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-02-39/cli.log b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-02-39/cli.log new file mode 100644 index 0000000000000000000000000000000000000000..d979f7814a386643d7eef55e9b88275a3c283fdc --- /dev/null +++ b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-02-39/cli.log @@ -0,0 +1,113 @@ +[2024-10-01 19:02:39,712][launcher][INFO] - ََAllocating process launcher +[2024-10-01 19:02:39,712][process][INFO] - + Setting multiprocessing start method to spawn. +[2024-10-01 19:02:39,725][device-isolation][INFO] - + Launched device(s) isolation process 66 +[2024-10-01 19:02:39,726][device-isolation][INFO] - + Isolating device(s) [0] +[2024-10-01 19:02:39,732][process][INFO] - + Launched benchmark in isolated process 67. +[PROC-0][2024-10-01 19:02:42,331][datasets][INFO] - PyTorch version 2.4.0 available. +[PROC-0][2024-10-01 19:02:43,278][backend][INFO] - َAllocating pytorch backend +[PROC-0][2024-10-01 19:02:43,278][backend][INFO] - + Setting random seed to 42 +[PROC-0][2024-10-01 19:02:43,590][pytorch][INFO] - + Using AutoModel class AutoModelForSequenceClassification +[PROC-0][2024-10-01 19:02:43,591][pytorch][INFO] - + Creating backend temporary directory +[PROC-0][2024-10-01 19:02:43,591][pytorch][INFO] - + Loading model with random weights +[PROC-0][2024-10-01 19:02:43,591][pytorch][INFO] - + Creating no weights model +[PROC-0][2024-10-01 19:02:43,591][pytorch][INFO] - + Creating no weights model directory +[PROC-0][2024-10-01 19:02:43,591][pytorch][INFO] - + Creating no weights model state dict +[PROC-0][2024-10-01 19:02:43,593][pytorch][INFO] - + Saving no weights model safetensors +[PROC-0][2024-10-01 19:02:43,594][pytorch][INFO] - + Saving no weights model pretrained config +[PROC-0][2024-10-01 19:02:43,595][pytorch][INFO] - + Loading no weights AutoModel +[PROC-0][2024-10-01 19:02:43,595][pytorch][INFO] - + Loading model directly on device: cuda +[PROC-0][2024-10-01 19:02:43,896][pytorch][INFO] - + Turning on model's eval mode +[PROC-0][2024-10-01 19:02:43,902][benchmark][INFO] - Allocating energy_star benchmark +[PROC-0][2024-10-01 19:02:43,902][energy_star][INFO] - + Loading raw dataset +[PROC-0][2024-10-01 19:02:44,447][energy_star][INFO] - + Initializing Inference report +[PROC-0][2024-10-01 19:02:44,448][energy][INFO] - + Tracking GPU energy on devices [0] +[PROC-0][2024-10-01 19:02:48,663][energy_star][INFO] - + Preprocessing dataset +[PROC-0][2024-10-01 19:02:49,098][energy][INFO] - + Saving codecarbon emission data to preprocess_codecarbon.json +[PROC-0][2024-10-01 19:02:49,100][energy_star][INFO] - + Preparing backend for Inference +[PROC-0][2024-10-01 19:02:49,100][energy_star][INFO] - + Initialising dataloader +[PROC-0][2024-10-01 19:02:49,100][energy_star][INFO] - + Warming up backend for Inference +[PROC-0][2024-10-01 19:02:49,663][energy_star][INFO] - + Running Inference energy tracking for 10 iterations +[PROC-0][2024-10-01 19:02:49,663][energy_star][INFO] - + Iteration 1/10 +[PROC-0][2024-10-01 19:02:53,729][energy][INFO] - + Saving codecarbon emission data to forward_codecarbon.json +[PROC-0][2024-10-01 19:02:53,731][energy_star][INFO] - + Iteration 2/10 +[PROC-0][2024-10-01 19:02:57,781][energy][INFO] - + Saving codecarbon emission data to forward_codecarbon.json +[PROC-0][2024-10-01 19:02:57,784][energy_star][INFO] - + Iteration 3/10 +[PROC-0][2024-10-01 19:03:01,819][energy][INFO] - + Saving codecarbon emission data to forward_codecarbon.json +[PROC-0][2024-10-01 19:03:01,821][energy_star][INFO] - + Iteration 4/10 +[PROC-0][2024-10-01 19:03:05,936][energy][INFO] - + Saving codecarbon emission data to forward_codecarbon.json +[PROC-0][2024-10-01 19:03:05,938][energy_star][INFO] - + Iteration 5/10 +[PROC-0][2024-10-01 19:03:10,054][energy][INFO] - + Saving codecarbon emission data to forward_codecarbon.json +[PROC-0][2024-10-01 19:03:10,057][energy_star][INFO] - + Iteration 6/10 +[PROC-0][2024-10-01 19:03:14,093][energy][INFO] - + Saving codecarbon emission data to forward_codecarbon.json +[PROC-0][2024-10-01 19:03:14,095][energy_star][INFO] - + Iteration 7/10 +[PROC-0][2024-10-01 19:03:18,137][energy][INFO] - + Saving codecarbon emission data to forward_codecarbon.json +[PROC-0][2024-10-01 19:03:18,139][energy_star][INFO] - + Iteration 8/10 +[PROC-0][2024-10-01 19:03:22,181][energy][INFO] - + Saving codecarbon emission data to forward_codecarbon.json +[PROC-0][2024-10-01 19:03:22,183][energy_star][INFO] - + Iteration 9/10 +[PROC-0][2024-10-01 19:03:26,190][energy][INFO] - + Saving codecarbon emission data to forward_codecarbon.json +[PROC-0][2024-10-01 19:03:26,192][energy_star][INFO] - + Iteration 10/10 +[PROC-0][2024-10-01 19:03:30,198][energy][INFO] - + Saving codecarbon emission data to forward_codecarbon.json +[PROC-0][2024-10-01 19:03:30,200][energy][INFO] - + forward energy consumption: +[PROC-0][2024-10-01 19:03:30,200][energy][INFO] - + CPU: 0.000043 (kWh) +[PROC-0][2024-10-01 19:03:30,201][energy][INFO] - + GPU: 0.000167 (kWh) +[PROC-0][2024-10-01 19:03:30,201][energy][INFO] - + RAM: 0.000000 (kWh) +[PROC-0][2024-10-01 19:03:30,201][energy][INFO] - + total: 0.000210 (kWh) +[PROC-0][2024-10-01 19:03:30,201][energy][INFO] - + forward_iteration_1 energy consumption: +[PROC-0][2024-10-01 19:03:30,201][energy][INFO] - + CPU: 0.000048 (kWh) +[PROC-0][2024-10-01 19:03:30,201][energy][INFO] - + GPU: 0.000184 (kWh) +[PROC-0][2024-10-01 19:03:30,201][energy][INFO] - + RAM: 0.000000 (kWh) +[PROC-0][2024-10-01 19:03:30,201][energy][INFO] - + total: 0.000233 (kWh) +[PROC-0][2024-10-01 19:03:30,201][energy][INFO] - + forward_iteration_2 energy consumption: +[PROC-0][2024-10-01 19:03:30,201][energy][INFO] - + CPU: 0.000048 (kWh) +[PROC-0][2024-10-01 19:03:30,201][energy][INFO] - + GPU: 0.000181 (kWh) +[PROC-0][2024-10-01 19:03:30,202][energy][INFO] - + RAM: 0.000000 (kWh) +[PROC-0][2024-10-01 19:03:30,202][energy][INFO] - + total: 0.000230 (kWh) +[PROC-0][2024-10-01 19:03:30,202][energy][INFO] - + forward_iteration_3 energy consumption: +[PROC-0][2024-10-01 19:03:30,202][energy][INFO] - + CPU: 0.000048 (kWh) +[PROC-0][2024-10-01 19:03:30,202][energy][INFO] - + GPU: 0.000187 (kWh) +[PROC-0][2024-10-01 19:03:30,202][energy][INFO] - + RAM: 0.000000 (kWh) +[PROC-0][2024-10-01 19:03:30,202][energy][INFO] - + total: 0.000235 (kWh) +[PROC-0][2024-10-01 19:03:30,202][energy][INFO] - + forward_iteration_4 energy consumption: +[PROC-0][2024-10-01 19:03:30,202][energy][INFO] - + CPU: 0.000049 (kWh) +[PROC-0][2024-10-01 19:03:30,202][energy][INFO] - + GPU: 0.000186 (kWh) +[PROC-0][2024-10-01 19:03:30,202][energy][INFO] - + RAM: 0.000000 (kWh) +[PROC-0][2024-10-01 19:03:30,202][energy][INFO] - + total: 0.000235 (kWh) +[PROC-0][2024-10-01 19:03:30,202][energy][INFO] - + forward_iteration_5 energy consumption: +[PROC-0][2024-10-01 19:03:30,203][energy][INFO] - + CPU: 0.000049 (kWh) +[PROC-0][2024-10-01 19:03:30,203][energy][INFO] - + GPU: 0.000187 (kWh) +[PROC-0][2024-10-01 19:03:30,203][energy][INFO] - + RAM: 0.000000 (kWh) +[PROC-0][2024-10-01 19:03:30,203][energy][INFO] - + total: 0.000236 (kWh) +[PROC-0][2024-10-01 19:03:30,203][energy][INFO] - + forward_iteration_6 energy consumption: +[PROC-0][2024-10-01 19:03:30,203][energy][INFO] - + CPU: 0.000048 (kWh) +[PROC-0][2024-10-01 19:03:30,203][energy][INFO] - + GPU: 0.000185 (kWh) +[PROC-0][2024-10-01 19:03:30,203][energy][INFO] - + RAM: 0.000000 (kWh) +[PROC-0][2024-10-01 19:03:30,203][energy][INFO] - + total: 0.000233 (kWh) +[PROC-0][2024-10-01 19:03:30,203][energy][INFO] - + forward_iteration_7 energy consumption: +[PROC-0][2024-10-01 19:03:30,203][energy][INFO] - + CPU: 0.000000 (kWh) +[PROC-0][2024-10-01 19:03:30,203][energy][INFO] - + GPU: 0.000000 (kWh) +[PROC-0][2024-10-01 19:03:30,204][energy][INFO] - + RAM: 0.000000 (kWh) +[PROC-0][2024-10-01 19:03:30,204][energy][INFO] - + total: 0.000000 (kWh) +[PROC-0][2024-10-01 19:03:30,204][energy][INFO] - + forward_iteration_8 energy consumption: +[PROC-0][2024-10-01 19:03:30,204][energy][INFO] - + CPU: 0.000048 (kWh) +[PROC-0][2024-10-01 19:03:30,204][energy][INFO] - + GPU: 0.000186 (kWh) +[PROC-0][2024-10-01 19:03:30,204][energy][INFO] - + RAM: 0.000000 (kWh) +[PROC-0][2024-10-01 19:03:30,204][energy][INFO] - + total: 0.000234 (kWh) +[PROC-0][2024-10-01 19:03:30,204][energy][INFO] - + forward_iteration_9 energy consumption: +[PROC-0][2024-10-01 19:03:30,204][energy][INFO] - + CPU: 0.000047 (kWh) +[PROC-0][2024-10-01 19:03:30,204][energy][INFO] - + GPU: 0.000188 (kWh) +[PROC-0][2024-10-01 19:03:30,204][energy][INFO] - + RAM: 0.000000 (kWh) +[PROC-0][2024-10-01 19:03:30,204][energy][INFO] - + total: 0.000235 (kWh) +[PROC-0][2024-10-01 19:03:30,205][energy][INFO] - + forward_iteration_10 energy consumption: +[PROC-0][2024-10-01 19:03:30,205][energy][INFO] - + CPU: 0.000047 (kWh) +[PROC-0][2024-10-01 19:03:30,205][energy][INFO] - + GPU: 0.000186 (kWh) +[PROC-0][2024-10-01 19:03:30,205][energy][INFO] - + RAM: 0.000000 (kWh) +[PROC-0][2024-10-01 19:03:30,205][energy][INFO] - + total: 0.000234 (kWh) +[PROC-0][2024-10-01 19:03:30,205][energy][INFO] - + preprocess energy consumption: +[PROC-0][2024-10-01 19:03:30,205][energy][INFO] - + CPU: 0.000005 (kWh) +[PROC-0][2024-10-01 19:03:30,205][energy][INFO] - + GPU: 0.000007 (kWh) +[PROC-0][2024-10-01 19:03:30,205][energy][INFO] - + RAM: 0.000000 (kWh) +[PROC-0][2024-10-01 19:03:30,205][energy][INFO] - + total: 0.000013 (kWh) +[PROC-0][2024-10-01 19:03:30,206][energy][INFO] - + forward energy efficiency: 4753317.121386 (samples/kWh) +[PROC-0][2024-10-01 19:03:30,206][energy][INFO] - + preprocess energy efficiency: 79429823.952320 (samples/kWh) +[2024-10-01 19:03:30,837][device-isolation][INFO] - + Closing device(s) isolation process... +[2024-10-01 19:03:30,891][datasets][INFO] - PyTorch version 2.4.0 available. diff --git a/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-02-39/experiment_config.json b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-02-39/experiment_config.json new file mode 100644 index 0000000000000000000000000000000000000000..772aa352d839018f6caa9107fd24acf54618c6aa --- /dev/null +++ b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-02-39/experiment_config.json @@ -0,0 +1,107 @@ +{ + "experiment_name": "text_classification", + "backend": { + "name": "pytorch", + "version": "2.4.0", + "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", + "task": "text-classification", + "model": "lvwerra/distilbert-imdb", + "processor": "lvwerra/distilbert-imdb", + "library": "transformers", + "device": "cuda", + "device_ids": "0", + "seed": 42, + "inter_op_num_threads": null, + "intra_op_num_threads": null, + "hub_kwargs": { + "revision": "main", + "force_download": false, + "local_files_only": false, + "trust_remote_code": true + }, + "no_weights": true, + "device_map": null, + "torch_dtype": null, + "amp_autocast": false, + "amp_dtype": null, + "eval_mode": true, + "to_bettertransformer": false, + "low_cpu_mem_usage": null, + "attn_implementation": null, + "cache_implementation": null, + "torch_compile": false, + "torch_compile_config": {}, + "quantization_scheme": null, + "quantization_config": {}, + "deepspeed_inference": false, + "deepspeed_inference_config": {}, + "peft_type": null, + "peft_config": {} + }, + "launcher": { + "name": "process", + "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", + "device_isolation": true, + "device_isolation_action": "warn", + "start_method": "spawn" + }, + "benchmark": { + "name": "energy_star", + "_target_": "optimum_benchmark.benchmarks.energy_star.benchmark.EnergyStarBenchmark", + "dataset_name": "EnergyStarAI/text_classification", + "dataset_config": "", + "dataset_split": "train", + "num_samples": 1000, + "input_shapes": { + "batch_size": 1 + }, + "text_column_name": "text", + "truncation": true, + "max_length": -1, + "dataset_prefix1": "", + "dataset_prefix2": "", + "t5_task": "", + "image_column_name": "image", + "resize": false, + "question_column_name": "question", + "context_column_name": "context", + "sentence1_column_name": "sentence1", + "sentence2_column_name": "sentence2", + "audio_column_name": "audio", + "iterations": 10, + "warmup_runs": 10, + "energy": true, + "forward_kwargs": {}, + "generate_kwargs": {}, + "call_kwargs": {} + }, + "environment": { + "cpu": " AMD EPYC 7R32", + "cpu_count": 48, + "cpu_ram_mb": 200472.73984, + "system": "Linux", + "machine": "x86_64", + "platform": "Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35", + "processor": "x86_64", + "python_version": "3.9.20", + "gpu": [ + "NVIDIA A10G" + ], + "gpu_count": 1, + "gpu_vram_mb": 24146608128, + "optimum_benchmark_version": "0.2.0", + "optimum_benchmark_commit": null, + "transformers_version": "4.44.0", + "transformers_commit": null, + "accelerate_version": "0.33.0", + "accelerate_commit": null, + "diffusers_version": "0.30.0", + "diffusers_commit": null, + "optimum_version": null, + "optimum_commit": null, + "timm_version": null, + "timm_commit": null, + "peft_version": null, + "peft_commit": null + } +} \ No newline at end of file diff --git a/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-02-39/forward_codecarbon.json b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-02-39/forward_codecarbon.json new file mode 100644 index 0000000000000000000000000000000000000000..7da707bbbfc48b81e44626dba62e2d0195eb34b9 --- /dev/null +++ b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-02-39/forward_codecarbon.json @@ -0,0 +1,33 @@ +{ + "timestamp": "2024-10-01T19:03:30", + "project_name": "codecarbon", + "run_id": "87c20af8-09ee-4935-b57c-c55ba981f72f", + "duration": -1726624715.2796082, + "emissions": 8.632960488643251e-05, + "emissions_rate": 2.1573364543800052e-05, + "cpu_power": 42.5, + "gpu_power": 167.54585718780086, + "ram_power": 0.3451881408691406, + "cpu_energy": 4.726442728375936e-05, + "gpu_energy": 0.00018622237120169416, + "ram_energy": 3.836022605458718e-07, + "energy_consumed": 0.0002338704007459995, + "country_name": "United States", + "country_iso_code": "USA", + "region": "virginia", + "cloud_provider": "", + "cloud_region": "", + "os": "Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35", + "python_version": "3.9.20", + "codecarbon_version": "2.5.1", + "cpu_count": 48, + "cpu_model": "AMD EPYC 7R32", + "gpu_count": 1, + "gpu_model": "1 x NVIDIA A10G", + "longitude": -77.4903, + "latitude": 39.0469, + "ram_total_size": 186.7047882080078, + "tracking_mode": "process", + "on_cloud": "N", + "pue": 1.0 +} \ No newline at end of file diff --git a/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-02-39/preprocess_codecarbon.json b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-02-39/preprocess_codecarbon.json new file mode 100644 index 0000000000000000000000000000000000000000..cb235d083bbd9eec42203962225ee94de72d0de9 --- /dev/null +++ b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-02-39/preprocess_codecarbon.json @@ -0,0 +1,33 @@ +{ + "timestamp": "2024-10-01T19:02:49", + "project_name": "codecarbon", + "run_id": "87c20af8-09ee-4935-b57c-c55ba981f72f", + "duration": -1726624718.8521838, + "emissions": 4.647301977054882e-06, + "emissions_rate": 1.0827489819885996e-05, + "cpu_power": 42.5, + "gpu_power": 62.65163325440906, + "ram_power": 0.27443647384643555, + "cpu_energy": 5.090969423731116e-06, + "gpu_energy": 7.466117082799428e-06, + "ram_energy": 3.264302126335614e-08, + "energy_consumed": 1.2589729527793902e-05, + "country_name": "United States", + "country_iso_code": "USA", + "region": "virginia", + "cloud_provider": "", + "cloud_region": "", + "os": "Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35", + "python_version": "3.9.20", + "codecarbon_version": "2.5.1", + "cpu_count": 48, + "cpu_model": "AMD EPYC 7R32", + "gpu_count": 1, + "gpu_model": "1 x NVIDIA A10G", + "longitude": -77.4903, + "latitude": 39.0469, + "ram_total_size": 186.7047882080078, + "tracking_mode": "process", + "on_cloud": "N", + "pue": 1.0 +} \ No newline at end of file diff --git a/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-03-36/.hydra/config.yaml b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-03-36/.hydra/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..ca9c7622983c00dc182a28ef5197b99186dbcca4 --- /dev/null +++ b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-03-36/.hydra/config.yaml @@ -0,0 +1,94 @@ +backend: + name: pytorch + version: 2.4.0 + _target_: optimum_benchmark.backends.pytorch.backend.PyTorchBackend + task: text-classification + model: lvwerra/distilbert-imdb + processor: lvwerra/distilbert-imdb + library: null + device: cuda + device_ids: '0' + seed: 42 + inter_op_num_threads: null + intra_op_num_threads: null + hub_kwargs: {} + no_weights: true + device_map: null + torch_dtype: null + amp_autocast: false + amp_dtype: null + eval_mode: true + to_bettertransformer: false + low_cpu_mem_usage: null + attn_implementation: null + cache_implementation: null + torch_compile: false + torch_compile_config: {} + quantization_scheme: null + quantization_config: {} + deepspeed_inference: false + deepspeed_inference_config: {} + peft_type: null + peft_config: {} +launcher: + name: process + _target_: optimum_benchmark.launchers.process.launcher.ProcessLauncher + device_isolation: true + device_isolation_action: warn + start_method: spawn +benchmark: + name: energy_star + _target_: optimum_benchmark.benchmarks.energy_star.benchmark.EnergyStarBenchmark + dataset_name: EnergyStarAI/text_classification + dataset_config: '' + dataset_split: train + num_samples: 1000 + input_shapes: + batch_size: 1 + text_column_name: text + truncation: true + max_length: -1 + dataset_prefix1: '' + dataset_prefix2: '' + t5_task: '' + image_column_name: image + resize: false + question_column_name: question + context_column_name: context + sentence1_column_name: sentence1 + sentence2_column_name: sentence2 + audio_column_name: audio + iterations: 10 + warmup_runs: 10 + energy: true + forward_kwargs: {} + generate_kwargs: {} + call_kwargs: {} +experiment_name: text_classification +environment: + cpu: ' AMD EPYC 7R32' + cpu_count: 48 + cpu_ram_mb: 200472.73984 + system: Linux + machine: x86_64 + platform: Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35 + processor: x86_64 + python_version: 3.9.20 + gpu: + - NVIDIA A10G + gpu_count: 1 + gpu_vram_mb: 24146608128 + optimum_benchmark_version: 0.2.0 + optimum_benchmark_commit: null + transformers_version: 4.44.0 + transformers_commit: null + accelerate_version: 0.33.0 + accelerate_commit: null + diffusers_version: 0.30.0 + diffusers_commit: null + optimum_version: null + optimum_commit: null + timm_version: null + timm_commit: null + peft_version: null + peft_commit: null diff --git a/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-03-36/.hydra/hydra.yaml b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-03-36/.hydra/hydra.yaml new file mode 100644 index 0000000000000000000000000000000000000000..6f3e594f81555ff11ad4855a370191ffa569e5fe --- /dev/null +++ b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-03-36/.hydra/hydra.yaml @@ -0,0 +1,173 @@ +hydra: + run: + dir: /data/runs/${experiment_name}/${backend.model}/${now:%Y-%m-%d-%H-%M-%S} + sweep: + dir: sweeps/${experiment_name}/${now:%Y-%m-%d-%H-%M-%S} + subdir: ${hydra.job.num} + launcher: + _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher + sweeper: + _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper + max_batch_size: null + params: null + help: + app_name: ${hydra.job.name} + header: '${hydra.help.app_name} is powered by Hydra. + + ' + footer: 'Powered by Hydra (https://hydra.cc) + + Use --hydra-help to view Hydra specific help + + ' + template: '${hydra.help.header} + + == Configuration groups == + + Compose your configuration from those groups (group=option) + + + $APP_CONFIG_GROUPS + + + == Config == + + Override anything in the config (foo.bar=value) + + + $CONFIG + + + ${hydra.help.footer} + + ' + hydra_help: + template: 'Hydra (${hydra.runtime.version}) + + See https://hydra.cc for more info. + + + == Flags == + + $FLAGS_HELP + + + == Configuration groups == + + Compose your configuration from those groups (For example, append hydra/job_logging=disabled + to command line) + + + $HYDRA_CONFIG_GROUPS + + + Use ''--cfg hydra'' to Show the Hydra config. + + ' + hydra_help: ??? + hydra_logging: + version: 1 + formatters: + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s' + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + root: + level: INFO + handlers: + - console + disable_existing_loggers: false + job_logging: + version: 1 + formatters: + simple: + format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s' + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s] + - %(message)s' + log_colors: + DEBUG: purple + INFO: green + WARNING: yellow + ERROR: red + CRITICAL: red + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + file: + class: logging.FileHandler + formatter: simple + filename: ${hydra.job.name}.log + root: + level: INFO + handlers: + - console + - file + disable_existing_loggers: false + env: {} + mode: RUN + searchpath: [] + callbacks: {} + output_subdir: .hydra + overrides: + hydra: + - hydra.run.dir=/data/runs/${experiment_name}/${backend.model}/${now:%Y-%m-%d-%H-%M-%S} + - hydra.mode=RUN + task: [] + job: + name: cli + chdir: true + override_dirname: '' + id: ??? + num: ??? + config_name: text_classification + env_set: + OVERRIDE_BENCHMARKS: '1' + env_copy: [] + config: + override_dirname: + kv_sep: '=' + item_sep: ',' + exclude_keys: [] + runtime: + version: 1.3.2 + version_base: '1.3' + cwd: / + config_sources: + - path: hydra.conf + schema: pkg + provider: hydra + - path: optimum_benchmark + schema: pkg + provider: main + - path: hydra_plugins.hydra_colorlog.conf + schema: pkg + provider: hydra-colorlog + - path: /optimum-benchmark/examples/energy_star + schema: file + provider: command-line + - path: '' + schema: structured + provider: schema + output_dir: /data/runs/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-03-36 + choices: + benchmark: energy_star + launcher: process + backend: pytorch + hydra/env: default + hydra/callbacks: null + hydra/job_logging: colorlog + hydra/hydra_logging: colorlog + hydra/hydra_help: default + hydra/help: default + hydra/sweeper: basic + hydra/launcher: basic + hydra/output: default + verbose: false diff --git a/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-03-36/.hydra/overrides.yaml b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-03-36/.hydra/overrides.yaml new file mode 100644 index 0000000000000000000000000000000000000000..fe51488c7066f6687ef680d6bfaa4f7768ef205c --- /dev/null +++ b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-03-36/.hydra/overrides.yaml @@ -0,0 +1 @@ +[] diff --git a/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-03-36/cli.log b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-03-36/cli.log new file mode 100644 index 0000000000000000000000000000000000000000..871f7642e5ef154b7e0e146003a6b1741e9641e0 --- /dev/null +++ b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-03-36/cli.log @@ -0,0 +1,28 @@ +[2024-10-01 19:03:36,282][launcher][INFO] - ََAllocating process launcher +[2024-10-01 19:03:36,282][process][INFO] - + Setting multiprocessing start method to spawn. +[2024-10-01 19:03:36,295][device-isolation][INFO] - + Launched device(s) isolation process 66 +[2024-10-01 19:03:36,296][device-isolation][INFO] - + Isolating device(s) [0] +[2024-10-01 19:03:36,302][process][INFO] - + Launched benchmark in isolated process 67. +[PROC-0][2024-10-01 19:03:38,922][datasets][INFO] - PyTorch version 2.4.0 available. +[PROC-0][2024-10-01 19:03:39,859][backend][INFO] - َAllocating pytorch backend +[PROC-0][2024-10-01 19:03:39,859][backend][INFO] - + Setting random seed to 42 +[PROC-0][2024-10-01 19:03:40,168][pytorch][INFO] - + Using AutoModel class AutoModelForSequenceClassification +[PROC-0][2024-10-01 19:03:40,168][pytorch][INFO] - + Creating backend temporary directory +[PROC-0][2024-10-01 19:03:40,168][pytorch][INFO] - + Loading model with random weights +[PROC-0][2024-10-01 19:03:40,169][pytorch][INFO] - + Creating no weights model +[PROC-0][2024-10-01 19:03:40,169][pytorch][INFO] - + Creating no weights model directory +[PROC-0][2024-10-01 19:03:40,169][pytorch][INFO] - + Creating no weights model state dict +[PROC-0][2024-10-01 19:03:40,171][pytorch][INFO] - + Saving no weights model safetensors +[PROC-0][2024-10-01 19:03:40,171][pytorch][INFO] - + Saving no weights model pretrained config +[PROC-0][2024-10-01 19:03:40,172][pytorch][INFO] - + Loading no weights AutoModel +[PROC-0][2024-10-01 19:03:40,172][pytorch][INFO] - + Loading model directly on device: cuda +[PROC-0][2024-10-01 19:03:40,425][pytorch][INFO] - + Turning on model's eval mode +[PROC-0][2024-10-01 19:03:40,431][benchmark][INFO] - Allocating energy_star benchmark +[PROC-0][2024-10-01 19:03:40,432][energy_star][INFO] - + Loading raw dataset +[PROC-0][2024-10-01 19:03:41,213][energy_star][INFO] - + Initializing Inference report +[PROC-0][2024-10-01 19:03:41,214][energy][INFO] - + Tracking GPU energy on devices [0] +[PROC-0][2024-10-01 19:03:45,415][energy_star][INFO] - + Preprocessing dataset +[PROC-0][2024-10-01 19:03:45,847][energy][INFO] - + Saving codecarbon emission data to preprocess_codecarbon.json +[PROC-0][2024-10-01 19:03:45,849][energy_star][INFO] - + Preparing backend for Inference +[PROC-0][2024-10-01 19:03:45,849][energy_star][INFO] - + Initialising dataloader +[PROC-0][2024-10-01 19:03:45,849][energy_star][INFO] - + Warming up backend for Inference diff --git a/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-03-36/experiment_config.json b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-03-36/experiment_config.json new file mode 100644 index 0000000000000000000000000000000000000000..772aa352d839018f6caa9107fd24acf54618c6aa --- /dev/null +++ b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-03-36/experiment_config.json @@ -0,0 +1,107 @@ +{ + "experiment_name": "text_classification", + "backend": { + "name": "pytorch", + "version": "2.4.0", + "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", + "task": "text-classification", + "model": "lvwerra/distilbert-imdb", + "processor": "lvwerra/distilbert-imdb", + "library": "transformers", + "device": "cuda", + "device_ids": "0", + "seed": 42, + "inter_op_num_threads": null, + "intra_op_num_threads": null, + "hub_kwargs": { + "revision": "main", + "force_download": false, + "local_files_only": false, + "trust_remote_code": true + }, + "no_weights": true, + "device_map": null, + "torch_dtype": null, + "amp_autocast": false, + "amp_dtype": null, + "eval_mode": true, + "to_bettertransformer": false, + "low_cpu_mem_usage": null, + "attn_implementation": null, + "cache_implementation": null, + "torch_compile": false, + "torch_compile_config": {}, + "quantization_scheme": null, + "quantization_config": {}, + "deepspeed_inference": false, + "deepspeed_inference_config": {}, + "peft_type": null, + "peft_config": {} + }, + "launcher": { + "name": "process", + "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", + "device_isolation": true, + "device_isolation_action": "warn", + "start_method": "spawn" + }, + "benchmark": { + "name": "energy_star", + "_target_": "optimum_benchmark.benchmarks.energy_star.benchmark.EnergyStarBenchmark", + "dataset_name": "EnergyStarAI/text_classification", + "dataset_config": "", + "dataset_split": "train", + "num_samples": 1000, + "input_shapes": { + "batch_size": 1 + }, + "text_column_name": "text", + "truncation": true, + "max_length": -1, + "dataset_prefix1": "", + "dataset_prefix2": "", + "t5_task": "", + "image_column_name": "image", + "resize": false, + "question_column_name": "question", + "context_column_name": "context", + "sentence1_column_name": "sentence1", + "sentence2_column_name": "sentence2", + "audio_column_name": "audio", + "iterations": 10, + "warmup_runs": 10, + "energy": true, + "forward_kwargs": {}, + "generate_kwargs": {}, + "call_kwargs": {} + }, + "environment": { + "cpu": " AMD EPYC 7R32", + "cpu_count": 48, + "cpu_ram_mb": 200472.73984, + "system": "Linux", + "machine": "x86_64", + "platform": "Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35", + "processor": "x86_64", + "python_version": "3.9.20", + "gpu": [ + "NVIDIA A10G" + ], + "gpu_count": 1, + "gpu_vram_mb": 24146608128, + "optimum_benchmark_version": "0.2.0", + "optimum_benchmark_commit": null, + "transformers_version": "4.44.0", + "transformers_commit": null, + "accelerate_version": "0.33.0", + "accelerate_commit": null, + "diffusers_version": "0.30.0", + "diffusers_commit": null, + "optimum_version": null, + "optimum_commit": null, + "timm_version": null, + "timm_commit": null, + "peft_version": null, + "peft_commit": null + } +} \ No newline at end of file diff --git a/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-03-36/preprocess_codecarbon.json b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-03-36/preprocess_codecarbon.json new file mode 100644 index 0000000000000000000000000000000000000000..7a770f40c026cbcc0b74b78c9fb711d943c87712 --- /dev/null +++ b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-03-36/preprocess_codecarbon.json @@ -0,0 +1,33 @@ +{ + "timestamp": "2024-10-01T19:03:45", + "project_name": "codecarbon", + "run_id": "9729c6fe-e1e1-4c51-afc4-bd96489f187f", + "duration": -1726624718.85701, + "emissions": 4.7184334872045504e-06, + "emissions_rate": 1.1118273961949702e-05, + "cpu_power": 42.5, + "gpu_power": 64.96423630092855, + "ram_power": 0.2747225761413574, + "cpu_energy": 5.0585669730531264e-06, + "gpu_energy": 7.691395044417959e-06, + "ram_energy": 3.246565606615931e-08, + "energy_consumed": 1.2782427673537246e-05, + "country_name": "United States", + "country_iso_code": "USA", + "region": "virginia", + "cloud_provider": "", + "cloud_region": "", + "os": "Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35", + "python_version": "3.9.20", + "codecarbon_version": "2.5.1", + "cpu_count": 48, + "cpu_model": "AMD EPYC 7R32", + "gpu_count": 1, + "gpu_model": "1 x NVIDIA A10G", + "longitude": -77.4903, + "latitude": 39.0469, + "ram_total_size": 186.7047882080078, + "tracking_mode": "process", + "on_cloud": "N", + "pue": 1.0 +} \ No newline at end of file diff --git a/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-10-43/.hydra/config.yaml b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-10-43/.hydra/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..ca9c7622983c00dc182a28ef5197b99186dbcca4 --- /dev/null +++ b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-10-43/.hydra/config.yaml @@ -0,0 +1,94 @@ +backend: + name: pytorch + version: 2.4.0 + _target_: optimum_benchmark.backends.pytorch.backend.PyTorchBackend + task: text-classification + model: lvwerra/distilbert-imdb + processor: lvwerra/distilbert-imdb + library: null + device: cuda + device_ids: '0' + seed: 42 + inter_op_num_threads: null + intra_op_num_threads: null + hub_kwargs: {} + no_weights: true + device_map: null + torch_dtype: null + amp_autocast: false + amp_dtype: null + eval_mode: true + to_bettertransformer: false + low_cpu_mem_usage: null + attn_implementation: null + cache_implementation: null + torch_compile: false + torch_compile_config: {} + quantization_scheme: null + quantization_config: {} + deepspeed_inference: false + deepspeed_inference_config: {} + peft_type: null + peft_config: {} +launcher: + name: process + _target_: optimum_benchmark.launchers.process.launcher.ProcessLauncher + device_isolation: true + device_isolation_action: warn + start_method: spawn +benchmark: + name: energy_star + _target_: optimum_benchmark.benchmarks.energy_star.benchmark.EnergyStarBenchmark + dataset_name: EnergyStarAI/text_classification + dataset_config: '' + dataset_split: train + num_samples: 1000 + input_shapes: + batch_size: 1 + text_column_name: text + truncation: true + max_length: -1 + dataset_prefix1: '' + dataset_prefix2: '' + t5_task: '' + image_column_name: image + resize: false + question_column_name: question + context_column_name: context + sentence1_column_name: sentence1 + sentence2_column_name: sentence2 + audio_column_name: audio + iterations: 10 + warmup_runs: 10 + energy: true + forward_kwargs: {} + generate_kwargs: {} + call_kwargs: {} +experiment_name: text_classification +environment: + cpu: ' AMD EPYC 7R32' + cpu_count: 48 + cpu_ram_mb: 200472.73984 + system: Linux + machine: x86_64 + platform: Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35 + processor: x86_64 + python_version: 3.9.20 + gpu: + - NVIDIA A10G + gpu_count: 1 + gpu_vram_mb: 24146608128 + optimum_benchmark_version: 0.2.0 + optimum_benchmark_commit: null + transformers_version: 4.44.0 + transformers_commit: null + accelerate_version: 0.33.0 + accelerate_commit: null + diffusers_version: 0.30.0 + diffusers_commit: null + optimum_version: null + optimum_commit: null + timm_version: null + timm_commit: null + peft_version: null + peft_commit: null diff --git a/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-10-43/.hydra/hydra.yaml b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-10-43/.hydra/hydra.yaml new file mode 100644 index 0000000000000000000000000000000000000000..f44eb017363a75f5e0f3146fa35b4f79d0dd015e --- /dev/null +++ b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-10-43/.hydra/hydra.yaml @@ -0,0 +1,173 @@ +hydra: + run: + dir: /data/runs/${experiment_name}/${backend.model}/${now:%Y-%m-%d-%H-%M-%S} + sweep: + dir: sweeps/${experiment_name}/${now:%Y-%m-%d-%H-%M-%S} + subdir: ${hydra.job.num} + launcher: + _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher + sweeper: + _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper + max_batch_size: null + params: null + help: + app_name: ${hydra.job.name} + header: '${hydra.help.app_name} is powered by Hydra. + + ' + footer: 'Powered by Hydra (https://hydra.cc) + + Use --hydra-help to view Hydra specific help + + ' + template: '${hydra.help.header} + + == Configuration groups == + + Compose your configuration from those groups (group=option) + + + $APP_CONFIG_GROUPS + + + == Config == + + Override anything in the config (foo.bar=value) + + + $CONFIG + + + ${hydra.help.footer} + + ' + hydra_help: + template: 'Hydra (${hydra.runtime.version}) + + See https://hydra.cc for more info. + + + == Flags == + + $FLAGS_HELP + + + == Configuration groups == + + Compose your configuration from those groups (For example, append hydra/job_logging=disabled + to command line) + + + $HYDRA_CONFIG_GROUPS + + + Use ''--cfg hydra'' to Show the Hydra config. + + ' + hydra_help: ??? + hydra_logging: + version: 1 + formatters: + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s' + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + root: + level: INFO + handlers: + - console + disable_existing_loggers: false + job_logging: + version: 1 + formatters: + simple: + format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s' + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s] + - %(message)s' + log_colors: + DEBUG: purple + INFO: green + WARNING: yellow + ERROR: red + CRITICAL: red + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + file: + class: logging.FileHandler + formatter: simple + filename: ${hydra.job.name}.log + root: + level: INFO + handlers: + - console + - file + disable_existing_loggers: false + env: {} + mode: RUN + searchpath: [] + callbacks: {} + output_subdir: .hydra + overrides: + hydra: + - hydra.run.dir=/data/runs/${experiment_name}/${backend.model}/${now:%Y-%m-%d-%H-%M-%S} + - hydra.mode=RUN + task: [] + job: + name: cli + chdir: true + override_dirname: '' + id: ??? + num: ??? + config_name: text_classification + env_set: + OVERRIDE_BENCHMARKS: '1' + env_copy: [] + config: + override_dirname: + kv_sep: '=' + item_sep: ',' + exclude_keys: [] + runtime: + version: 1.3.2 + version_base: '1.3' + cwd: / + config_sources: + - path: hydra.conf + schema: pkg + provider: hydra + - path: optimum_benchmark + schema: pkg + provider: main + - path: hydra_plugins.hydra_colorlog.conf + schema: pkg + provider: hydra-colorlog + - path: /optimum-benchmark/examples/energy_star + schema: file + provider: command-line + - path: '' + schema: structured + provider: schema + output_dir: /data/runs/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-10-43 + choices: + benchmark: energy_star + launcher: process + backend: pytorch + hydra/env: default + hydra/callbacks: null + hydra/job_logging: colorlog + hydra/hydra_logging: colorlog + hydra/hydra_help: default + hydra/help: default + hydra/sweeper: basic + hydra/launcher: basic + hydra/output: default + verbose: false diff --git a/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-10-43/.hydra/overrides.yaml b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-10-43/.hydra/overrides.yaml new file mode 100644 index 0000000000000000000000000000000000000000..fe51488c7066f6687ef680d6bfaa4f7768ef205c --- /dev/null +++ b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-10-43/.hydra/overrides.yaml @@ -0,0 +1 @@ +[] diff --git a/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-10-43/benchmark_report.json b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-10-43/benchmark_report.json new file mode 100644 index 0000000000000000000000000000000000000000..a9ab63615e337781fdeff6df7ba79b6042668bf1 --- /dev/null +++ b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-10-43/benchmark_report.json @@ -0,0 +1,107 @@ +{ + "forward": { + "memory": null, + "latency": null, + "throughput": null, + "energy": { + "unit": "kWh", + "cpu": 4.230587705198155e-05, + "ram": 3.431121791114993e-07, + "gpu": 0.00016923796872365672, + "total": 0.00021188695795474976 + }, + "efficiency": { + "unit": "samples/kWh", + "value": 4719497.649371881 + }, + "measures": [ + { + "unit": "kWh", + "cpu": 4.747376785780944e-05, + "ram": 3.849488878078334e-07, + "gpu": 0.00018165792310398388, + "total": 0.00022951663984960112 + }, + { + "unit": "kWh", + "cpu": 4.715741071437757e-05, + "ram": 3.824990357918675e-07, + "gpu": 0.00018548153727415695, + "total": 0.00023302144702432639 + }, + { + "unit": "kWh", + "cpu": 4.722432519249399e-05, + "ram": 3.830795513851654e-07, + "gpu": 0.00018511181475577843, + "total": 0.00023271921949965765 + }, + { + "unit": "kWh", + "cpu": 4.6876172406498694e-05, + "ram": 3.802514148299537e-07, + "gpu": 0.00018767126124785705, + "total": 0.00023492768506918563 + }, + { + "unit": "kWh", + "cpu": 4.690202185055468e-05, + "ram": 3.801120580377572e-07, + "gpu": 0.0001881120949338566, + "total": 0.0002353942288424491 + }, + { + "unit": "kWh", + "cpu": 4.7141369560672024e-05, + "ram": 3.822386915096002e-07, + "gpu": 0.0001907123747928452, + "total": 0.00023823598304502694 + }, + { + "unit": "kWh", + "cpu": 0.0, + "ram": 0.0, + "gpu": 0.0, + "total": 0.0 + }, + { + "unit": "kWh", + "cpu": 4.723005958823907e-05, + "ram": 3.8313577770360217e-07, + "gpu": 0.0001910398750535336, + "total": 0.00023865307041947637 + }, + { + "unit": "kWh", + "cpu": 4.6333836789482234e-05, + "ram": 3.758602951283193e-07, + "gpu": 0.0001891912624643055, + "total": 0.00023590095954891606 + }, + { + "unit": "kWh", + "cpu": 4.6719806559687824e-05, + "ram": 3.789960789208947e-07, + "gpu": 0.00019340154361024986, + "total": 0.0002405003462488584 + } + ] + }, + "preprocess": { + "memory": null, + "latency": null, + "throughput": null, + "energy": { + "unit": "kWh", + "cpu": 5.039374399186475e-06, + "ram": 3.228873847000398e-08, + "gpu": 7.653617234204546e-06, + "total": 1.2725280371861023e-05 + }, + "efficiency": { + "unit": "samples/kWh", + "value": 78583730.24230301 + }, + "measures": null + } +} \ No newline at end of file diff --git a/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-10-43/cli.log b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-10-43/cli.log new file mode 100644 index 0000000000000000000000000000000000000000..e90d2a5ac80d63b59d3dc9f324b2a3af76fc95df --- /dev/null +++ b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-10-43/cli.log @@ -0,0 +1,113 @@ +[2024-10-01 19:10:43,629][launcher][INFO] - ََAllocating process launcher +[2024-10-01 19:10:43,630][process][INFO] - + Setting multiprocessing start method to spawn. +[2024-10-01 19:10:43,640][device-isolation][INFO] - + Launched device(s) isolation process 66 +[2024-10-01 19:10:43,640][device-isolation][INFO] - + Isolating device(s) [0] +[2024-10-01 19:10:43,646][process][INFO] - + Launched benchmark in isolated process 67. +[PROC-0][2024-10-01 19:10:46,225][datasets][INFO] - PyTorch version 2.4.0 available. +[PROC-0][2024-10-01 19:10:47,152][backend][INFO] - َAllocating pytorch backend +[PROC-0][2024-10-01 19:10:47,152][backend][INFO] - + Setting random seed to 42 +[PROC-0][2024-10-01 19:10:47,446][pytorch][INFO] - + Using AutoModel class AutoModelForSequenceClassification +[PROC-0][2024-10-01 19:10:47,446][pytorch][INFO] - + Creating backend temporary directory +[PROC-0][2024-10-01 19:10:47,446][pytorch][INFO] - + Loading model with random weights +[PROC-0][2024-10-01 19:10:47,447][pytorch][INFO] - + Creating no weights model +[PROC-0][2024-10-01 19:10:47,447][pytorch][INFO] - + Creating no weights model directory +[PROC-0][2024-10-01 19:10:47,447][pytorch][INFO] - + Creating no weights model state dict +[PROC-0][2024-10-01 19:10:47,449][pytorch][INFO] - + Saving no weights model safetensors +[PROC-0][2024-10-01 19:10:47,449][pytorch][INFO] - + Saving no weights model pretrained config +[PROC-0][2024-10-01 19:10:47,450][pytorch][INFO] - + Loading no weights AutoModel +[PROC-0][2024-10-01 19:10:47,450][pytorch][INFO] - + Loading model directly on device: cuda +[PROC-0][2024-10-01 19:10:47,709][pytorch][INFO] - + Turning on model's eval mode +[PROC-0][2024-10-01 19:10:47,715][benchmark][INFO] - Allocating energy_star benchmark +[PROC-0][2024-10-01 19:10:47,716][energy_star][INFO] - + Loading raw dataset +[PROC-0][2024-10-01 19:10:48,367][energy_star][INFO] - + Initializing Inference report +[PROC-0][2024-10-01 19:10:48,368][energy][INFO] - + Tracking GPU energy on devices [0] +[PROC-0][2024-10-01 19:10:52,549][energy_star][INFO] - + Preprocessing dataset +[PROC-0][2024-10-01 19:10:52,979][energy][INFO] - + Saving codecarbon emission data to preprocess_codecarbon.json +[PROC-0][2024-10-01 19:10:52,982][energy_star][INFO] - + Preparing backend for Inference +[PROC-0][2024-10-01 19:10:52,982][energy_star][INFO] - + Initialising dataloader +[PROC-0][2024-10-01 19:10:52,982][energy_star][INFO] - + Warming up backend for Inference +[PROC-0][2024-10-01 19:10:53,580][energy_star][INFO] - + Running Inference energy tracking for 10 iterations +[PROC-0][2024-10-01 19:10:53,580][energy_star][INFO] - + Iteration 1/10 +[PROC-0][2024-10-01 19:10:57,604][energy][INFO] - + Saving codecarbon emission data to forward_codecarbon.json +[PROC-0][2024-10-01 19:10:57,607][energy_star][INFO] - + Iteration 2/10 +[PROC-0][2024-10-01 19:11:01,606][energy][INFO] - + Saving codecarbon emission data to forward_codecarbon.json +[PROC-0][2024-10-01 19:11:01,609][energy_star][INFO] - + Iteration 3/10 +[PROC-0][2024-10-01 19:11:05,612][energy][INFO] - + Saving codecarbon emission data to forward_codecarbon.json +[PROC-0][2024-10-01 19:11:05,615][energy_star][INFO] - + Iteration 4/10 +[PROC-0][2024-10-01 19:11:09,588][energy][INFO] - + Saving codecarbon emission data to forward_codecarbon.json +[PROC-0][2024-10-01 19:11:09,591][energy_star][INFO] - + Iteration 5/10 +[PROC-0][2024-10-01 19:11:13,567][energy][INFO] - + Saving codecarbon emission data to forward_codecarbon.json +[PROC-0][2024-10-01 19:11:13,569][energy_star][INFO] - + Iteration 6/10 +[PROC-0][2024-10-01 19:11:17,565][energy][INFO] - + Saving codecarbon emission data to forward_codecarbon.json +[PROC-0][2024-10-01 19:11:17,568][energy_star][INFO] - + Iteration 7/10 +[PROC-0][2024-10-01 19:11:21,531][energy][INFO] - + Saving codecarbon emission data to forward_codecarbon.json +[PROC-0][2024-10-01 19:11:21,534][energy_star][INFO] - + Iteration 8/10 +[PROC-0][2024-10-01 19:11:25,538][energy][INFO] - + Saving codecarbon emission data to forward_codecarbon.json +[PROC-0][2024-10-01 19:11:25,540][energy_star][INFO] - + Iteration 9/10 +[PROC-0][2024-10-01 19:11:29,469][energy][INFO] - + Saving codecarbon emission data to forward_codecarbon.json +[PROC-0][2024-10-01 19:11:29,472][energy_star][INFO] - + Iteration 10/10 +[PROC-0][2024-10-01 19:11:33,433][energy][INFO] - + Saving codecarbon emission data to forward_codecarbon.json +[PROC-0][2024-10-01 19:11:33,436][energy][INFO] - + forward energy consumption: +[PROC-0][2024-10-01 19:11:33,436][energy][INFO] - + CPU: 0.000042 (kWh) +[PROC-0][2024-10-01 19:11:33,436][energy][INFO] - + GPU: 0.000169 (kWh) +[PROC-0][2024-10-01 19:11:33,436][energy][INFO] - + RAM: 0.000000 (kWh) +[PROC-0][2024-10-01 19:11:33,436][energy][INFO] - + total: 0.000212 (kWh) +[PROC-0][2024-10-01 19:11:33,437][energy][INFO] - + forward_iteration_1 energy consumption: +[PROC-0][2024-10-01 19:11:33,437][energy][INFO] - + CPU: 0.000047 (kWh) +[PROC-0][2024-10-01 19:11:33,437][energy][INFO] - + GPU: 0.000182 (kWh) +[PROC-0][2024-10-01 19:11:33,437][energy][INFO] - + RAM: 0.000000 (kWh) +[PROC-0][2024-10-01 19:11:33,437][energy][INFO] - + total: 0.000230 (kWh) +[PROC-0][2024-10-01 19:11:33,437][energy][INFO] - + forward_iteration_2 energy consumption: +[PROC-0][2024-10-01 19:11:33,437][energy][INFO] - + CPU: 0.000047 (kWh) +[PROC-0][2024-10-01 19:11:33,437][energy][INFO] - + GPU: 0.000185 (kWh) +[PROC-0][2024-10-01 19:11:33,437][energy][INFO] - + RAM: 0.000000 (kWh) +[PROC-0][2024-10-01 19:11:33,437][energy][INFO] - + total: 0.000233 (kWh) +[PROC-0][2024-10-01 19:11:33,437][energy][INFO] - + forward_iteration_3 energy consumption: +[PROC-0][2024-10-01 19:11:33,437][energy][INFO] - + CPU: 0.000047 (kWh) +[PROC-0][2024-10-01 19:11:33,438][energy][INFO] - + GPU: 0.000185 (kWh) +[PROC-0][2024-10-01 19:11:33,438][energy][INFO] - + RAM: 0.000000 (kWh) +[PROC-0][2024-10-01 19:11:33,438][energy][INFO] - + total: 0.000233 (kWh) +[PROC-0][2024-10-01 19:11:33,438][energy][INFO] - + forward_iteration_4 energy consumption: +[PROC-0][2024-10-01 19:11:33,438][energy][INFO] - + CPU: 0.000047 (kWh) +[PROC-0][2024-10-01 19:11:33,438][energy][INFO] - + GPU: 0.000188 (kWh) +[PROC-0][2024-10-01 19:11:33,438][energy][INFO] - + RAM: 0.000000 (kWh) +[PROC-0][2024-10-01 19:11:33,438][energy][INFO] - + total: 0.000235 (kWh) +[PROC-0][2024-10-01 19:11:33,438][energy][INFO] - + forward_iteration_5 energy consumption: +[PROC-0][2024-10-01 19:11:33,438][energy][INFO] - + CPU: 0.000047 (kWh) +[PROC-0][2024-10-01 19:11:33,438][energy][INFO] - + GPU: 0.000188 (kWh) +[PROC-0][2024-10-01 19:11:33,438][energy][INFO] - + RAM: 0.000000 (kWh) +[PROC-0][2024-10-01 19:11:33,438][energy][INFO] - + total: 0.000235 (kWh) +[PROC-0][2024-10-01 19:11:33,439][energy][INFO] - + forward_iteration_6 energy consumption: +[PROC-0][2024-10-01 19:11:33,439][energy][INFO] - + CPU: 0.000047 (kWh) +[PROC-0][2024-10-01 19:11:33,439][energy][INFO] - + GPU: 0.000191 (kWh) +[PROC-0][2024-10-01 19:11:33,439][energy][INFO] - + RAM: 0.000000 (kWh) +[PROC-0][2024-10-01 19:11:33,439][energy][INFO] - + total: 0.000238 (kWh) +[PROC-0][2024-10-01 19:11:33,439][energy][INFO] - + forward_iteration_7 energy consumption: +[PROC-0][2024-10-01 19:11:33,439][energy][INFO] - + CPU: 0.000000 (kWh) +[PROC-0][2024-10-01 19:11:33,439][energy][INFO] - + GPU: 0.000000 (kWh) +[PROC-0][2024-10-01 19:11:33,439][energy][INFO] - + RAM: 0.000000 (kWh) +[PROC-0][2024-10-01 19:11:33,439][energy][INFO] - + total: 0.000000 (kWh) +[PROC-0][2024-10-01 19:11:33,439][energy][INFO] - + forward_iteration_8 energy consumption: +[PROC-0][2024-10-01 19:11:33,439][energy][INFO] - + CPU: 0.000047 (kWh) +[PROC-0][2024-10-01 19:11:33,440][energy][INFO] - + GPU: 0.000191 (kWh) +[PROC-0][2024-10-01 19:11:33,440][energy][INFO] - + RAM: 0.000000 (kWh) +[PROC-0][2024-10-01 19:11:33,440][energy][INFO] - + total: 0.000239 (kWh) +[PROC-0][2024-10-01 19:11:33,440][energy][INFO] - + forward_iteration_9 energy consumption: +[PROC-0][2024-10-01 19:11:33,440][energy][INFO] - + CPU: 0.000046 (kWh) +[PROC-0][2024-10-01 19:11:33,440][energy][INFO] - + GPU: 0.000189 (kWh) +[PROC-0][2024-10-01 19:11:33,440][energy][INFO] - + RAM: 0.000000 (kWh) +[PROC-0][2024-10-01 19:11:33,440][energy][INFO] - + total: 0.000236 (kWh) +[PROC-0][2024-10-01 19:11:33,440][energy][INFO] - + forward_iteration_10 energy consumption: +[PROC-0][2024-10-01 19:11:33,440][energy][INFO] - + CPU: 0.000047 (kWh) +[PROC-0][2024-10-01 19:11:33,440][energy][INFO] - + GPU: 0.000193 (kWh) +[PROC-0][2024-10-01 19:11:33,440][energy][INFO] - + RAM: 0.000000 (kWh) +[PROC-0][2024-10-01 19:11:33,440][energy][INFO] - + total: 0.000241 (kWh) +[PROC-0][2024-10-01 19:11:33,441][energy][INFO] - + preprocess energy consumption: +[PROC-0][2024-10-01 19:11:33,441][energy][INFO] - + CPU: 0.000005 (kWh) +[PROC-0][2024-10-01 19:11:33,441][energy][INFO] - + GPU: 0.000008 (kWh) +[PROC-0][2024-10-01 19:11:33,441][energy][INFO] - + RAM: 0.000000 (kWh) +[PROC-0][2024-10-01 19:11:33,441][energy][INFO] - + total: 0.000013 (kWh) +[PROC-0][2024-10-01 19:11:33,441][energy][INFO] - + forward energy efficiency: 4719497.649372 (samples/kWh) +[PROC-0][2024-10-01 19:11:33,441][energy][INFO] - + preprocess energy efficiency: 78583730.242303 (samples/kWh) +[2024-10-01 19:11:34,040][device-isolation][INFO] - + Closing device(s) isolation process... +[2024-10-01 19:11:34,093][datasets][INFO] - PyTorch version 2.4.0 available. diff --git a/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-10-43/experiment_config.json b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-10-43/experiment_config.json new file mode 100644 index 0000000000000000000000000000000000000000..772aa352d839018f6caa9107fd24acf54618c6aa --- /dev/null +++ b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-10-43/experiment_config.json @@ -0,0 +1,107 @@ +{ + "experiment_name": "text_classification", + "backend": { + "name": "pytorch", + "version": "2.4.0", + "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", + "task": "text-classification", + "model": "lvwerra/distilbert-imdb", + "processor": "lvwerra/distilbert-imdb", + "library": "transformers", + "device": "cuda", + "device_ids": "0", + "seed": 42, + "inter_op_num_threads": null, + "intra_op_num_threads": null, + "hub_kwargs": { + "revision": "main", + "force_download": false, + "local_files_only": false, + "trust_remote_code": true + }, + "no_weights": true, + "device_map": null, + "torch_dtype": null, + "amp_autocast": false, + "amp_dtype": null, + "eval_mode": true, + "to_bettertransformer": false, + "low_cpu_mem_usage": null, + "attn_implementation": null, + "cache_implementation": null, + "torch_compile": false, + "torch_compile_config": {}, + "quantization_scheme": null, + "quantization_config": {}, + "deepspeed_inference": false, + "deepspeed_inference_config": {}, + "peft_type": null, + "peft_config": {} + }, + "launcher": { + "name": "process", + "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", + "device_isolation": true, + "device_isolation_action": "warn", + "start_method": "spawn" + }, + "benchmark": { + "name": "energy_star", + "_target_": "optimum_benchmark.benchmarks.energy_star.benchmark.EnergyStarBenchmark", + "dataset_name": "EnergyStarAI/text_classification", + "dataset_config": "", + "dataset_split": "train", + "num_samples": 1000, + "input_shapes": { + "batch_size": 1 + }, + "text_column_name": "text", + "truncation": true, + "max_length": -1, + "dataset_prefix1": "", + "dataset_prefix2": "", + "t5_task": "", + "image_column_name": "image", + "resize": false, + "question_column_name": "question", + "context_column_name": "context", + "sentence1_column_name": "sentence1", + "sentence2_column_name": "sentence2", + "audio_column_name": "audio", + "iterations": 10, + "warmup_runs": 10, + "energy": true, + "forward_kwargs": {}, + "generate_kwargs": {}, + "call_kwargs": {} + }, + "environment": { + "cpu": " AMD EPYC 7R32", + "cpu_count": 48, + "cpu_ram_mb": 200472.73984, + "system": "Linux", + "machine": "x86_64", + "platform": "Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35", + "processor": "x86_64", + "python_version": "3.9.20", + "gpu": [ + "NVIDIA A10G" + ], + "gpu_count": 1, + "gpu_vram_mb": 24146608128, + "optimum_benchmark_version": "0.2.0", + "optimum_benchmark_commit": null, + "transformers_version": "4.44.0", + "transformers_commit": null, + "accelerate_version": "0.33.0", + "accelerate_commit": null, + "diffusers_version": "0.30.0", + "diffusers_commit": null, + "optimum_version": null, + "optimum_commit": null, + "timm_version": null, + "timm_commit": null, + "peft_version": null, + "peft_commit": null + } +} \ No newline at end of file diff --git a/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-10-43/forward_codecarbon.json b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-10-43/forward_codecarbon.json new file mode 100644 index 0000000000000000000000000000000000000000..a6b87eed8d5d34d59d8b3553aaeb0b394250d07b --- /dev/null +++ b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-10-43/forward_codecarbon.json @@ -0,0 +1,33 @@ +{ + "timestamp": "2024-10-01T19:11:33", + "project_name": "codecarbon", + "run_id": "c8fcecc8-7a1e-4547-8a77-8af5c2b8f80e", + "duration": -1727489511.32649, + "emissions": 8.877694569508002e-05, + "emissions_rate": 2.246503756843967e-05, + "cpu_power": 42.5, + "gpu_power": 176.03588127710825, + "ram_power": 0.34502363204956055, + "cpu_energy": 4.6719806559687824e-05, + "gpu_energy": 0.00019340154361024986, + "ram_energy": 3.789960789208947e-07, + "energy_consumed": 0.0002405003462488584, + "country_name": "United States", + "country_iso_code": "USA", + "region": "virginia", + "cloud_provider": "", + "cloud_region": "", + "os": "Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35", + "python_version": "3.9.20", + "codecarbon_version": "2.5.1", + "cpu_count": 48, + "cpu_model": "AMD EPYC 7R32", + "gpu_count": 1, + "gpu_model": "1 x NVIDIA A10G", + "longitude": -77.4903, + "latitude": 39.0469, + "ram_total_size": 186.7047882080078, + "tracking_mode": "process", + "on_cloud": "N", + "pue": 1.0 +} \ No newline at end of file diff --git a/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-10-43/preprocess_codecarbon.json b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-10-43/preprocess_codecarbon.json new file mode 100644 index 0000000000000000000000000000000000000000..de64eb2cc9681a59b48c845ec34999ac3e696831 --- /dev/null +++ b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-10-43/preprocess_codecarbon.json @@ -0,0 +1,33 @@ +{ + "timestamp": "2024-10-01T19:10:52", + "project_name": "codecarbon", + "run_id": "c8fcecc8-7a1e-4547-8a77-8af5c2b8f80e", + "duration": -1727489514.8535163, + "emissions": 4.697338453552163e-06, + "emissions_rate": 1.105667657056027e-05, + "cpu_power": 42.5, + "gpu_power": 64.89345926326264, + "ram_power": 0.27426910400390625, + "cpu_energy": 5.039374399186475e-06, + "gpu_energy": 7.653617234204546e-06, + "ram_energy": 3.228873847000398e-08, + "energy_consumed": 1.2725280371861023e-05, + "country_name": "United States", + "country_iso_code": "USA", + "region": "virginia", + "cloud_provider": "", + "cloud_region": "", + "os": "Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35", + "python_version": "3.9.20", + "codecarbon_version": "2.5.1", + "cpu_count": 48, + "cpu_model": "AMD EPYC 7R32", + "gpu_count": 1, + "gpu_model": "1 x NVIDIA A10G", + "longitude": -77.4903, + "latitude": 39.0469, + "ram_total_size": 186.7047882080078, + "tracking_mode": "process", + "on_cloud": "N", + "pue": 1.0 +} \ No newline at end of file diff --git a/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-11-38/.hydra/config.yaml b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-11-38/.hydra/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..ca9c7622983c00dc182a28ef5197b99186dbcca4 --- /dev/null +++ b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-11-38/.hydra/config.yaml @@ -0,0 +1,94 @@ +backend: + name: pytorch + version: 2.4.0 + _target_: optimum_benchmark.backends.pytorch.backend.PyTorchBackend + task: text-classification + model: lvwerra/distilbert-imdb + processor: lvwerra/distilbert-imdb + library: null + device: cuda + device_ids: '0' + seed: 42 + inter_op_num_threads: null + intra_op_num_threads: null + hub_kwargs: {} + no_weights: true + device_map: null + torch_dtype: null + amp_autocast: false + amp_dtype: null + eval_mode: true + to_bettertransformer: false + low_cpu_mem_usage: null + attn_implementation: null + cache_implementation: null + torch_compile: false + torch_compile_config: {} + quantization_scheme: null + quantization_config: {} + deepspeed_inference: false + deepspeed_inference_config: {} + peft_type: null + peft_config: {} +launcher: + name: process + _target_: optimum_benchmark.launchers.process.launcher.ProcessLauncher + device_isolation: true + device_isolation_action: warn + start_method: spawn +benchmark: + name: energy_star + _target_: optimum_benchmark.benchmarks.energy_star.benchmark.EnergyStarBenchmark + dataset_name: EnergyStarAI/text_classification + dataset_config: '' + dataset_split: train + num_samples: 1000 + input_shapes: + batch_size: 1 + text_column_name: text + truncation: true + max_length: -1 + dataset_prefix1: '' + dataset_prefix2: '' + t5_task: '' + image_column_name: image + resize: false + question_column_name: question + context_column_name: context + sentence1_column_name: sentence1 + sentence2_column_name: sentence2 + audio_column_name: audio + iterations: 10 + warmup_runs: 10 + energy: true + forward_kwargs: {} + generate_kwargs: {} + call_kwargs: {} +experiment_name: text_classification +environment: + cpu: ' AMD EPYC 7R32' + cpu_count: 48 + cpu_ram_mb: 200472.73984 + system: Linux + machine: x86_64 + platform: Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35 + processor: x86_64 + python_version: 3.9.20 + gpu: + - NVIDIA A10G + gpu_count: 1 + gpu_vram_mb: 24146608128 + optimum_benchmark_version: 0.2.0 + optimum_benchmark_commit: null + transformers_version: 4.44.0 + transformers_commit: null + accelerate_version: 0.33.0 + accelerate_commit: null + diffusers_version: 0.30.0 + diffusers_commit: null + optimum_version: null + optimum_commit: null + timm_version: null + timm_commit: null + peft_version: null + peft_commit: null diff --git a/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-11-38/.hydra/hydra.yaml b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-11-38/.hydra/hydra.yaml new file mode 100644 index 0000000000000000000000000000000000000000..6642aa1eb9ad114eeccb8862d660ca0da8f5931d --- /dev/null +++ b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-11-38/.hydra/hydra.yaml @@ -0,0 +1,173 @@ +hydra: + run: + dir: /data/runs/${experiment_name}/${backend.model}/${now:%Y-%m-%d-%H-%M-%S} + sweep: + dir: sweeps/${experiment_name}/${now:%Y-%m-%d-%H-%M-%S} + subdir: ${hydra.job.num} + launcher: + _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher + sweeper: + _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper + max_batch_size: null + params: null + help: + app_name: ${hydra.job.name} + header: '${hydra.help.app_name} is powered by Hydra. + + ' + footer: 'Powered by Hydra (https://hydra.cc) + + Use --hydra-help to view Hydra specific help + + ' + template: '${hydra.help.header} + + == Configuration groups == + + Compose your configuration from those groups (group=option) + + + $APP_CONFIG_GROUPS + + + == Config == + + Override anything in the config (foo.bar=value) + + + $CONFIG + + + ${hydra.help.footer} + + ' + hydra_help: + template: 'Hydra (${hydra.runtime.version}) + + See https://hydra.cc for more info. + + + == Flags == + + $FLAGS_HELP + + + == Configuration groups == + + Compose your configuration from those groups (For example, append hydra/job_logging=disabled + to command line) + + + $HYDRA_CONFIG_GROUPS + + + Use ''--cfg hydra'' to Show the Hydra config. + + ' + hydra_help: ??? + hydra_logging: + version: 1 + formatters: + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s' + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + root: + level: INFO + handlers: + - console + disable_existing_loggers: false + job_logging: + version: 1 + formatters: + simple: + format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s' + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s] + - %(message)s' + log_colors: + DEBUG: purple + INFO: green + WARNING: yellow + ERROR: red + CRITICAL: red + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + file: + class: logging.FileHandler + formatter: simple + filename: ${hydra.job.name}.log + root: + level: INFO + handlers: + - console + - file + disable_existing_loggers: false + env: {} + mode: RUN + searchpath: [] + callbacks: {} + output_subdir: .hydra + overrides: + hydra: + - hydra.run.dir=/data/runs/${experiment_name}/${backend.model}/${now:%Y-%m-%d-%H-%M-%S} + - hydra.mode=RUN + task: [] + job: + name: cli + chdir: true + override_dirname: '' + id: ??? + num: ??? + config_name: text_classification + env_set: + OVERRIDE_BENCHMARKS: '1' + env_copy: [] + config: + override_dirname: + kv_sep: '=' + item_sep: ',' + exclude_keys: [] + runtime: + version: 1.3.2 + version_base: '1.3' + cwd: / + config_sources: + - path: hydra.conf + schema: pkg + provider: hydra + - path: optimum_benchmark + schema: pkg + provider: main + - path: hydra_plugins.hydra_colorlog.conf + schema: pkg + provider: hydra-colorlog + - path: /optimum-benchmark/examples/energy_star + schema: file + provider: command-line + - path: '' + schema: structured + provider: schema + output_dir: /data/runs/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-11-38 + choices: + benchmark: energy_star + launcher: process + backend: pytorch + hydra/env: default + hydra/callbacks: null + hydra/job_logging: colorlog + hydra/hydra_logging: colorlog + hydra/hydra_help: default + hydra/help: default + hydra/sweeper: basic + hydra/launcher: basic + hydra/output: default + verbose: false diff --git a/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-11-38/.hydra/overrides.yaml b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-11-38/.hydra/overrides.yaml new file mode 100644 index 0000000000000000000000000000000000000000..fe51488c7066f6687ef680d6bfaa4f7768ef205c --- /dev/null +++ b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-11-38/.hydra/overrides.yaml @@ -0,0 +1 @@ +[] diff --git a/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-11-38/cli.log b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-11-38/cli.log new file mode 100644 index 0000000000000000000000000000000000000000..9197931b33a27bdf218c2f8187420419f2d660d3 --- /dev/null +++ b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-11-38/cli.log @@ -0,0 +1,30 @@ +[2024-10-01 19:11:38,705][launcher][INFO] - ََAllocating process launcher +[2024-10-01 19:11:38,706][process][INFO] - + Setting multiprocessing start method to spawn. +[2024-10-01 19:11:38,718][device-isolation][INFO] - + Launched device(s) isolation process 66 +[2024-10-01 19:11:38,718][device-isolation][INFO] - + Isolating device(s) [0] +[2024-10-01 19:11:38,725][process][INFO] - + Launched benchmark in isolated process 67. +[PROC-0][2024-10-01 19:11:41,368][datasets][INFO] - PyTorch version 2.4.0 available. +[PROC-0][2024-10-01 19:11:42,337][backend][INFO] - َAllocating pytorch backend +[PROC-0][2024-10-01 19:11:42,338][backend][INFO] - + Setting random seed to 42 +[PROC-0][2024-10-01 19:11:42,663][pytorch][INFO] - + Using AutoModel class AutoModelForSequenceClassification +[PROC-0][2024-10-01 19:11:42,663][pytorch][INFO] - + Creating backend temporary directory +[PROC-0][2024-10-01 19:11:42,663][pytorch][INFO] - + Loading model with random weights +[PROC-0][2024-10-01 19:11:42,663][pytorch][INFO] - + Creating no weights model +[PROC-0][2024-10-01 19:11:42,663][pytorch][INFO] - + Creating no weights model directory +[PROC-0][2024-10-01 19:11:42,663][pytorch][INFO] - + Creating no weights model state dict +[PROC-0][2024-10-01 19:11:42,665][pytorch][INFO] - + Saving no weights model safetensors +[PROC-0][2024-10-01 19:11:42,666][pytorch][INFO] - + Saving no weights model pretrained config +[PROC-0][2024-10-01 19:11:42,667][pytorch][INFO] - + Loading no weights AutoModel +[PROC-0][2024-10-01 19:11:42,667][pytorch][INFO] - + Loading model directly on device: cuda +[PROC-0][2024-10-01 19:11:42,919][pytorch][INFO] - + Turning on model's eval mode +[PROC-0][2024-10-01 19:11:42,925][benchmark][INFO] - Allocating energy_star benchmark +[PROC-0][2024-10-01 19:11:42,925][energy_star][INFO] - + Loading raw dataset +[PROC-0][2024-10-01 19:11:43,473][energy_star][INFO] - + Initializing Inference report +[PROC-0][2024-10-01 19:11:43,473][energy][INFO] - + Tracking GPU energy on devices [0] +[PROC-0][2024-10-01 19:11:47,649][energy_star][INFO] - + Preprocessing dataset +[PROC-0][2024-10-01 19:11:48,079][energy][INFO] - + Saving codecarbon emission data to preprocess_codecarbon.json +[PROC-0][2024-10-01 19:11:48,082][energy_star][INFO] - + Preparing backend for Inference +[PROC-0][2024-10-01 19:11:48,082][energy_star][INFO] - + Initialising dataloader +[PROC-0][2024-10-01 19:11:48,082][energy_star][INFO] - + Warming up backend for Inference +[PROC-0][2024-10-01 19:11:48,570][energy_star][INFO] - + Running Inference energy tracking for 10 iterations +[PROC-0][2024-10-01 19:11:48,570][energy_star][INFO] - + Iteration 1/10 diff --git a/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-11-38/experiment_config.json b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-11-38/experiment_config.json new file mode 100644 index 0000000000000000000000000000000000000000..772aa352d839018f6caa9107fd24acf54618c6aa --- /dev/null +++ b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-11-38/experiment_config.json @@ -0,0 +1,107 @@ +{ + "experiment_name": "text_classification", + "backend": { + "name": "pytorch", + "version": "2.4.0", + "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", + "task": "text-classification", + "model": "lvwerra/distilbert-imdb", + "processor": "lvwerra/distilbert-imdb", + "library": "transformers", + "device": "cuda", + "device_ids": "0", + "seed": 42, + "inter_op_num_threads": null, + "intra_op_num_threads": null, + "hub_kwargs": { + "revision": "main", + "force_download": false, + "local_files_only": false, + "trust_remote_code": true + }, + "no_weights": true, + "device_map": null, + "torch_dtype": null, + "amp_autocast": false, + "amp_dtype": null, + "eval_mode": true, + "to_bettertransformer": false, + "low_cpu_mem_usage": null, + "attn_implementation": null, + "cache_implementation": null, + "torch_compile": false, + "torch_compile_config": {}, + "quantization_scheme": null, + "quantization_config": {}, + "deepspeed_inference": false, + "deepspeed_inference_config": {}, + "peft_type": null, + "peft_config": {} + }, + "launcher": { + "name": "process", + "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", + "device_isolation": true, + "device_isolation_action": "warn", + "start_method": "spawn" + }, + "benchmark": { + "name": "energy_star", + "_target_": "optimum_benchmark.benchmarks.energy_star.benchmark.EnergyStarBenchmark", + "dataset_name": "EnergyStarAI/text_classification", + "dataset_config": "", + "dataset_split": "train", + "num_samples": 1000, + "input_shapes": { + "batch_size": 1 + }, + "text_column_name": "text", + "truncation": true, + "max_length": -1, + "dataset_prefix1": "", + "dataset_prefix2": "", + "t5_task": "", + "image_column_name": "image", + "resize": false, + "question_column_name": "question", + "context_column_name": "context", + "sentence1_column_name": "sentence1", + "sentence2_column_name": "sentence2", + "audio_column_name": "audio", + "iterations": 10, + "warmup_runs": 10, + "energy": true, + "forward_kwargs": {}, + "generate_kwargs": {}, + "call_kwargs": {} + }, + "environment": { + "cpu": " AMD EPYC 7R32", + "cpu_count": 48, + "cpu_ram_mb": 200472.73984, + "system": "Linux", + "machine": "x86_64", + "platform": "Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35", + "processor": "x86_64", + "python_version": "3.9.20", + "gpu": [ + "NVIDIA A10G" + ], + "gpu_count": 1, + "gpu_vram_mb": 24146608128, + "optimum_benchmark_version": "0.2.0", + "optimum_benchmark_commit": null, + "transformers_version": "4.44.0", + "transformers_commit": null, + "accelerate_version": "0.33.0", + "accelerate_commit": null, + "diffusers_version": "0.30.0", + "diffusers_commit": null, + "optimum_version": null, + "optimum_commit": null, + "timm_version": null, + "timm_commit": null, + "peft_version": null, + "peft_commit": null + } +} \ No newline at end of file diff --git a/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-11-38/preprocess_codecarbon.json b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-11-38/preprocess_codecarbon.json new file mode 100644 index 0000000000000000000000000000000000000000..a12528715ef153b59d012abb5dcb55650c94ccbe --- /dev/null +++ b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-11-38/preprocess_codecarbon.json @@ -0,0 +1,33 @@ +{ + "timestamp": "2024-10-01T19:11:48", + "project_name": "codecarbon", + "run_id": "5a0a9fa7-9c60-4e13-86a2-6ceb176f9706", + "duration": -1727489514.8532324, + "emissions": 4.820076293574065e-06, + "emissions_rate": 1.1339286230938756e-05, + "cpu_power": 42.5, + "gpu_power": 67.66018489643034, + "ram_power": 0.2745237350463867, + "cpu_energy": 5.041561779823194e-06, + "gpu_energy": 7.983895276453268e-06, + "ram_energy": 3.232510185338502e-08, + "energy_consumed": 1.3057782158129846e-05, + "country_name": "United States", + "country_iso_code": "USA", + "region": "virginia", + "cloud_provider": "", + "cloud_region": "", + "os": "Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35", + "python_version": "3.9.20", + "codecarbon_version": "2.5.1", + "cpu_count": 48, + "cpu_model": "AMD EPYC 7R32", + "gpu_count": 1, + "gpu_model": "1 x NVIDIA A10G", + "longitude": -77.4903, + "latitude": 39.0469, + "ram_total_size": 186.7047882080078, + "tracking_mode": "process", + "on_cloud": "N", + "pue": 1.0 +} \ No newline at end of file diff --git a/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-17-35/.hydra/config.yaml b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-17-35/.hydra/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..ca9c7622983c00dc182a28ef5197b99186dbcca4 --- /dev/null +++ b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-17-35/.hydra/config.yaml @@ -0,0 +1,94 @@ +backend: + name: pytorch + version: 2.4.0 + _target_: optimum_benchmark.backends.pytorch.backend.PyTorchBackend + task: text-classification + model: lvwerra/distilbert-imdb + processor: lvwerra/distilbert-imdb + library: null + device: cuda + device_ids: '0' + seed: 42 + inter_op_num_threads: null + intra_op_num_threads: null + hub_kwargs: {} + no_weights: true + device_map: null + torch_dtype: null + amp_autocast: false + amp_dtype: null + eval_mode: true + to_bettertransformer: false + low_cpu_mem_usage: null + attn_implementation: null + cache_implementation: null + torch_compile: false + torch_compile_config: {} + quantization_scheme: null + quantization_config: {} + deepspeed_inference: false + deepspeed_inference_config: {} + peft_type: null + peft_config: {} +launcher: + name: process + _target_: optimum_benchmark.launchers.process.launcher.ProcessLauncher + device_isolation: true + device_isolation_action: warn + start_method: spawn +benchmark: + name: energy_star + _target_: optimum_benchmark.benchmarks.energy_star.benchmark.EnergyStarBenchmark + dataset_name: EnergyStarAI/text_classification + dataset_config: '' + dataset_split: train + num_samples: 1000 + input_shapes: + batch_size: 1 + text_column_name: text + truncation: true + max_length: -1 + dataset_prefix1: '' + dataset_prefix2: '' + t5_task: '' + image_column_name: image + resize: false + question_column_name: question + context_column_name: context + sentence1_column_name: sentence1 + sentence2_column_name: sentence2 + audio_column_name: audio + iterations: 10 + warmup_runs: 10 + energy: true + forward_kwargs: {} + generate_kwargs: {} + call_kwargs: {} +experiment_name: text_classification +environment: + cpu: ' AMD EPYC 7R32' + cpu_count: 48 + cpu_ram_mb: 200472.73984 + system: Linux + machine: x86_64 + platform: Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35 + processor: x86_64 + python_version: 3.9.20 + gpu: + - NVIDIA A10G + gpu_count: 1 + gpu_vram_mb: 24146608128 + optimum_benchmark_version: 0.2.0 + optimum_benchmark_commit: null + transformers_version: 4.44.0 + transformers_commit: null + accelerate_version: 0.33.0 + accelerate_commit: null + diffusers_version: 0.30.0 + diffusers_commit: null + optimum_version: null + optimum_commit: null + timm_version: null + timm_commit: null + peft_version: null + peft_commit: null diff --git a/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-17-35/.hydra/hydra.yaml b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-17-35/.hydra/hydra.yaml new file mode 100644 index 0000000000000000000000000000000000000000..789e0087c89e391e4642ea821cd343007e216917 --- /dev/null +++ b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-17-35/.hydra/hydra.yaml @@ -0,0 +1,173 @@ +hydra: + run: + dir: /data/runs/${experiment_name}/${backend.model}/${now:%Y-%m-%d-%H-%M-%S} + sweep: + dir: sweeps/${experiment_name}/${now:%Y-%m-%d-%H-%M-%S} + subdir: ${hydra.job.num} + launcher: + _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher + sweeper: + _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper + max_batch_size: null + params: null + help: + app_name: ${hydra.job.name} + header: '${hydra.help.app_name} is powered by Hydra. + + ' + footer: 'Powered by Hydra (https://hydra.cc) + + Use --hydra-help to view Hydra specific help + + ' + template: '${hydra.help.header} + + == Configuration groups == + + Compose your configuration from those groups (group=option) + + + $APP_CONFIG_GROUPS + + + == Config == + + Override anything in the config (foo.bar=value) + + + $CONFIG + + + ${hydra.help.footer} + + ' + hydra_help: + template: 'Hydra (${hydra.runtime.version}) + + See https://hydra.cc for more info. + + + == Flags == + + $FLAGS_HELP + + + == Configuration groups == + + Compose your configuration from those groups (For example, append hydra/job_logging=disabled + to command line) + + + $HYDRA_CONFIG_GROUPS + + + Use ''--cfg hydra'' to Show the Hydra config. + + ' + hydra_help: ??? + hydra_logging: + version: 1 + formatters: + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s' + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + root: + level: INFO + handlers: + - console + disable_existing_loggers: false + job_logging: + version: 1 + formatters: + simple: + format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s' + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s] + - %(message)s' + log_colors: + DEBUG: purple + INFO: green + WARNING: yellow + ERROR: red + CRITICAL: red + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + file: + class: logging.FileHandler + formatter: simple + filename: ${hydra.job.name}.log + root: + level: INFO + handlers: + - console + - file + disable_existing_loggers: false + env: {} + mode: RUN + searchpath: [] + callbacks: {} + output_subdir: .hydra + overrides: + hydra: + - hydra.run.dir=/data/runs/${experiment_name}/${backend.model}/${now:%Y-%m-%d-%H-%M-%S} + - hydra.mode=RUN + task: [] + job: + name: cli + chdir: true + override_dirname: '' + id: ??? + num: ??? + config_name: text_classification + env_set: + OVERRIDE_BENCHMARKS: '1' + env_copy: [] + config: + override_dirname: + kv_sep: '=' + item_sep: ',' + exclude_keys: [] + runtime: + version: 1.3.2 + version_base: '1.3' + cwd: / + config_sources: + - path: hydra.conf + schema: pkg + provider: hydra + - path: optimum_benchmark + schema: pkg + provider: main + - path: hydra_plugins.hydra_colorlog.conf + schema: pkg + provider: hydra-colorlog + - path: /optimum-benchmark/examples/energy_star + schema: file + provider: command-line + - path: '' + schema: structured + provider: schema + output_dir: /data/runs/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-17-35 + choices: + benchmark: energy_star + launcher: process + backend: pytorch + hydra/env: default + hydra/callbacks: null + hydra/job_logging: colorlog + hydra/hydra_logging: colorlog + hydra/hydra_help: default + hydra/help: default + hydra/sweeper: basic + hydra/launcher: basic + hydra/output: default + verbose: false diff --git a/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-17-35/.hydra/overrides.yaml b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-17-35/.hydra/overrides.yaml new file mode 100644 index 0000000000000000000000000000000000000000..fe51488c7066f6687ef680d6bfaa4f7768ef205c --- /dev/null +++ b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-17-35/.hydra/overrides.yaml @@ -0,0 +1 @@ +[] diff --git a/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-17-35/benchmark_report.json b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-17-35/benchmark_report.json new file mode 100644 index 0000000000000000000000000000000000000000..2c032883fca54459914da0c285657e8c48caa66d --- /dev/null +++ b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-17-35/benchmark_report.json @@ -0,0 +1,107 @@ +{ + "forward": { + "memory": null, + "latency": null, + "throughput": null, + "energy": { + "unit": "kWh", + "cpu": 4.260452147775747e-05, + "ram": 3.4592761775505556e-07, + "gpu": 0.0001664303275886514, + "total": 0.00020938077668416393 + }, + "efficiency": { + "unit": "samples/kWh", + "value": 4775987.632849548 + }, + "measures": [ + { + "unit": "kWh", + "cpu": 4.768930666646661e-05, + "ram": 3.8696443611667573e-07, + "gpu": 0.0001821493123861373, + "total": 0.00023022558348872058 + }, + { + "unit": "kWh", + "cpu": 4.6940381795846194e-05, + "ram": 3.8121489929549824e-07, + "gpu": 0.00018246292374790585, + "total": 0.00022978452044304757 + }, + { + "unit": "kWh", + "cpu": 4.6714137484963844e-05, + "ram": 3.793954227984908e-07, + "gpu": 0.00017757097538995126, + "total": 0.00022466450829771356 + }, + { + "unit": "kWh", + "cpu": 4.6761999083739494e-05, + "ram": 3.795202562096357e-07, + "gpu": 0.0001828726462980157, + "total": 0.00023001416563796478 + }, + { + "unit": "kWh", + "cpu": 4.669321016598487e-05, + "ram": 3.792265216560366e-07, + "gpu": 0.00018768403903601794, + "total": 0.00023475647572365884 + }, + { + "unit": "kWh", + "cpu": 4.767671408163651e-05, + "ram": 3.872217933203246e-07, + "gpu": 0.00018459459212039064, + "total": 0.00023265852799534748 + }, + { + "unit": "kWh", + "cpu": 0.0, + "ram": 0.0, + "gpu": 0.0, + "total": 0.0 + }, + { + "unit": "kWh", + "cpu": 4.784057620854583e-05, + "ram": 3.8854325756228024e-07, + "gpu": 0.00019145376427420757, + "total": 0.0002396828837403157 + }, + { + "unit": "kWh", + "cpu": 4.801785357792847e-05, + "ram": 3.8969878714738733e-07, + "gpu": 0.00018962904059183217, + "total": 0.00023803659295690825 + }, + { + "unit": "kWh", + "cpu": 4.771103571246286e-05, + "ram": 3.8749080344422673e-07, + "gpu": 0.00018588598204205553, + "total": 0.00023398450855796225 + } + ] + }, + "preprocess": { + "memory": null, + "latency": null, + "throughput": null, + "energy": { + "unit": "kWh", + "cpu": 5.0026916104090765e-06, + "ram": 3.210765448103274e-08, + "gpu": 7.458894855982834e-06, + "total": 1.2493694120872942e-05 + }, + "efficiency": { + "unit": "samples/kWh", + "value": 80040377.9959141 + }, + "measures": null + } +} \ No newline at end of file diff --git a/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-17-35/cli.log b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-17-35/cli.log new file mode 100644 index 0000000000000000000000000000000000000000..dbc6a7aac5ee04e649477d03b97db9d4026aaad7 --- /dev/null +++ b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-17-35/cli.log @@ -0,0 +1,113 @@ +[2024-10-01 19:17:35,237][launcher][INFO] - ََAllocating process launcher +[2024-10-01 19:17:35,237][process][INFO] - + Setting multiprocessing start method to spawn. +[2024-10-01 19:17:35,269][device-isolation][INFO] - + Launched device(s) isolation process 66 +[2024-10-01 19:17:35,270][device-isolation][INFO] - + Isolating device(s) [0] +[2024-10-01 19:17:35,277][process][INFO] - + Launched benchmark in isolated process 67. +[PROC-0][2024-10-01 19:17:37,871][datasets][INFO] - PyTorch version 2.4.0 available. +[PROC-0][2024-10-01 19:17:38,845][backend][INFO] - َAllocating pytorch backend +[PROC-0][2024-10-01 19:17:38,846][backend][INFO] - + Setting random seed to 42 +[PROC-0][2024-10-01 19:17:39,114][pytorch][INFO] - + Using AutoModel class AutoModelForSequenceClassification +[PROC-0][2024-10-01 19:17:39,114][pytorch][INFO] - + Creating backend temporary directory +[PROC-0][2024-10-01 19:17:39,114][pytorch][INFO] - + Loading model with random weights +[PROC-0][2024-10-01 19:17:39,114][pytorch][INFO] - + Creating no weights model +[PROC-0][2024-10-01 19:17:39,114][pytorch][INFO] - + Creating no weights model directory +[PROC-0][2024-10-01 19:17:39,115][pytorch][INFO] - + Creating no weights model state dict +[PROC-0][2024-10-01 19:17:39,117][pytorch][INFO] - + Saving no weights model safetensors +[PROC-0][2024-10-01 19:17:39,117][pytorch][INFO] - + Saving no weights model pretrained config +[PROC-0][2024-10-01 19:17:39,118][pytorch][INFO] - + Loading no weights AutoModel +[PROC-0][2024-10-01 19:17:39,118][pytorch][INFO] - + Loading model directly on device: cuda +[PROC-0][2024-10-01 19:17:39,380][pytorch][INFO] - + Turning on model's eval mode +[PROC-0][2024-10-01 19:17:39,386][benchmark][INFO] - Allocating energy_star benchmark +[PROC-0][2024-10-01 19:17:39,387][energy_star][INFO] - + Loading raw dataset +[PROC-0][2024-10-01 19:17:40,056][energy_star][INFO] - + Initializing Inference report +[PROC-0][2024-10-01 19:17:40,056][energy][INFO] - + Tracking GPU energy on devices [0] +[PROC-0][2024-10-01 19:17:44,248][energy_star][INFO] - + Preprocessing dataset +[PROC-0][2024-10-01 19:17:44,676][energy][INFO] - + Saving codecarbon emission data to preprocess_codecarbon.json +[PROC-0][2024-10-01 19:17:44,679][energy_star][INFO] - + Preparing backend for Inference +[PROC-0][2024-10-01 19:17:44,679][energy_star][INFO] - + Initialising dataloader +[PROC-0][2024-10-01 19:17:44,680][energy_star][INFO] - + Warming up backend for Inference +[PROC-0][2024-10-01 19:17:45,201][energy_star][INFO] - + Running Inference energy tracking for 10 iterations +[PROC-0][2024-10-01 19:17:45,201][energy_star][INFO] - + Iteration 1/10 +[PROC-0][2024-10-01 19:17:49,243][energy][INFO] - + Saving codecarbon emission data to forward_codecarbon.json +[PROC-0][2024-10-01 19:17:49,246][energy_star][INFO] - + Iteration 2/10 +[PROC-0][2024-10-01 19:17:53,228][energy][INFO] - + Saving codecarbon emission data to forward_codecarbon.json +[PROC-0][2024-10-01 19:17:53,233][energy_star][INFO] - + Iteration 3/10 +[PROC-0][2024-10-01 19:17:57,193][energy][INFO] - + Saving codecarbon emission data to forward_codecarbon.json +[PROC-0][2024-10-01 19:17:57,196][energy_star][INFO] - + Iteration 4/10 +[PROC-0][2024-10-01 19:18:01,160][energy][INFO] - + Saving codecarbon emission data to forward_codecarbon.json +[PROC-0][2024-10-01 19:18:01,163][energy_star][INFO] - + Iteration 5/10 +[PROC-0][2024-10-01 19:18:05,124][energy][INFO] - + Saving codecarbon emission data to forward_codecarbon.json +[PROC-0][2024-10-01 19:18:05,127][energy_star][INFO] - + Iteration 6/10 +[PROC-0][2024-10-01 19:18:09,169][energy][INFO] - + Saving codecarbon emission data to forward_codecarbon.json +[PROC-0][2024-10-01 19:18:09,172][energy_star][INFO] - + Iteration 7/10 +[PROC-0][2024-10-01 19:18:13,182][energy][INFO] - + Saving codecarbon emission data to forward_codecarbon.json +[PROC-0][2024-10-01 19:18:13,185][energy_star][INFO] - + Iteration 8/10 +[PROC-0][2024-10-01 19:18:17,240][energy][INFO] - + Saving codecarbon emission data to forward_codecarbon.json +[PROC-0][2024-10-01 19:18:17,243][energy_star][INFO] - + Iteration 9/10 +[PROC-0][2024-10-01 19:18:21,315][energy][INFO] - + Saving codecarbon emission data to forward_codecarbon.json +[PROC-0][2024-10-01 19:18:21,319][energy_star][INFO] - + Iteration 10/10 +[PROC-0][2024-10-01 19:18:25,364][energy][INFO] - + Saving codecarbon emission data to forward_codecarbon.json +[PROC-0][2024-10-01 19:18:25,367][energy][INFO] - + forward energy consumption: +[PROC-0][2024-10-01 19:18:25,367][energy][INFO] - + CPU: 0.000043 (kWh) +[PROC-0][2024-10-01 19:18:25,367][energy][INFO] - + GPU: 0.000166 (kWh) +[PROC-0][2024-10-01 19:18:25,367][energy][INFO] - + RAM: 0.000000 (kWh) +[PROC-0][2024-10-01 19:18:25,368][energy][INFO] - + total: 0.000209 (kWh) +[PROC-0][2024-10-01 19:18:25,368][energy][INFO] - + forward_iteration_1 energy consumption: +[PROC-0][2024-10-01 19:18:25,368][energy][INFO] - + CPU: 0.000048 (kWh) +[PROC-0][2024-10-01 19:18:25,368][energy][INFO] - + GPU: 0.000182 (kWh) +[PROC-0][2024-10-01 19:18:25,368][energy][INFO] - + RAM: 0.000000 (kWh) +[PROC-0][2024-10-01 19:18:25,368][energy][INFO] - + total: 0.000230 (kWh) +[PROC-0][2024-10-01 19:18:25,368][energy][INFO] - + forward_iteration_2 energy consumption: +[PROC-0][2024-10-01 19:18:25,368][energy][INFO] - + CPU: 0.000047 (kWh) +[PROC-0][2024-10-01 19:18:25,368][energy][INFO] - + GPU: 0.000182 (kWh) +[PROC-0][2024-10-01 19:18:25,369][energy][INFO] - + RAM: 0.000000 (kWh) +[PROC-0][2024-10-01 19:18:25,369][energy][INFO] - + total: 0.000230 (kWh) +[PROC-0][2024-10-01 19:18:25,369][energy][INFO] - + forward_iteration_3 energy consumption: +[PROC-0][2024-10-01 19:18:25,369][energy][INFO] - + CPU: 0.000047 (kWh) +[PROC-0][2024-10-01 19:18:25,369][energy][INFO] - + GPU: 0.000178 (kWh) +[PROC-0][2024-10-01 19:18:25,369][energy][INFO] - + RAM: 0.000000 (kWh) +[PROC-0][2024-10-01 19:18:25,369][energy][INFO] - + total: 0.000225 (kWh) +[PROC-0][2024-10-01 19:18:25,369][energy][INFO] - + forward_iteration_4 energy consumption: +[PROC-0][2024-10-01 19:18:25,369][energy][INFO] - + CPU: 0.000047 (kWh) +[PROC-0][2024-10-01 19:18:25,370][energy][INFO] - + GPU: 0.000183 (kWh) +[PROC-0][2024-10-01 19:18:25,370][energy][INFO] - + RAM: 0.000000 (kWh) +[PROC-0][2024-10-01 19:18:25,370][energy][INFO] - + total: 0.000230 (kWh) +[PROC-0][2024-10-01 19:18:25,370][energy][INFO] - + forward_iteration_5 energy consumption: +[PROC-0][2024-10-01 19:18:25,370][energy][INFO] - + CPU: 0.000047 (kWh) +[PROC-0][2024-10-01 19:18:25,370][energy][INFO] - + GPU: 0.000188 (kWh) +[PROC-0][2024-10-01 19:18:25,370][energy][INFO] - + RAM: 0.000000 (kWh) +[PROC-0][2024-10-01 19:18:25,370][energy][INFO] - + total: 0.000235 (kWh) +[PROC-0][2024-10-01 19:18:25,370][energy][INFO] - + forward_iteration_6 energy consumption: +[PROC-0][2024-10-01 19:18:25,370][energy][INFO] - + CPU: 0.000048 (kWh) +[PROC-0][2024-10-01 19:18:25,371][energy][INFO] - + GPU: 0.000185 (kWh) +[PROC-0][2024-10-01 19:18:25,371][energy][INFO] - + RAM: 0.000000 (kWh) +[PROC-0][2024-10-01 19:18:25,371][energy][INFO] - + total: 0.000233 (kWh) +[PROC-0][2024-10-01 19:18:25,371][energy][INFO] - + forward_iteration_7 energy consumption: +[PROC-0][2024-10-01 19:18:25,371][energy][INFO] - + CPU: 0.000000 (kWh) +[PROC-0][2024-10-01 19:18:25,371][energy][INFO] - + GPU: 0.000000 (kWh) +[PROC-0][2024-10-01 19:18:25,371][energy][INFO] - + RAM: 0.000000 (kWh) +[PROC-0][2024-10-01 19:18:25,371][energy][INFO] - + total: 0.000000 (kWh) +[PROC-0][2024-10-01 19:18:25,371][energy][INFO] - + forward_iteration_8 energy consumption: +[PROC-0][2024-10-01 19:18:25,371][energy][INFO] - + CPU: 0.000048 (kWh) +[PROC-0][2024-10-01 19:18:25,371][energy][INFO] - + GPU: 0.000191 (kWh) +[PROC-0][2024-10-01 19:18:25,372][energy][INFO] - + RAM: 0.000000 (kWh) +[PROC-0][2024-10-01 19:18:25,372][energy][INFO] - + total: 0.000240 (kWh) +[PROC-0][2024-10-01 19:18:25,372][energy][INFO] - + forward_iteration_9 energy consumption: +[PROC-0][2024-10-01 19:18:25,372][energy][INFO] - + CPU: 0.000048 (kWh) +[PROC-0][2024-10-01 19:18:25,372][energy][INFO] - + GPU: 0.000190 (kWh) +[PROC-0][2024-10-01 19:18:25,372][energy][INFO] - + RAM: 0.000000 (kWh) +[PROC-0][2024-10-01 19:18:25,372][energy][INFO] - + total: 0.000238 (kWh) +[PROC-0][2024-10-01 19:18:25,372][energy][INFO] - + forward_iteration_10 energy consumption: +[PROC-0][2024-10-01 19:18:25,372][energy][INFO] - + CPU: 0.000048 (kWh) +[PROC-0][2024-10-01 19:18:25,372][energy][INFO] - + GPU: 0.000186 (kWh) +[PROC-0][2024-10-01 19:18:25,372][energy][INFO] - + RAM: 0.000000 (kWh) +[PROC-0][2024-10-01 19:18:25,373][energy][INFO] - + total: 0.000234 (kWh) +[PROC-0][2024-10-01 19:18:25,373][energy][INFO] - + preprocess energy consumption: +[PROC-0][2024-10-01 19:18:25,373][energy][INFO] - + CPU: 0.000005 (kWh) +[PROC-0][2024-10-01 19:18:25,373][energy][INFO] - + GPU: 0.000007 (kWh) +[PROC-0][2024-10-01 19:18:25,373][energy][INFO] - + RAM: 0.000000 (kWh) +[PROC-0][2024-10-01 19:18:25,373][energy][INFO] - + total: 0.000012 (kWh) +[PROC-0][2024-10-01 19:18:25,373][energy][INFO] - + forward energy efficiency: 4775987.632850 (samples/kWh) +[PROC-0][2024-10-01 19:18:25,373][energy][INFO] - + preprocess energy efficiency: 80040377.995914 (samples/kWh) +[2024-10-01 19:18:26,002][device-isolation][INFO] - + Closing device(s) isolation process... +[2024-10-01 19:18:26,056][datasets][INFO] - PyTorch version 2.4.0 available. diff --git a/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-17-35/experiment_config.json b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-17-35/experiment_config.json new file mode 100644 index 0000000000000000000000000000000000000000..772aa352d839018f6caa9107fd24acf54618c6aa --- /dev/null +++ b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-17-35/experiment_config.json @@ -0,0 +1,107 @@ +{ + "experiment_name": "text_classification", + "backend": { + "name": "pytorch", + "version": "2.4.0", + "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", + "task": "text-classification", + "model": "lvwerra/distilbert-imdb", + "processor": "lvwerra/distilbert-imdb", + "library": "transformers", + "device": "cuda", + "device_ids": "0", + "seed": 42, + "inter_op_num_threads": null, + "intra_op_num_threads": null, + "hub_kwargs": { + "revision": "main", + "force_download": false, + "local_files_only": false, + "trust_remote_code": true + }, + "no_weights": true, + "device_map": null, + "torch_dtype": null, + "amp_autocast": false, + "amp_dtype": null, + "eval_mode": true, + "to_bettertransformer": false, + "low_cpu_mem_usage": null, + "attn_implementation": null, + "cache_implementation": null, + "torch_compile": false, + "torch_compile_config": {}, + "quantization_scheme": null, + "quantization_config": {}, + "deepspeed_inference": false, + "deepspeed_inference_config": {}, + "peft_type": null, + "peft_config": {} + }, + "launcher": { + "name": "process", + "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", + "device_isolation": true, + "device_isolation_action": "warn", + "start_method": "spawn" + }, + "benchmark": { + "name": "energy_star", + "_target_": "optimum_benchmark.benchmarks.energy_star.benchmark.EnergyStarBenchmark", + "dataset_name": "EnergyStarAI/text_classification", + "dataset_config": "", + "dataset_split": "train", + "num_samples": 1000, + "input_shapes": { + "batch_size": 1 + }, + "text_column_name": "text", + "truncation": true, + "max_length": -1, + "dataset_prefix1": "", + "dataset_prefix2": "", + "t5_task": "", + "image_column_name": "image", + "resize": false, + "question_column_name": "question", + "context_column_name": "context", + "sentence1_column_name": "sentence1", + "sentence2_column_name": "sentence2", + "audio_column_name": "audio", + "iterations": 10, + "warmup_runs": 10, + "energy": true, + "forward_kwargs": {}, + "generate_kwargs": {}, + "call_kwargs": {} + }, + "environment": { + "cpu": " AMD EPYC 7R32", + "cpu_count": 48, + "cpu_ram_mb": 200472.73984, + "system": "Linux", + "machine": "x86_64", + "platform": "Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35", + "processor": "x86_64", + "python_version": "3.9.20", + "gpu": [ + "NVIDIA A10G" + ], + "gpu_count": 1, + "gpu_vram_mb": 24146608128, + "optimum_benchmark_version": "0.2.0", + "optimum_benchmark_commit": null, + "transformers_version": "4.44.0", + "transformers_commit": null, + "accelerate_version": "0.33.0", + "accelerate_commit": null, + "diffusers_version": "0.30.0", + "diffusers_commit": null, + "optimum_version": null, + "optimum_commit": null, + "timm_version": null, + "timm_commit": null, + "peft_version": null, + "peft_commit": null + } +} \ No newline at end of file diff --git a/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-17-35/forward_codecarbon.json b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-17-35/forward_codecarbon.json new file mode 100644 index 0000000000000000000000000000000000000000..e45315f8c5dcdea9f8795c1374631c251f6077b8 --- /dev/null +++ b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-17-35/forward_codecarbon.json @@ -0,0 +1,33 @@ +{ + "timestamp": "2024-10-01T19:18:25", + "project_name": "codecarbon", + "run_id": "a6f42b22-00c0-4a36-9d59-397582f53253", + "duration": -1727580654.2330382, + "emissions": 8.637172600261395e-05, + "emissions_rate": 2.13816591502115e-05, + "cpu_power": 42.5, + "gpu_power": 165.67962005758133, + "ram_power": 0.34543418884277344, + "cpu_energy": 4.771103571246286e-05, + "gpu_energy": 0.00018588598204205553, + "ram_energy": 3.8749080344422673e-07, + "energy_consumed": 0.00023398450855796225, + "country_name": "United States", + "country_iso_code": "USA", + "region": "virginia", + "cloud_provider": "", + "cloud_region": "", + "os": "Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35", + "python_version": "3.9.20", + "codecarbon_version": "2.5.1", + "cpu_count": 48, + "cpu_model": "AMD EPYC 7R32", + "gpu_count": 1, + "gpu_model": "1 x NVIDIA A10G", + "longitude": -77.4903, + "latitude": 39.0469, + "ram_total_size": 186.7047882080078, + "tracking_mode": "process", + "on_cloud": "N", + "pue": 1.0 +} \ No newline at end of file diff --git a/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-17-35/preprocess_codecarbon.json b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-17-35/preprocess_codecarbon.json new file mode 100644 index 0000000000000000000000000000000000000000..1bd6ba385ca87147b03681a4db8c28a978c7b86c --- /dev/null +++ b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-17-35/preprocess_codecarbon.json @@ -0,0 +1,33 @@ +{ + "timestamp": "2024-10-01T19:17:44", + "project_name": "codecarbon", + "run_id": "a6f42b22-00c0-4a36-9d59-397582f53253", + "duration": -1727580657.850911, + "emissions": 4.611852006865629e-06, + "emissions_rate": 1.093499566834929e-05, + "cpu_power": 42.5, + "gpu_power": 63.710496193228956, + "ram_power": 0.2747640609741211, + "cpu_energy": 5.0026916104090765e-06, + "gpu_energy": 7.458894855982834e-06, + "ram_energy": 3.210765448103274e-08, + "energy_consumed": 1.2493694120872942e-05, + "country_name": "United States", + "country_iso_code": "USA", + "region": "virginia", + "cloud_provider": "", + "cloud_region": "", + "os": "Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35", + "python_version": "3.9.20", + "codecarbon_version": "2.5.1", + "cpu_count": 48, + "cpu_model": "AMD EPYC 7R32", + "gpu_count": 1, + "gpu_model": "1 x NVIDIA A10G", + "longitude": -77.4903, + "latitude": 39.0469, + "ram_total_size": 186.7047882080078, + "tracking_mode": "process", + "on_cloud": "N", + "pue": 1.0 +} \ No newline at end of file diff --git a/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-18-30/.hydra/config.yaml b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-18-30/.hydra/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..ca9c7622983c00dc182a28ef5197b99186dbcca4 --- /dev/null +++ b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-18-30/.hydra/config.yaml @@ -0,0 +1,94 @@ +backend: + name: pytorch + version: 2.4.0 + _target_: optimum_benchmark.backends.pytorch.backend.PyTorchBackend + task: text-classification + model: lvwerra/distilbert-imdb + processor: lvwerra/distilbert-imdb + library: null + device: cuda + device_ids: '0' + seed: 42 + inter_op_num_threads: null + intra_op_num_threads: null + hub_kwargs: {} + no_weights: true + device_map: null + torch_dtype: null + amp_autocast: false + amp_dtype: null + eval_mode: true + to_bettertransformer: false + low_cpu_mem_usage: null + attn_implementation: null + cache_implementation: null + torch_compile: false + torch_compile_config: {} + quantization_scheme: null + quantization_config: {} + deepspeed_inference: false + deepspeed_inference_config: {} + peft_type: null + peft_config: {} +launcher: + name: process + _target_: optimum_benchmark.launchers.process.launcher.ProcessLauncher + device_isolation: true + device_isolation_action: warn + start_method: spawn +benchmark: + name: energy_star + _target_: optimum_benchmark.benchmarks.energy_star.benchmark.EnergyStarBenchmark + dataset_name: EnergyStarAI/text_classification + dataset_config: '' + dataset_split: train + num_samples: 1000 + input_shapes: + batch_size: 1 + text_column_name: text + truncation: true + max_length: -1 + dataset_prefix1: '' + dataset_prefix2: '' + t5_task: '' + image_column_name: image + resize: false + question_column_name: question + context_column_name: context + sentence1_column_name: sentence1 + sentence2_column_name: sentence2 + audio_column_name: audio + iterations: 10 + warmup_runs: 10 + energy: true + forward_kwargs: {} + generate_kwargs: {} + call_kwargs: {} +experiment_name: text_classification +environment: + cpu: ' AMD EPYC 7R32' + cpu_count: 48 + cpu_ram_mb: 200472.73984 + system: Linux + machine: x86_64 + platform: Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35 + processor: x86_64 + python_version: 3.9.20 + gpu: + - NVIDIA A10G + gpu_count: 1 + gpu_vram_mb: 24146608128 + optimum_benchmark_version: 0.2.0 + optimum_benchmark_commit: null + transformers_version: 4.44.0 + transformers_commit: null + accelerate_version: 0.33.0 + accelerate_commit: null + diffusers_version: 0.30.0 + diffusers_commit: null + optimum_version: null + optimum_commit: null + timm_version: null + timm_commit: null + peft_version: null + peft_commit: null diff --git a/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-18-30/.hydra/hydra.yaml b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-18-30/.hydra/hydra.yaml new file mode 100644 index 0000000000000000000000000000000000000000..50ae65efefa411f63a16b188cd9bf73425ce0348 --- /dev/null +++ b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-18-30/.hydra/hydra.yaml @@ -0,0 +1,173 @@ +hydra: + run: + dir: /data/runs/${experiment_name}/${backend.model}/${now:%Y-%m-%d-%H-%M-%S} + sweep: + dir: sweeps/${experiment_name}/${now:%Y-%m-%d-%H-%M-%S} + subdir: ${hydra.job.num} + launcher: + _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher + sweeper: + _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper + max_batch_size: null + params: null + help: + app_name: ${hydra.job.name} + header: '${hydra.help.app_name} is powered by Hydra. + + ' + footer: 'Powered by Hydra (https://hydra.cc) + + Use --hydra-help to view Hydra specific help + + ' + template: '${hydra.help.header} + + == Configuration groups == + + Compose your configuration from those groups (group=option) + + + $APP_CONFIG_GROUPS + + + == Config == + + Override anything in the config (foo.bar=value) + + + $CONFIG + + + ${hydra.help.footer} + + ' + hydra_help: + template: 'Hydra (${hydra.runtime.version}) + + See https://hydra.cc for more info. + + + == Flags == + + $FLAGS_HELP + + + == Configuration groups == + + Compose your configuration from those groups (For example, append hydra/job_logging=disabled + to command line) + + + $HYDRA_CONFIG_GROUPS + + + Use ''--cfg hydra'' to Show the Hydra config. + + ' + hydra_help: ??? + hydra_logging: + version: 1 + formatters: + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s' + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + root: + level: INFO + handlers: + - console + disable_existing_loggers: false + job_logging: + version: 1 + formatters: + simple: + format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s' + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s] + - %(message)s' + log_colors: + DEBUG: purple + INFO: green + WARNING: yellow + ERROR: red + CRITICAL: red + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + file: + class: logging.FileHandler + formatter: simple + filename: ${hydra.job.name}.log + root: + level: INFO + handlers: + - console + - file + disable_existing_loggers: false + env: {} + mode: RUN + searchpath: [] + callbacks: {} + output_subdir: .hydra + overrides: + hydra: + - hydra.run.dir=/data/runs/${experiment_name}/${backend.model}/${now:%Y-%m-%d-%H-%M-%S} + - hydra.mode=RUN + task: [] + job: + name: cli + chdir: true + override_dirname: '' + id: ??? + num: ??? + config_name: text_classification + env_set: + OVERRIDE_BENCHMARKS: '1' + env_copy: [] + config: + override_dirname: + kv_sep: '=' + item_sep: ',' + exclude_keys: [] + runtime: + version: 1.3.2 + version_base: '1.3' + cwd: / + config_sources: + - path: hydra.conf + schema: pkg + provider: hydra + - path: optimum_benchmark + schema: pkg + provider: main + - path: hydra_plugins.hydra_colorlog.conf + schema: pkg + provider: hydra-colorlog + - path: /optimum-benchmark/examples/energy_star + schema: file + provider: command-line + - path: '' + schema: structured + provider: schema + output_dir: /data/runs/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-18-30 + choices: + benchmark: energy_star + launcher: process + backend: pytorch + hydra/env: default + hydra/callbacks: null + hydra/job_logging: colorlog + hydra/hydra_logging: colorlog + hydra/hydra_help: default + hydra/help: default + hydra/sweeper: basic + hydra/launcher: basic + hydra/output: default + verbose: false diff --git a/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-18-30/.hydra/overrides.yaml b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-18-30/.hydra/overrides.yaml new file mode 100644 index 0000000000000000000000000000000000000000..fe51488c7066f6687ef680d6bfaa4f7768ef205c --- /dev/null +++ b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-18-30/.hydra/overrides.yaml @@ -0,0 +1 @@ +[] diff --git a/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-18-30/cli.log b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-18-30/cli.log new file mode 100644 index 0000000000000000000000000000000000000000..9100e9d1df7f2ff431d11b1a08384f390f8d35a5 --- /dev/null +++ b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-18-30/cli.log @@ -0,0 +1,24 @@ +[2024-10-01 19:18:30,814][launcher][INFO] - ََAllocating process launcher +[2024-10-01 19:18:30,815][process][INFO] - + Setting multiprocessing start method to spawn. +[2024-10-01 19:18:30,826][device-isolation][INFO] - + Launched device(s) isolation process 66 +[2024-10-01 19:18:30,826][device-isolation][INFO] - + Isolating device(s) [0] +[2024-10-01 19:18:30,833][process][INFO] - + Launched benchmark in isolated process 67. +[PROC-0][2024-10-01 19:18:33,466][datasets][INFO] - PyTorch version 2.4.0 available. +[PROC-0][2024-10-01 19:18:34,452][backend][INFO] - َAllocating pytorch backend +[PROC-0][2024-10-01 19:18:34,452][backend][INFO] - + Setting random seed to 42 +[PROC-0][2024-10-01 19:18:35,428][pytorch][INFO] - + Using AutoModel class AutoModelForSequenceClassification +[PROC-0][2024-10-01 19:18:35,428][pytorch][INFO] - + Creating backend temporary directory +[PROC-0][2024-10-01 19:18:35,429][pytorch][INFO] - + Loading model with random weights +[PROC-0][2024-10-01 19:18:35,429][pytorch][INFO] - + Creating no weights model +[PROC-0][2024-10-01 19:18:35,429][pytorch][INFO] - + Creating no weights model directory +[PROC-0][2024-10-01 19:18:35,429][pytorch][INFO] - + Creating no weights model state dict +[PROC-0][2024-10-01 19:18:35,431][pytorch][INFO] - + Saving no weights model safetensors +[PROC-0][2024-10-01 19:18:35,432][pytorch][INFO] - + Saving no weights model pretrained config +[PROC-0][2024-10-01 19:18:35,433][pytorch][INFO] - + Loading no weights AutoModel +[PROC-0][2024-10-01 19:18:35,433][pytorch][INFO] - + Loading model directly on device: cuda +[PROC-0][2024-10-01 19:18:35,717][pytorch][INFO] - + Turning on model's eval mode +[PROC-0][2024-10-01 19:18:35,723][benchmark][INFO] - Allocating energy_star benchmark +[PROC-0][2024-10-01 19:18:35,723][energy_star][INFO] - + Loading raw dataset +[PROC-0][2024-10-01 19:18:36,671][energy_star][INFO] - + Initializing Inference report +[PROC-0][2024-10-01 19:18:36,672][energy][INFO] - + Tracking GPU energy on devices [0] +[PROC-0][2024-10-01 19:18:40,851][energy_star][INFO] - + Preprocessing dataset diff --git a/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-18-30/experiment_config.json b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-18-30/experiment_config.json new file mode 100644 index 0000000000000000000000000000000000000000..772aa352d839018f6caa9107fd24acf54618c6aa --- /dev/null +++ b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-18-30/experiment_config.json @@ -0,0 +1,107 @@ +{ + "experiment_name": "text_classification", + "backend": { + "name": "pytorch", + "version": "2.4.0", + "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", + "task": "text-classification", + "model": "lvwerra/distilbert-imdb", + "processor": "lvwerra/distilbert-imdb", + "library": "transformers", + "device": "cuda", + "device_ids": "0", + "seed": 42, + "inter_op_num_threads": null, + "intra_op_num_threads": null, + "hub_kwargs": { + "revision": "main", + "force_download": false, + "local_files_only": false, + "trust_remote_code": true + }, + "no_weights": true, + "device_map": null, + "torch_dtype": null, + "amp_autocast": false, + "amp_dtype": null, + "eval_mode": true, + "to_bettertransformer": false, + "low_cpu_mem_usage": null, + "attn_implementation": null, + "cache_implementation": null, + "torch_compile": false, + "torch_compile_config": {}, + "quantization_scheme": null, + "quantization_config": {}, + "deepspeed_inference": false, + "deepspeed_inference_config": {}, + "peft_type": null, + "peft_config": {} + }, + "launcher": { + "name": "process", + "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", + "device_isolation": true, + "device_isolation_action": "warn", + "start_method": "spawn" + }, + "benchmark": { + "name": "energy_star", + "_target_": "optimum_benchmark.benchmarks.energy_star.benchmark.EnergyStarBenchmark", + "dataset_name": "EnergyStarAI/text_classification", + "dataset_config": "", + "dataset_split": "train", + "num_samples": 1000, + "input_shapes": { + "batch_size": 1 + }, + "text_column_name": "text", + "truncation": true, + "max_length": -1, + "dataset_prefix1": "", + "dataset_prefix2": "", + "t5_task": "", + "image_column_name": "image", + "resize": false, + "question_column_name": "question", + "context_column_name": "context", + "sentence1_column_name": "sentence1", + "sentence2_column_name": "sentence2", + "audio_column_name": "audio", + "iterations": 10, + "warmup_runs": 10, + "energy": true, + "forward_kwargs": {}, + "generate_kwargs": {}, + "call_kwargs": {} + }, + "environment": { + "cpu": " AMD EPYC 7R32", + "cpu_count": 48, + "cpu_ram_mb": 200472.73984, + "system": "Linux", + "machine": "x86_64", + "platform": "Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35", + "processor": "x86_64", + "python_version": "3.9.20", + "gpu": [ + "NVIDIA A10G" + ], + "gpu_count": 1, + "gpu_vram_mb": 24146608128, + "optimum_benchmark_version": "0.2.0", + "optimum_benchmark_commit": null, + "transformers_version": "4.44.0", + "transformers_commit": null, + "accelerate_version": "0.33.0", + "accelerate_commit": null, + "diffusers_version": "0.30.0", + "diffusers_commit": null, + "optimum_version": null, + "optimum_commit": null, + "timm_version": null, + "timm_commit": null, + "peft_version": null, + "peft_commit": null + } +} \ No newline at end of file diff --git a/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-21-19/.hydra/config.yaml b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-21-19/.hydra/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..ca9c7622983c00dc182a28ef5197b99186dbcca4 --- /dev/null +++ b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-21-19/.hydra/config.yaml @@ -0,0 +1,94 @@ +backend: + name: pytorch + version: 2.4.0 + _target_: optimum_benchmark.backends.pytorch.backend.PyTorchBackend + task: text-classification + model: lvwerra/distilbert-imdb + processor: lvwerra/distilbert-imdb + library: null + device: cuda + device_ids: '0' + seed: 42 + inter_op_num_threads: null + intra_op_num_threads: null + hub_kwargs: {} + no_weights: true + device_map: null + torch_dtype: null + amp_autocast: false + amp_dtype: null + eval_mode: true + to_bettertransformer: false + low_cpu_mem_usage: null + attn_implementation: null + cache_implementation: null + torch_compile: false + torch_compile_config: {} + quantization_scheme: null + quantization_config: {} + deepspeed_inference: false + deepspeed_inference_config: {} + peft_type: null + peft_config: {} +launcher: + name: process + _target_: optimum_benchmark.launchers.process.launcher.ProcessLauncher + device_isolation: true + device_isolation_action: warn + start_method: spawn +benchmark: + name: energy_star + _target_: optimum_benchmark.benchmarks.energy_star.benchmark.EnergyStarBenchmark + dataset_name: EnergyStarAI/text_classification + dataset_config: '' + dataset_split: train + num_samples: 1000 + input_shapes: + batch_size: 1 + text_column_name: text + truncation: true + max_length: -1 + dataset_prefix1: '' + dataset_prefix2: '' + t5_task: '' + image_column_name: image + resize: false + question_column_name: question + context_column_name: context + sentence1_column_name: sentence1 + sentence2_column_name: sentence2 + audio_column_name: audio + iterations: 10 + warmup_runs: 10 + energy: true + forward_kwargs: {} + generate_kwargs: {} + call_kwargs: {} +experiment_name: text_classification +environment: + cpu: ' AMD EPYC 7R32' + cpu_count: 48 + cpu_ram_mb: 200472.73984 + system: Linux + machine: x86_64 + platform: Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35 + processor: x86_64 + python_version: 3.9.20 + gpu: + - NVIDIA A10G + gpu_count: 1 + gpu_vram_mb: 24146608128 + optimum_benchmark_version: 0.2.0 + optimum_benchmark_commit: null + transformers_version: 4.44.0 + transformers_commit: null + accelerate_version: 0.33.0 + accelerate_commit: null + diffusers_version: 0.30.0 + diffusers_commit: null + optimum_version: null + optimum_commit: null + timm_version: null + timm_commit: null + peft_version: null + peft_commit: null diff --git a/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-21-19/.hydra/hydra.yaml b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-21-19/.hydra/hydra.yaml new file mode 100644 index 0000000000000000000000000000000000000000..d60a7f3a2d148ad01583601fe15d64811f2b044f --- /dev/null +++ b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-21-19/.hydra/hydra.yaml @@ -0,0 +1,173 @@ +hydra: + run: + dir: /data/runs/${experiment_name}/${backend.model}/${now:%Y-%m-%d-%H-%M-%S} + sweep: + dir: sweeps/${experiment_name}/${now:%Y-%m-%d-%H-%M-%S} + subdir: ${hydra.job.num} + launcher: + _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher + sweeper: + _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper + max_batch_size: null + params: null + help: + app_name: ${hydra.job.name} + header: '${hydra.help.app_name} is powered by Hydra. + + ' + footer: 'Powered by Hydra (https://hydra.cc) + + Use --hydra-help to view Hydra specific help + + ' + template: '${hydra.help.header} + + == Configuration groups == + + Compose your configuration from those groups (group=option) + + + $APP_CONFIG_GROUPS + + + == Config == + + Override anything in the config (foo.bar=value) + + + $CONFIG + + + ${hydra.help.footer} + + ' + hydra_help: + template: 'Hydra (${hydra.runtime.version}) + + See https://hydra.cc for more info. + + + == Flags == + + $FLAGS_HELP + + + == Configuration groups == + + Compose your configuration from those groups (For example, append hydra/job_logging=disabled + to command line) + + + $HYDRA_CONFIG_GROUPS + + + Use ''--cfg hydra'' to Show the Hydra config. + + ' + hydra_help: ??? + hydra_logging: + version: 1 + formatters: + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s' + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + root: + level: INFO + handlers: + - console + disable_existing_loggers: false + job_logging: + version: 1 + formatters: + simple: + format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s' + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s] + - %(message)s' + log_colors: + DEBUG: purple + INFO: green + WARNING: yellow + ERROR: red + CRITICAL: red + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + file: + class: logging.FileHandler + formatter: simple + filename: ${hydra.job.name}.log + root: + level: INFO + handlers: + - console + - file + disable_existing_loggers: false + env: {} + mode: RUN + searchpath: [] + callbacks: {} + output_subdir: .hydra + overrides: + hydra: + - hydra.run.dir=/data/runs/${experiment_name}/${backend.model}/${now:%Y-%m-%d-%H-%M-%S} + - hydra.mode=RUN + task: [] + job: + name: cli + chdir: true + override_dirname: '' + id: ??? + num: ??? + config_name: text_classification + env_set: + OVERRIDE_BENCHMARKS: '1' + env_copy: [] + config: + override_dirname: + kv_sep: '=' + item_sep: ',' + exclude_keys: [] + runtime: + version: 1.3.2 + version_base: '1.3' + cwd: / + config_sources: + - path: hydra.conf + schema: pkg + provider: hydra + - path: optimum_benchmark + schema: pkg + provider: main + - path: hydra_plugins.hydra_colorlog.conf + schema: pkg + provider: hydra-colorlog + - path: /optimum-benchmark/examples/energy_star + schema: file + provider: command-line + - path: '' + schema: structured + provider: schema + output_dir: /data/runs/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-21-19 + choices: + benchmark: energy_star + launcher: process + backend: pytorch + hydra/env: default + hydra/callbacks: null + hydra/job_logging: colorlog + hydra/hydra_logging: colorlog + hydra/hydra_help: default + hydra/help: default + hydra/sweeper: basic + hydra/launcher: basic + hydra/output: default + verbose: false diff --git a/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-21-19/.hydra/overrides.yaml b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-21-19/.hydra/overrides.yaml new file mode 100644 index 0000000000000000000000000000000000000000..fe51488c7066f6687ef680d6bfaa4f7768ef205c --- /dev/null +++ b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-21-19/.hydra/overrides.yaml @@ -0,0 +1 @@ +[] diff --git a/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-21-19/benchmark_report.json b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-21-19/benchmark_report.json new file mode 100644 index 0000000000000000000000000000000000000000..43274ee954f40a21b91e2b640346f01d2bb9915e --- /dev/null +++ b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-21-19/benchmark_report.json @@ -0,0 +1,107 @@ +{ + "forward": { + "memory": null, + "latency": null, + "throughput": null, + "energy": { + "unit": "kWh", + "cpu": 4.266393853194031e-05, + "ram": 3.4665989535233225e-07, + "gpu": 0.00016836457913599334, + "total": 0.00021137517756328596 + }, + "efficiency": { + "unit": "samples/kWh", + "value": 4730924.470545265 + }, + "measures": [ + { + "unit": "kWh", + "cpu": 4.7530642727077886e-05, + "ram": 3.8606025568381555e-07, + "gpu": 0.0001853851483080149, + "total": 0.00023330185129077655 + }, + { + "unit": "kWh", + "cpu": 4.738324727292012e-05, + "ram": 3.849445379185505e-07, + "gpu": 0.0001832601466079664, + "total": 0.00023102833841880512 + }, + { + "unit": "kWh", + "cpu": 4.739282785904935e-05, + "ram": 3.8512579460033693e-07, + "gpu": 0.00018434098080599703, + "total": 0.00023211893445964667 + }, + { + "unit": "kWh", + "cpu": 4.697727643540223e-05, + "ram": 3.817560603545082e-07, + "gpu": 0.0001861507044760291, + "total": 0.00023350973697178591 + }, + { + "unit": "kWh", + "cpu": 4.7635639366671966e-05, + "ram": 3.871280953400939e-07, + "gpu": 0.0001859029265000034, + "total": 0.00023392569396201552 + }, + { + "unit": "kWh", + "cpu": 4.691363451527451e-05, + "ram": 3.8125916546713195e-07, + "gpu": 0.00018823459503192197, + "total": 0.0002355294887126636 + }, + { + "unit": "kWh", + "cpu": 0.0, + "ram": 0.0, + "gpu": 0.0, + "total": 0.0 + }, + { + "unit": "kWh", + "cpu": 4.7470307519451016e-05, + "ram": 3.857454732798674e-07, + "gpu": 0.00018852265081792385, + "total": 0.0002363787038106547 + }, + { + "unit": "kWh", + "cpu": 4.797277234024983e-05, + "ram": 3.89746227983841e-07, + "gpu": 0.00019253932069807078, + "total": 0.00024090183926630442 + }, + { + "unit": "kWh", + "cpu": 4.736303728330619e-05, + "ram": 3.8483334289517675e-07, + "gpu": 0.00018930931811400598, + "total": 0.0002370571887402072 + } + ] + }, + "preprocess": { + "memory": null, + "latency": null, + "throughput": null, + "energy": { + "unit": "kWh", + "cpu": 5.02041641244533e-06, + "ram": 3.224512999262444e-08, + "gpu": 7.722228399953401e-06, + "total": 1.2774889942391356e-05 + }, + "efficiency": { + "unit": "samples/kWh", + "value": 78278560.87289377 + }, + "measures": null + } +} \ No newline at end of file diff --git a/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-21-19/cli.log b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-21-19/cli.log new file mode 100644 index 0000000000000000000000000000000000000000..e384d26cb6ee2e34a93b262d85ee282d28f3cc7b --- /dev/null +++ b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-21-19/cli.log @@ -0,0 +1,113 @@ +[2024-10-01 19:21:19,583][launcher][INFO] - ََAllocating process launcher +[2024-10-01 19:21:19,583][process][INFO] - + Setting multiprocessing start method to spawn. +[2024-10-01 19:21:19,594][device-isolation][INFO] - + Launched device(s) isolation process 66 +[2024-10-01 19:21:19,595][device-isolation][INFO] - + Isolating device(s) [0] +[2024-10-01 19:21:19,602][process][INFO] - + Launched benchmark in isolated process 67. +[PROC-0][2024-10-01 19:21:22,226][datasets][INFO] - PyTorch version 2.4.0 available. +[PROC-0][2024-10-01 19:21:23,166][backend][INFO] - َAllocating pytorch backend +[PROC-0][2024-10-01 19:21:23,166][backend][INFO] - + Setting random seed to 42 +[PROC-0][2024-10-01 19:21:24,202][pytorch][INFO] - + Using AutoModel class AutoModelForSequenceClassification +[PROC-0][2024-10-01 19:21:24,202][pytorch][INFO] - + Creating backend temporary directory +[PROC-0][2024-10-01 19:21:24,202][pytorch][INFO] - + Loading model with random weights +[PROC-0][2024-10-01 19:21:24,202][pytorch][INFO] - + Creating no weights model +[PROC-0][2024-10-01 19:21:24,202][pytorch][INFO] - + Creating no weights model directory +[PROC-0][2024-10-01 19:21:24,203][pytorch][INFO] - + Creating no weights model state dict +[PROC-0][2024-10-01 19:21:24,205][pytorch][INFO] - + Saving no weights model safetensors +[PROC-0][2024-10-01 19:21:24,205][pytorch][INFO] - + Saving no weights model pretrained config +[PROC-0][2024-10-01 19:21:24,206][pytorch][INFO] - + Loading no weights AutoModel +[PROC-0][2024-10-01 19:21:24,206][pytorch][INFO] - + Loading model directly on device: cuda +[PROC-0][2024-10-01 19:21:24,479][pytorch][INFO] - + Turning on model's eval mode +[PROC-0][2024-10-01 19:21:24,485][benchmark][INFO] - Allocating energy_star benchmark +[PROC-0][2024-10-01 19:21:24,485][energy_star][INFO] - + Loading raw dataset +[PROC-0][2024-10-01 19:21:25,223][energy_star][INFO] - + Initializing Inference report +[PROC-0][2024-10-01 19:21:25,223][energy][INFO] - + Tracking GPU energy on devices [0] +[PROC-0][2024-10-01 19:21:29,422][energy_star][INFO] - + Preprocessing dataset +[PROC-0][2024-10-01 19:21:29,850][energy][INFO] - + Saving codecarbon emission data to preprocess_codecarbon.json +[PROC-0][2024-10-01 19:21:29,853][energy_star][INFO] - + Preparing backend for Inference +[PROC-0][2024-10-01 19:21:29,853][energy_star][INFO] - + Initialising dataloader +[PROC-0][2024-10-01 19:21:29,854][energy_star][INFO] - + Warming up backend for Inference +[PROC-0][2024-10-01 19:21:30,349][energy_star][INFO] - + Running Inference energy tracking for 10 iterations +[PROC-0][2024-10-01 19:21:30,349][energy_star][INFO] - + Iteration 1/10 +[PROC-0][2024-10-01 19:21:34,378][energy][INFO] - + Saving codecarbon emission data to forward_codecarbon.json +[PROC-0][2024-10-01 19:21:34,380][energy_star][INFO] - + Iteration 2/10 +[PROC-0][2024-10-01 19:21:38,397][energy][INFO] - + Saving codecarbon emission data to forward_codecarbon.json +[PROC-0][2024-10-01 19:21:38,401][energy_star][INFO] - + Iteration 3/10 +[PROC-0][2024-10-01 19:21:42,418][energy][INFO] - + Saving codecarbon emission data to forward_codecarbon.json +[PROC-0][2024-10-01 19:21:42,421][energy_star][INFO] - + Iteration 4/10 +[PROC-0][2024-10-01 19:21:46,404][energy][INFO] - + Saving codecarbon emission data to forward_codecarbon.json +[PROC-0][2024-10-01 19:21:46,407][energy_star][INFO] - + Iteration 5/10 +[PROC-0][2024-10-01 19:21:50,445][energy][INFO] - + Saving codecarbon emission data to forward_codecarbon.json +[PROC-0][2024-10-01 19:21:50,448][energy_star][INFO] - + Iteration 6/10 +[PROC-0][2024-10-01 19:21:54,425][energy][INFO] - + Saving codecarbon emission data to forward_codecarbon.json +[PROC-0][2024-10-01 19:21:54,427][energy_star][INFO] - + Iteration 7/10 +[PROC-0][2024-10-01 19:21:58,411][energy][INFO] - + Saving codecarbon emission data to forward_codecarbon.json +[PROC-0][2024-10-01 19:21:58,414][energy_star][INFO] - + Iteration 8/10 +[PROC-0][2024-10-01 19:22:02,438][energy][INFO] - + Saving codecarbon emission data to forward_codecarbon.json +[PROC-0][2024-10-01 19:22:02,442][energy_star][INFO] - + Iteration 9/10 +[PROC-0][2024-10-01 19:22:06,509][energy][INFO] - + Saving codecarbon emission data to forward_codecarbon.json +[PROC-0][2024-10-01 19:22:06,512][energy_star][INFO] - + Iteration 10/10 +[PROC-0][2024-10-01 19:22:10,527][energy][INFO] - + Saving codecarbon emission data to forward_codecarbon.json +[PROC-0][2024-10-01 19:22:10,530][energy][INFO] - + forward energy consumption: +[PROC-0][2024-10-01 19:22:10,530][energy][INFO] - + CPU: 0.000043 (kWh) +[PROC-0][2024-10-01 19:22:10,530][energy][INFO] - + GPU: 0.000168 (kWh) +[PROC-0][2024-10-01 19:22:10,530][energy][INFO] - + RAM: 0.000000 (kWh) +[PROC-0][2024-10-01 19:22:10,530][energy][INFO] - + total: 0.000211 (kWh) +[PROC-0][2024-10-01 19:22:10,530][energy][INFO] - + forward_iteration_1 energy consumption: +[PROC-0][2024-10-01 19:22:10,530][energy][INFO] - + CPU: 0.000048 (kWh) +[PROC-0][2024-10-01 19:22:10,530][energy][INFO] - + GPU: 0.000185 (kWh) +[PROC-0][2024-10-01 19:22:10,530][energy][INFO] - + RAM: 0.000000 (kWh) +[PROC-0][2024-10-01 19:22:10,530][energy][INFO] - + total: 0.000233 (kWh) +[PROC-0][2024-10-01 19:22:10,531][energy][INFO] - + forward_iteration_2 energy consumption: +[PROC-0][2024-10-01 19:22:10,531][energy][INFO] - + CPU: 0.000047 (kWh) +[PROC-0][2024-10-01 19:22:10,531][energy][INFO] - + GPU: 0.000183 (kWh) +[PROC-0][2024-10-01 19:22:10,531][energy][INFO] - + RAM: 0.000000 (kWh) +[PROC-0][2024-10-01 19:22:10,531][energy][INFO] - + total: 0.000231 (kWh) +[PROC-0][2024-10-01 19:22:10,531][energy][INFO] - + forward_iteration_3 energy consumption: +[PROC-0][2024-10-01 19:22:10,531][energy][INFO] - + CPU: 0.000047 (kWh) +[PROC-0][2024-10-01 19:22:10,531][energy][INFO] - + GPU: 0.000184 (kWh) +[PROC-0][2024-10-01 19:22:10,531][energy][INFO] - + RAM: 0.000000 (kWh) +[PROC-0][2024-10-01 19:22:10,531][energy][INFO] - + total: 0.000232 (kWh) +[PROC-0][2024-10-01 19:22:10,532][energy][INFO] - + forward_iteration_4 energy consumption: +[PROC-0][2024-10-01 19:22:10,532][energy][INFO] - + CPU: 0.000047 (kWh) +[PROC-0][2024-10-01 19:22:10,532][energy][INFO] - + GPU: 0.000186 (kWh) +[PROC-0][2024-10-01 19:22:10,532][energy][INFO] - + RAM: 0.000000 (kWh) +[PROC-0][2024-10-01 19:22:10,532][energy][INFO] - + total: 0.000234 (kWh) +[PROC-0][2024-10-01 19:22:10,532][energy][INFO] - + forward_iteration_5 energy consumption: +[PROC-0][2024-10-01 19:22:10,532][energy][INFO] - + CPU: 0.000048 (kWh) +[PROC-0][2024-10-01 19:22:10,532][energy][INFO] - + GPU: 0.000186 (kWh) +[PROC-0][2024-10-01 19:22:10,532][energy][INFO] - + RAM: 0.000000 (kWh) +[PROC-0][2024-10-01 19:22:10,532][energy][INFO] - + total: 0.000234 (kWh) +[PROC-0][2024-10-01 19:22:10,533][energy][INFO] - + forward_iteration_6 energy consumption: +[PROC-0][2024-10-01 19:22:10,533][energy][INFO] - + CPU: 0.000047 (kWh) +[PROC-0][2024-10-01 19:22:10,533][energy][INFO] - + GPU: 0.000188 (kWh) +[PROC-0][2024-10-01 19:22:10,533][energy][INFO] - + RAM: 0.000000 (kWh) +[PROC-0][2024-10-01 19:22:10,533][energy][INFO] - + total: 0.000236 (kWh) +[PROC-0][2024-10-01 19:22:10,533][energy][INFO] - + forward_iteration_7 energy consumption: +[PROC-0][2024-10-01 19:22:10,533][energy][INFO] - + CPU: 0.000000 (kWh) +[PROC-0][2024-10-01 19:22:10,533][energy][INFO] - + GPU: 0.000000 (kWh) +[PROC-0][2024-10-01 19:22:10,533][energy][INFO] - + RAM: 0.000000 (kWh) +[PROC-0][2024-10-01 19:22:10,533][energy][INFO] - + total: 0.000000 (kWh) +[PROC-0][2024-10-01 19:22:10,534][energy][INFO] - + forward_iteration_8 energy consumption: +[PROC-0][2024-10-01 19:22:10,534][energy][INFO] - + CPU: 0.000047 (kWh) +[PROC-0][2024-10-01 19:22:10,534][energy][INFO] - + GPU: 0.000189 (kWh) +[PROC-0][2024-10-01 19:22:10,534][energy][INFO] - + RAM: 0.000000 (kWh) +[PROC-0][2024-10-01 19:22:10,534][energy][INFO] - + total: 0.000236 (kWh) +[PROC-0][2024-10-01 19:22:10,534][energy][INFO] - + forward_iteration_9 energy consumption: +[PROC-0][2024-10-01 19:22:10,534][energy][INFO] - + CPU: 0.000048 (kWh) +[PROC-0][2024-10-01 19:22:10,534][energy][INFO] - + GPU: 0.000193 (kWh) +[PROC-0][2024-10-01 19:22:10,534][energy][INFO] - + RAM: 0.000000 (kWh) +[PROC-0][2024-10-01 19:22:10,534][energy][INFO] - + total: 0.000241 (kWh) +[PROC-0][2024-10-01 19:22:10,535][energy][INFO] - + forward_iteration_10 energy consumption: +[PROC-0][2024-10-01 19:22:10,535][energy][INFO] - + CPU: 0.000047 (kWh) +[PROC-0][2024-10-01 19:22:10,535][energy][INFO] - + GPU: 0.000189 (kWh) +[PROC-0][2024-10-01 19:22:10,535][energy][INFO] - + RAM: 0.000000 (kWh) +[PROC-0][2024-10-01 19:22:10,535][energy][INFO] - + total: 0.000237 (kWh) +[PROC-0][2024-10-01 19:22:10,535][energy][INFO] - + preprocess energy consumption: +[PROC-0][2024-10-01 19:22:10,535][energy][INFO] - + CPU: 0.000005 (kWh) +[PROC-0][2024-10-01 19:22:10,535][energy][INFO] - + GPU: 0.000008 (kWh) +[PROC-0][2024-10-01 19:22:10,535][energy][INFO] - + RAM: 0.000000 (kWh) +[PROC-0][2024-10-01 19:22:10,535][energy][INFO] - + total: 0.000013 (kWh) +[PROC-0][2024-10-01 19:22:10,536][energy][INFO] - + forward energy efficiency: 4730924.470545 (samples/kWh) +[PROC-0][2024-10-01 19:22:10,536][energy][INFO] - + preprocess energy efficiency: 78278560.872894 (samples/kWh) +[2024-10-01 19:22:11,131][device-isolation][INFO] - + Closing device(s) isolation process... +[2024-10-01 19:22:11,179][datasets][INFO] - PyTorch version 2.4.0 available. diff --git a/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-21-19/experiment_config.json b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-21-19/experiment_config.json new file mode 100644 index 0000000000000000000000000000000000000000..772aa352d839018f6caa9107fd24acf54618c6aa --- /dev/null +++ b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-21-19/experiment_config.json @@ -0,0 +1,107 @@ +{ + "experiment_name": "text_classification", + "backend": { + "name": "pytorch", + "version": "2.4.0", + "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", + "task": "text-classification", + "model": "lvwerra/distilbert-imdb", + "processor": "lvwerra/distilbert-imdb", + "library": "transformers", + "device": "cuda", + "device_ids": "0", + "seed": 42, + "inter_op_num_threads": null, + "intra_op_num_threads": null, + "hub_kwargs": { + "revision": "main", + "force_download": false, + "local_files_only": false, + "trust_remote_code": true + }, + "no_weights": true, + "device_map": null, + "torch_dtype": null, + "amp_autocast": false, + "amp_dtype": null, + "eval_mode": true, + "to_bettertransformer": false, + "low_cpu_mem_usage": null, + "attn_implementation": null, + "cache_implementation": null, + "torch_compile": false, + "torch_compile_config": {}, + "quantization_scheme": null, + "quantization_config": {}, + "deepspeed_inference": false, + "deepspeed_inference_config": {}, + "peft_type": null, + "peft_config": {} + }, + "launcher": { + "name": "process", + "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", + "device_isolation": true, + "device_isolation_action": "warn", + "start_method": "spawn" + }, + "benchmark": { + "name": "energy_star", + "_target_": "optimum_benchmark.benchmarks.energy_star.benchmark.EnergyStarBenchmark", + "dataset_name": "EnergyStarAI/text_classification", + "dataset_config": "", + "dataset_split": "train", + "num_samples": 1000, + "input_shapes": { + "batch_size": 1 + }, + "text_column_name": "text", + "truncation": true, + "max_length": -1, + "dataset_prefix1": "", + "dataset_prefix2": "", + "t5_task": "", + "image_column_name": "image", + "resize": false, + "question_column_name": "question", + "context_column_name": "context", + "sentence1_column_name": "sentence1", + "sentence2_column_name": "sentence2", + "audio_column_name": "audio", + "iterations": 10, + "warmup_runs": 10, + "energy": true, + "forward_kwargs": {}, + "generate_kwargs": {}, + "call_kwargs": {} + }, + "environment": { + "cpu": " AMD EPYC 7R32", + "cpu_count": 48, + "cpu_ram_mb": 200472.73984, + "system": "Linux", + "machine": "x86_64", + "platform": "Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35", + "processor": "x86_64", + "python_version": "3.9.20", + "gpu": [ + "NVIDIA A10G" + ], + "gpu_count": 1, + "gpu_vram_mb": 24146608128, + "optimum_benchmark_version": "0.2.0", + "optimum_benchmark_commit": null, + "transformers_version": "4.44.0", + "transformers_commit": null, + "accelerate_version": "0.33.0", + "accelerate_commit": null, + "diffusers_version": "0.30.0", + "diffusers_commit": null, + "optimum_version": null, + "optimum_commit": null, + "timm_version": null, + "timm_commit": null, + "peft_version": null, + "peft_commit": null + } +} \ No newline at end of file diff --git a/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-21-19/forward_codecarbon.json b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-21-19/forward_codecarbon.json new file mode 100644 index 0000000000000000000000000000000000000000..f7ab537e64b59f6652b804352c259b244bb30e39 --- /dev/null +++ b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-21-19/forward_codecarbon.json @@ -0,0 +1,33 @@ +{ + "timestamp": "2024-10-01T19:22:10", + "project_name": "codecarbon", + "run_id": "a4609e1f-c533-456f-8bf6-28ba55ba7cdd", + "duration": -1727750784.2951422, + "emissions": 8.750595789014414e-05, + "emissions_rate": 2.182205858635412e-05, + "cpu_power": 42.5, + "gpu_power": 170.00339141921057, + "ram_power": 0.34564733505249023, + "cpu_energy": 4.736303728330619e-05, + "gpu_energy": 0.00018930931811400598, + "ram_energy": 3.8483334289517675e-07, + "energy_consumed": 0.0002370571887402072, + "country_name": "United States", + "country_iso_code": "USA", + "region": "virginia", + "cloud_provider": "", + "cloud_region": "", + "os": "Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35", + "python_version": "3.9.20", + "codecarbon_version": "2.5.1", + "cpu_count": 48, + "cpu_model": "AMD EPYC 7R32", + "gpu_count": 1, + "gpu_model": "1 x NVIDIA A10G", + "longitude": -77.4903, + "latitude": 39.0469, + "ram_total_size": 186.7047882080078, + "tracking_mode": "process", + "on_cloud": "N", + "pue": 1.0 +} \ No newline at end of file diff --git a/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-21-19/preprocess_codecarbon.json b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-21-19/preprocess_codecarbon.json new file mode 100644 index 0000000000000000000000000000000000000000..1b227138bf38ee09bbff0c7b6653eb440aa30421 --- /dev/null +++ b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-21-19/preprocess_codecarbon.json @@ -0,0 +1,33 @@ +{ + "timestamp": "2024-10-01T19:21:29", + "project_name": "codecarbon", + "run_id": "a4609e1f-c533-456f-8bf6-28ba55ba7cdd", + "duration": -1727750787.8838358, + "emissions": 4.715651051507304e-06, + "emissions_rate": 1.119145951252703e-05, + "cpu_power": 42.5, + "gpu_power": 65.72069424710368, + "ram_power": 0.2749357223510742, + "cpu_energy": 5.02041641244533e-06, + "gpu_energy": 7.722228399953401e-06, + "ram_energy": 3.224512999262444e-08, + "energy_consumed": 1.2774889942391356e-05, + "country_name": "United States", + "country_iso_code": "USA", + "region": "virginia", + "cloud_provider": "", + "cloud_region": "", + "os": "Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35", + "python_version": "3.9.20", + "codecarbon_version": "2.5.1", + "cpu_count": 48, + "cpu_model": "AMD EPYC 7R32", + "gpu_count": 1, + "gpu_model": "1 x NVIDIA A10G", + "longitude": -77.4903, + "latitude": 39.0469, + "ram_total_size": 186.7047882080078, + "tracking_mode": "process", + "on_cloud": "N", + "pue": 1.0 +} \ No newline at end of file diff --git a/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-53-10/.hydra/config.yaml b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-53-10/.hydra/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..ca9c7622983c00dc182a28ef5197b99186dbcca4 --- /dev/null +++ b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-53-10/.hydra/config.yaml @@ -0,0 +1,94 @@ +backend: + name: pytorch + version: 2.4.0 + _target_: optimum_benchmark.backends.pytorch.backend.PyTorchBackend + task: text-classification + model: lvwerra/distilbert-imdb + processor: lvwerra/distilbert-imdb + library: null + device: cuda + device_ids: '0' + seed: 42 + inter_op_num_threads: null + intra_op_num_threads: null + hub_kwargs: {} + no_weights: true + device_map: null + torch_dtype: null + amp_autocast: false + amp_dtype: null + eval_mode: true + to_bettertransformer: false + low_cpu_mem_usage: null + attn_implementation: null + cache_implementation: null + torch_compile: false + torch_compile_config: {} + quantization_scheme: null + quantization_config: {} + deepspeed_inference: false + deepspeed_inference_config: {} + peft_type: null + peft_config: {} +launcher: + name: process + _target_: optimum_benchmark.launchers.process.launcher.ProcessLauncher + device_isolation: true + device_isolation_action: warn + start_method: spawn +benchmark: + name: energy_star + _target_: optimum_benchmark.benchmarks.energy_star.benchmark.EnergyStarBenchmark + dataset_name: EnergyStarAI/text_classification + dataset_config: '' + dataset_split: train + num_samples: 1000 + input_shapes: + batch_size: 1 + text_column_name: text + truncation: true + max_length: -1 + dataset_prefix1: '' + dataset_prefix2: '' + t5_task: '' + image_column_name: image + resize: false + question_column_name: question + context_column_name: context + sentence1_column_name: sentence1 + sentence2_column_name: sentence2 + audio_column_name: audio + iterations: 10 + warmup_runs: 10 + energy: true + forward_kwargs: {} + generate_kwargs: {} + call_kwargs: {} +experiment_name: text_classification +environment: + cpu: ' AMD EPYC 7R32' + cpu_count: 48 + cpu_ram_mb: 200472.73984 + system: Linux + machine: x86_64 + platform: Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35 + processor: x86_64 + python_version: 3.9.20 + gpu: + - NVIDIA A10G + gpu_count: 1 + gpu_vram_mb: 24146608128 + optimum_benchmark_version: 0.2.0 + optimum_benchmark_commit: null + transformers_version: 4.44.0 + transformers_commit: null + accelerate_version: 0.33.0 + accelerate_commit: null + diffusers_version: 0.30.0 + diffusers_commit: null + optimum_version: null + optimum_commit: null + timm_version: null + timm_commit: null + peft_version: null + peft_commit: null diff --git a/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-53-10/.hydra/hydra.yaml b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-53-10/.hydra/hydra.yaml new file mode 100644 index 0000000000000000000000000000000000000000..5a51519a9426ef4733201aee39fe3da11d1ac113 --- /dev/null +++ b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-53-10/.hydra/hydra.yaml @@ -0,0 +1,173 @@ +hydra: + run: + dir: /data/runs/${experiment_name}/${backend.model}/${now:%Y-%m-%d-%H-%M-%S} + sweep: + dir: sweeps/${experiment_name}/${now:%Y-%m-%d-%H-%M-%S} + subdir: ${hydra.job.num} + launcher: + _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher + sweeper: + _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper + max_batch_size: null + params: null + help: + app_name: ${hydra.job.name} + header: '${hydra.help.app_name} is powered by Hydra. + + ' + footer: 'Powered by Hydra (https://hydra.cc) + + Use --hydra-help to view Hydra specific help + + ' + template: '${hydra.help.header} + + == Configuration groups == + + Compose your configuration from those groups (group=option) + + + $APP_CONFIG_GROUPS + + + == Config == + + Override anything in the config (foo.bar=value) + + + $CONFIG + + + ${hydra.help.footer} + + ' + hydra_help: + template: 'Hydra (${hydra.runtime.version}) + + See https://hydra.cc for more info. + + + == Flags == + + $FLAGS_HELP + + + == Configuration groups == + + Compose your configuration from those groups (For example, append hydra/job_logging=disabled + to command line) + + + $HYDRA_CONFIG_GROUPS + + + Use ''--cfg hydra'' to Show the Hydra config. + + ' + hydra_help: ??? + hydra_logging: + version: 1 + formatters: + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s' + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + root: + level: INFO + handlers: + - console + disable_existing_loggers: false + job_logging: + version: 1 + formatters: + simple: + format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s' + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s] + - %(message)s' + log_colors: + DEBUG: purple + INFO: green + WARNING: yellow + ERROR: red + CRITICAL: red + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + file: + class: logging.FileHandler + formatter: simple + filename: ${hydra.job.name}.log + root: + level: INFO + handlers: + - console + - file + disable_existing_loggers: false + env: {} + mode: RUN + searchpath: [] + callbacks: {} + output_subdir: .hydra + overrides: + hydra: + - hydra.run.dir=/data/runs/${experiment_name}/${backend.model}/${now:%Y-%m-%d-%H-%M-%S} + - hydra.mode=RUN + task: [] + job: + name: cli + chdir: true + override_dirname: '' + id: ??? + num: ??? + config_name: text_classification + env_set: + OVERRIDE_BENCHMARKS: '1' + env_copy: [] + config: + override_dirname: + kv_sep: '=' + item_sep: ',' + exclude_keys: [] + runtime: + version: 1.3.2 + version_base: '1.3' + cwd: / + config_sources: + - path: hydra.conf + schema: pkg + provider: hydra + - path: optimum_benchmark + schema: pkg + provider: main + - path: hydra_plugins.hydra_colorlog.conf + schema: pkg + provider: hydra-colorlog + - path: /optimum-benchmark/examples/energy_star + schema: file + provider: command-line + - path: '' + schema: structured + provider: schema + output_dir: /data/runs/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-53-10 + choices: + benchmark: energy_star + launcher: process + backend: pytorch + hydra/env: default + hydra/callbacks: null + hydra/job_logging: colorlog + hydra/hydra_logging: colorlog + hydra/hydra_help: default + hydra/help: default + hydra/sweeper: basic + hydra/launcher: basic + hydra/output: default + verbose: false diff --git a/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-53-10/.hydra/overrides.yaml b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-53-10/.hydra/overrides.yaml new file mode 100644 index 0000000000000000000000000000000000000000..fe51488c7066f6687ef680d6bfaa4f7768ef205c --- /dev/null +++ b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-53-10/.hydra/overrides.yaml @@ -0,0 +1 @@ +[] diff --git a/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-53-10/benchmark_report.json b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-53-10/benchmark_report.json new file mode 100644 index 0000000000000000000000000000000000000000..d7e98dd18baeea95c7c2cb06e1c48822f21b4410 --- /dev/null +++ b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-53-10/benchmark_report.json @@ -0,0 +1,107 @@ +{ + "forward": { + "memory": null, + "latency": null, + "throughput": null, + "energy": { + "unit": "kWh", + "cpu": 4.328938041541177e-05, + "ram": 3.5140342214037094e-07, + "gpu": 0.00016919610757900604, + "total": 0.00021283689141655813 + }, + "efficiency": { + "unit": "samples/kWh", + "value": 4698433.590832847 + }, + "measures": [ + { + "unit": "kWh", + "cpu": 4.837897983191195e-05, + "ram": 3.926649739371926e-07, + "gpu": 0.0001845559809779962, + "total": 0.00023332762578384537 + }, + { + "unit": "kWh", + "cpu": 4.784201455001191e-05, + "ram": 3.883550096135927e-07, + "gpu": 0.00018761320564608663, + "total": 0.00023584357520571212 + }, + { + "unit": "kWh", + "cpu": 4.8432672206935324e-05, + "ram": 3.931606991900786e-07, + "gpu": 0.00018643653803795956, + "total": 0.000235262370944085 + }, + { + "unit": "kWh", + "cpu": 4.787912353126559e-05, + "ram": 3.8866412265046843e-07, + "gpu": 0.0001842659807459457, + "total": 0.0002325337683998618 + }, + { + "unit": "kWh", + "cpu": 4.9012056703440985e-05, + "ram": 3.9783468990291465e-07, + "gpu": 0.00019145904205608577, + "total": 0.0002408689334494295 + }, + { + "unit": "kWh", + "cpu": 4.8268223981954694e-05, + "ram": 3.9183497590251925e-07, + "gpu": 0.00018912820685790255, + "total": 0.00023778826581575966 + }, + { + "unit": "kWh", + "cpu": 0.0, + "ram": 0.0, + "gpu": 0.0, + "total": 0.0 + }, + { + "unit": "kWh", + "cpu": 4.7899580304137e-05, + "ram": 3.8884532021784974e-07, + "gpu": 0.00019074904148796978, + "total": 0.00023903746711232462 + }, + { + "unit": "kWh", + "cpu": 4.752461423195199e-05, + "ram": 3.858019544743239e-07, + "gpu": 0.0001890387623421086, + "total": 0.0002369491785285348 + }, + { + "unit": "kWh", + "cpu": 4.765653881250826e-05, + "ram": 3.868724755147697e-07, + "gpu": 0.00018871431763800572, + "total": 0.00023675772892602868 + } + ] + }, + "preprocess": { + "memory": null, + "latency": null, + "throughput": null, + "energy": { + "unit": "kWh", + "cpu": 5.060345729154708e-06, + "ram": 3.2464527108885925e-08, + "gpu": 9.570563211935834e-06, + "total": 1.4663373468199429e-05 + }, + "efficiency": { + "unit": "samples/kWh", + "value": 68197130.91047621 + }, + "measures": null + } +} \ No newline at end of file diff --git a/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-53-10/cli.log b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-53-10/cli.log new file mode 100644 index 0000000000000000000000000000000000000000..997891d4135872f591750559b3bea466a11e31a6 --- /dev/null +++ b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-53-10/cli.log @@ -0,0 +1,113 @@ +[2024-10-01 19:53:10,577][launcher][INFO] - ََAllocating process launcher +[2024-10-01 19:53:10,577][process][INFO] - + Setting multiprocessing start method to spawn. +[2024-10-01 19:53:10,591][device-isolation][INFO] - + Launched device(s) isolation process 66 +[2024-10-01 19:53:10,592][device-isolation][INFO] - + Isolating device(s) [0] +[2024-10-01 19:53:10,598][process][INFO] - + Launched benchmark in isolated process 67. +[PROC-0][2024-10-01 19:53:13,240][datasets][INFO] - PyTorch version 2.4.0 available. +[PROC-0][2024-10-01 19:53:14,178][backend][INFO] - َAllocating pytorch backend +[PROC-0][2024-10-01 19:53:14,178][backend][INFO] - + Setting random seed to 42 +[PROC-0][2024-10-01 19:53:14,491][pytorch][INFO] - + Using AutoModel class AutoModelForSequenceClassification +[PROC-0][2024-10-01 19:53:14,492][pytorch][INFO] - + Creating backend temporary directory +[PROC-0][2024-10-01 19:53:14,492][pytorch][INFO] - + Loading model with random weights +[PROC-0][2024-10-01 19:53:14,492][pytorch][INFO] - + Creating no weights model +[PROC-0][2024-10-01 19:53:14,492][pytorch][INFO] - + Creating no weights model directory +[PROC-0][2024-10-01 19:53:14,492][pytorch][INFO] - + Creating no weights model state dict +[PROC-0][2024-10-01 19:53:14,494][pytorch][INFO] - + Saving no weights model safetensors +[PROC-0][2024-10-01 19:53:14,495][pytorch][INFO] - + Saving no weights model pretrained config +[PROC-0][2024-10-01 19:53:14,495][pytorch][INFO] - + Loading no weights AutoModel +[PROC-0][2024-10-01 19:53:14,496][pytorch][INFO] - + Loading model directly on device: cuda +[PROC-0][2024-10-01 19:53:14,749][pytorch][INFO] - + Turning on model's eval mode +[PROC-0][2024-10-01 19:53:14,755][benchmark][INFO] - Allocating energy_star benchmark +[PROC-0][2024-10-01 19:53:14,755][energy_star][INFO] - + Loading raw dataset +[PROC-0][2024-10-01 19:53:15,571][energy_star][INFO] - + Initializing Inference report +[PROC-0][2024-10-01 19:53:15,571][energy][INFO] - + Tracking GPU energy on devices [0] +[PROC-0][2024-10-01 19:53:19,761][energy_star][INFO] - + Preprocessing dataset +[PROC-0][2024-10-01 19:53:20,192][energy][INFO] - + Saving codecarbon emission data to preprocess_codecarbon.json +[PROC-0][2024-10-01 19:53:20,194][energy_star][INFO] - + Preparing backend for Inference +[PROC-0][2024-10-01 19:53:20,194][energy_star][INFO] - + Initialising dataloader +[PROC-0][2024-10-01 19:53:20,195][energy_star][INFO] - + Warming up backend for Inference +[PROC-0][2024-10-01 19:53:20,739][energy_star][INFO] - + Running Inference energy tracking for 10 iterations +[PROC-0][2024-10-01 19:53:20,739][energy_star][INFO] - + Iteration 1/10 +[PROC-0][2024-10-01 19:53:24,840][energy][INFO] - + Saving codecarbon emission data to forward_codecarbon.json +[PROC-0][2024-10-01 19:53:24,842][energy_star][INFO] - + Iteration 2/10 +[PROC-0][2024-10-01 19:53:28,897][energy][INFO] - + Saving codecarbon emission data to forward_codecarbon.json +[PROC-0][2024-10-01 19:53:28,899][energy_star][INFO] - + Iteration 3/10 +[PROC-0][2024-10-01 19:53:33,004][energy][INFO] - + Saving codecarbon emission data to forward_codecarbon.json +[PROC-0][2024-10-01 19:53:33,006][energy_star][INFO] - + Iteration 4/10 +[PROC-0][2024-10-01 19:53:37,064][energy][INFO] - + Saving codecarbon emission data to forward_codecarbon.json +[PROC-0][2024-10-01 19:53:37,066][energy_star][INFO] - + Iteration 5/10 +[PROC-0][2024-10-01 19:53:41,221][energy][INFO] - + Saving codecarbon emission data to forward_codecarbon.json +[PROC-0][2024-10-01 19:53:41,223][energy_star][INFO] - + Iteration 6/10 +[PROC-0][2024-10-01 19:53:45,314][energy][INFO] - + Saving codecarbon emission data to forward_codecarbon.json +[PROC-0][2024-10-01 19:53:45,316][energy_star][INFO] - + Iteration 7/10 +[PROC-0][2024-10-01 19:53:49,423][energy][INFO] - + Saving codecarbon emission data to forward_codecarbon.json +[PROC-0][2024-10-01 19:53:49,425][energy_star][INFO] - + Iteration 8/10 +[PROC-0][2024-10-01 19:53:53,485][energy][INFO] - + Saving codecarbon emission data to forward_codecarbon.json +[PROC-0][2024-10-01 19:53:53,487][energy_star][INFO] - + Iteration 9/10 +[PROC-0][2024-10-01 19:53:57,516][energy][INFO] - + Saving codecarbon emission data to forward_codecarbon.json +[PROC-0][2024-10-01 19:53:57,518][energy_star][INFO] - + Iteration 10/10 +[PROC-0][2024-10-01 19:54:01,557][energy][INFO] - + Saving codecarbon emission data to forward_codecarbon.json +[PROC-0][2024-10-01 19:54:01,560][energy][INFO] - + forward energy consumption: +[PROC-0][2024-10-01 19:54:01,560][energy][INFO] - + CPU: 0.000043 (kWh) +[PROC-0][2024-10-01 19:54:01,560][energy][INFO] - + GPU: 0.000169 (kWh) +[PROC-0][2024-10-01 19:54:01,560][energy][INFO] - + RAM: 0.000000 (kWh) +[PROC-0][2024-10-01 19:54:01,560][energy][INFO] - + total: 0.000213 (kWh) +[PROC-0][2024-10-01 19:54:01,560][energy][INFO] - + forward_iteration_1 energy consumption: +[PROC-0][2024-10-01 19:54:01,560][energy][INFO] - + CPU: 0.000048 (kWh) +[PROC-0][2024-10-01 19:54:01,560][energy][INFO] - + GPU: 0.000185 (kWh) +[PROC-0][2024-10-01 19:54:01,560][energy][INFO] - + RAM: 0.000000 (kWh) +[PROC-0][2024-10-01 19:54:01,560][energy][INFO] - + total: 0.000233 (kWh) +[PROC-0][2024-10-01 19:54:01,560][energy][INFO] - + forward_iteration_2 energy consumption: +[PROC-0][2024-10-01 19:54:01,561][energy][INFO] - + CPU: 0.000048 (kWh) +[PROC-0][2024-10-01 19:54:01,561][energy][INFO] - + GPU: 0.000188 (kWh) +[PROC-0][2024-10-01 19:54:01,561][energy][INFO] - + RAM: 0.000000 (kWh) +[PROC-0][2024-10-01 19:54:01,561][energy][INFO] - + total: 0.000236 (kWh) +[PROC-0][2024-10-01 19:54:01,561][energy][INFO] - + forward_iteration_3 energy consumption: +[PROC-0][2024-10-01 19:54:01,561][energy][INFO] - + CPU: 0.000048 (kWh) +[PROC-0][2024-10-01 19:54:01,561][energy][INFO] - + GPU: 0.000186 (kWh) +[PROC-0][2024-10-01 19:54:01,561][energy][INFO] - + RAM: 0.000000 (kWh) +[PROC-0][2024-10-01 19:54:01,561][energy][INFO] - + total: 0.000235 (kWh) +[PROC-0][2024-10-01 19:54:01,561][energy][INFO] - + forward_iteration_4 energy consumption: +[PROC-0][2024-10-01 19:54:01,561][energy][INFO] - + CPU: 0.000048 (kWh) +[PROC-0][2024-10-01 19:54:01,561][energy][INFO] - + GPU: 0.000184 (kWh) +[PROC-0][2024-10-01 19:54:01,561][energy][INFO] - + RAM: 0.000000 (kWh) +[PROC-0][2024-10-01 19:54:01,562][energy][INFO] - + total: 0.000233 (kWh) +[PROC-0][2024-10-01 19:54:01,562][energy][INFO] - + forward_iteration_5 energy consumption: +[PROC-0][2024-10-01 19:54:01,562][energy][INFO] - + CPU: 0.000049 (kWh) +[PROC-0][2024-10-01 19:54:01,562][energy][INFO] - + GPU: 0.000191 (kWh) +[PROC-0][2024-10-01 19:54:01,562][energy][INFO] - + RAM: 0.000000 (kWh) +[PROC-0][2024-10-01 19:54:01,562][energy][INFO] - + total: 0.000241 (kWh) +[PROC-0][2024-10-01 19:54:01,562][energy][INFO] - + forward_iteration_6 energy consumption: +[PROC-0][2024-10-01 19:54:01,562][energy][INFO] - + CPU: 0.000048 (kWh) +[PROC-0][2024-10-01 19:54:01,562][energy][INFO] - + GPU: 0.000189 (kWh) +[PROC-0][2024-10-01 19:54:01,562][energy][INFO] - + RAM: 0.000000 (kWh) +[PROC-0][2024-10-01 19:54:01,562][energy][INFO] - + total: 0.000238 (kWh) +[PROC-0][2024-10-01 19:54:01,562][energy][INFO] - + forward_iteration_7 energy consumption: +[PROC-0][2024-10-01 19:54:01,563][energy][INFO] - + CPU: 0.000000 (kWh) +[PROC-0][2024-10-01 19:54:01,563][energy][INFO] - + GPU: 0.000000 (kWh) +[PROC-0][2024-10-01 19:54:01,563][energy][INFO] - + RAM: 0.000000 (kWh) +[PROC-0][2024-10-01 19:54:01,563][energy][INFO] - + total: 0.000000 (kWh) +[PROC-0][2024-10-01 19:54:01,563][energy][INFO] - + forward_iteration_8 energy consumption: +[PROC-0][2024-10-01 19:54:01,563][energy][INFO] - + CPU: 0.000048 (kWh) +[PROC-0][2024-10-01 19:54:01,563][energy][INFO] - + GPU: 0.000191 (kWh) +[PROC-0][2024-10-01 19:54:01,563][energy][INFO] - + RAM: 0.000000 (kWh) +[PROC-0][2024-10-01 19:54:01,563][energy][INFO] - + total: 0.000239 (kWh) +[PROC-0][2024-10-01 19:54:01,563][energy][INFO] - + forward_iteration_9 energy consumption: +[PROC-0][2024-10-01 19:54:01,563][energy][INFO] - + CPU: 0.000048 (kWh) +[PROC-0][2024-10-01 19:54:01,563][energy][INFO] - + GPU: 0.000189 (kWh) +[PROC-0][2024-10-01 19:54:01,563][energy][INFO] - + RAM: 0.000000 (kWh) +[PROC-0][2024-10-01 19:54:01,564][energy][INFO] - + total: 0.000237 (kWh) +[PROC-0][2024-10-01 19:54:01,564][energy][INFO] - + forward_iteration_10 energy consumption: +[PROC-0][2024-10-01 19:54:01,564][energy][INFO] - + CPU: 0.000048 (kWh) +[PROC-0][2024-10-01 19:54:01,564][energy][INFO] - + GPU: 0.000189 (kWh) +[PROC-0][2024-10-01 19:54:01,564][energy][INFO] - + RAM: 0.000000 (kWh) +[PROC-0][2024-10-01 19:54:01,564][energy][INFO] - + total: 0.000237 (kWh) +[PROC-0][2024-10-01 19:54:01,564][energy][INFO] - + preprocess energy consumption: +[PROC-0][2024-10-01 19:54:01,564][energy][INFO] - + CPU: 0.000005 (kWh) +[PROC-0][2024-10-01 19:54:01,564][energy][INFO] - + GPU: 0.000010 (kWh) +[PROC-0][2024-10-01 19:54:01,564][energy][INFO] - + RAM: 0.000000 (kWh) +[PROC-0][2024-10-01 19:54:01,564][energy][INFO] - + total: 0.000015 (kWh) +[PROC-0][2024-10-01 19:54:01,565][energy][INFO] - + forward energy efficiency: 4698433.590833 (samples/kWh) +[PROC-0][2024-10-01 19:54:01,565][energy][INFO] - + preprocess energy efficiency: 68197130.910476 (samples/kWh) +[2024-10-01 19:54:02,194][device-isolation][INFO] - + Closing device(s) isolation process... +[2024-10-01 19:54:02,254][datasets][INFO] - PyTorch version 2.4.0 available. diff --git a/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-53-10/experiment_config.json b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-53-10/experiment_config.json new file mode 100644 index 0000000000000000000000000000000000000000..772aa352d839018f6caa9107fd24acf54618c6aa --- /dev/null +++ b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-53-10/experiment_config.json @@ -0,0 +1,107 @@ +{ + "experiment_name": "text_classification", + "backend": { + "name": "pytorch", + "version": "2.4.0", + "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend", + "task": "text-classification", + "model": "lvwerra/distilbert-imdb", + "processor": "lvwerra/distilbert-imdb", + "library": "transformers", + "device": "cuda", + "device_ids": "0", + "seed": 42, + "inter_op_num_threads": null, + "intra_op_num_threads": null, + "hub_kwargs": { + "revision": "main", + "force_download": false, + "local_files_only": false, + "trust_remote_code": true + }, + "no_weights": true, + "device_map": null, + "torch_dtype": null, + "amp_autocast": false, + "amp_dtype": null, + "eval_mode": true, + "to_bettertransformer": false, + "low_cpu_mem_usage": null, + "attn_implementation": null, + "cache_implementation": null, + "torch_compile": false, + "torch_compile_config": {}, + "quantization_scheme": null, + "quantization_config": {}, + "deepspeed_inference": false, + "deepspeed_inference_config": {}, + "peft_type": null, + "peft_config": {} + }, + "launcher": { + "name": "process", + "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher", + "device_isolation": true, + "device_isolation_action": "warn", + "start_method": "spawn" + }, + "benchmark": { + "name": "energy_star", + "_target_": "optimum_benchmark.benchmarks.energy_star.benchmark.EnergyStarBenchmark", + "dataset_name": "EnergyStarAI/text_classification", + "dataset_config": "", + "dataset_split": "train", + "num_samples": 1000, + "input_shapes": { + "batch_size": 1 + }, + "text_column_name": "text", + "truncation": true, + "max_length": -1, + "dataset_prefix1": "", + "dataset_prefix2": "", + "t5_task": "", + "image_column_name": "image", + "resize": false, + "question_column_name": "question", + "context_column_name": "context", + "sentence1_column_name": "sentence1", + "sentence2_column_name": "sentence2", + "audio_column_name": "audio", + "iterations": 10, + "warmup_runs": 10, + "energy": true, + "forward_kwargs": {}, + "generate_kwargs": {}, + "call_kwargs": {} + }, + "environment": { + "cpu": " AMD EPYC 7R32", + "cpu_count": 48, + "cpu_ram_mb": 200472.73984, + "system": "Linux", + "machine": "x86_64", + "platform": "Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35", + "processor": "x86_64", + "python_version": "3.9.20", + "gpu": [ + "NVIDIA A10G" + ], + "gpu_count": 1, + "gpu_vram_mb": 24146608128, + "optimum_benchmark_version": "0.2.0", + "optimum_benchmark_commit": null, + "transformers_version": "4.44.0", + "transformers_commit": null, + "accelerate_version": "0.33.0", + "accelerate_commit": null, + "diffusers_version": "0.30.0", + "diffusers_commit": null, + "optimum_version": null, + "optimum_commit": null, + "timm_version": null, + "timm_commit": null, + "peft_version": null, + "peft_commit": null + } +} \ No newline at end of file diff --git a/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-53-10/forward_codecarbon.json b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-53-10/forward_codecarbon.json new file mode 100644 index 0000000000000000000000000000000000000000..a4946256dc090648bad4545242fc44f77d2bd78c --- /dev/null +++ b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-53-10/forward_codecarbon.json @@ -0,0 +1,33 @@ +{ + "timestamp": "2024-10-01T19:54:01", + "project_name": "codecarbon", + "run_id": "e5fddf11-e7eb-4f78-a63d-2cbe3be7fc6f", + "duration": -1727749209.2494657, + "emissions": 8.739541697793405e-05, + "emissions_rate": 2.1659979687274563e-05, + "cpu_power": 42.5, + "gpu_power": 168.3900603830135, + "ram_power": 0.3452653884887696, + "cpu_energy": 4.765653881250826e-05, + "gpu_energy": 0.00018871431763800572, + "ram_energy": 3.868724755147697e-07, + "energy_consumed": 0.00023675772892602868, + "country_name": "United States", + "country_iso_code": "USA", + "region": "virginia", + "cloud_provider": "", + "cloud_region": "", + "os": "Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35", + "python_version": "3.9.20", + "codecarbon_version": "2.5.1", + "cpu_count": 48, + "cpu_model": "AMD EPYC 7R32", + "gpu_count": 1, + "gpu_model": "1 x NVIDIA A10G", + "longitude": -77.4903, + "latitude": 39.0469, + "ram_total_size": 186.7047882080078, + "tracking_mode": "process", + "on_cloud": "N", + "pue": 1.0 +} \ No newline at end of file diff --git a/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-53-10/preprocess_codecarbon.json b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-53-10/preprocess_codecarbon.json new file mode 100644 index 0000000000000000000000000000000000000000..dd06a82ba929ee828a1c70a24a627ddaaaf81545 --- /dev/null +++ b/text_classification/lvwerra/distilbert-imdb/2024-10-01-19-53-10/preprocess_codecarbon.json @@ -0,0 +1,33 @@ +{ + "timestamp": "2024-10-01T19:53:20", + "project_name": "codecarbon", + "run_id": "e5fddf11-e7eb-4f78-a63d-2cbe3be7fc6f", + "duration": -1727749212.8578753, + "emissions": 5.412755242963379e-06, + "emissions_rate": 1.2689263229337829e-05, + "cpu_power": 42.5, + "gpu_power": 80.80043302521536, + "ram_power": 0.2746009826660156, + "cpu_energy": 5.060345729154708e-06, + "gpu_energy": 9.570563211935834e-06, + "ram_energy": 3.2464527108885925e-08, + "energy_consumed": 1.4663373468199429e-05, + "country_name": "United States", + "country_iso_code": "USA", + "region": "virginia", + "cloud_provider": "", + "cloud_region": "", + "os": "Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35", + "python_version": "3.9.20", + "codecarbon_version": "2.5.1", + "cpu_count": 48, + "cpu_model": "AMD EPYC 7R32", + "gpu_count": 1, + "gpu_model": "1 x NVIDIA A10G", + "longitude": -77.4903, + "latitude": 39.0469, + "ram_total_size": 186.7047882080078, + "tracking_mode": "process", + "on_cloud": "N", + "pue": 1.0 +} \ No newline at end of file