Upload folder using huggingface_hub
Browse files- automatic_speech_recognition/jonatasgrosman/wav2vec2-large-xlsr-53-english/2024-12-04-15-35-05/.hydra/config.yaml +94 -0
- automatic_speech_recognition/jonatasgrosman/wav2vec2-large-xlsr-53-english/2024-12-04-15-35-05/.hydra/hydra.yaml +175 -0
- automatic_speech_recognition/jonatasgrosman/wav2vec2-large-xlsr-53-english/2024-12-04-15-35-05/.hydra/overrides.yaml +2 -0
- automatic_speech_recognition/jonatasgrosman/wav2vec2-large-xlsr-53-english/2024-12-04-15-35-05/cli.log +25 -0
- automatic_speech_recognition/jonatasgrosman/wav2vec2-large-xlsr-53-english/2024-12-04-15-35-05/error.log +93 -0
- automatic_speech_recognition/jonatasgrosman/wav2vec2-large-xlsr-53-english/2024-12-04-15-35-05/experiment_config.json +107 -0
- image_classification/timm/resnet50.a1_in1k/2024-12-04-17-13-03/.hydra/config.yaml +94 -0
- image_classification/timm/resnet50.a1_in1k/2024-12-04-17-13-03/.hydra/hydra.yaml +175 -0
- image_classification/timm/resnet50.a1_in1k/2024-12-04-17-13-03/.hydra/overrides.yaml +2 -0
- image_classification/timm/resnet50.a1_in1k/2024-12-04-17-13-03/cli.log +9 -0
- image_classification/timm/resnet50.a1_in1k/2024-12-04-17-13-03/error.log +32 -0
- image_classification/timm/resnet50.a1_in1k/2024-12-04-17-13-03/experiment_config.json +107 -0
- text_generation/bigscience/bloomz-560m/2024-12-04-16-38-24/.hydra/config.yaml +96 -0
- text_generation/bigscience/bloomz-560m/2024-12-04-16-38-24/.hydra/hydra.yaml +175 -0
- text_generation/bigscience/bloomz-560m/2024-12-04-16-38-24/.hydra/overrides.yaml +2 -0
- text_generation/bigscience/bloomz-560m/2024-12-04-16-38-24/benchmark_report.json +203 -0
- text_generation/bigscience/bloomz-560m/2024-12-04-16-38-24/cli.log +188 -0
- text_generation/bigscience/bloomz-560m/2024-12-04-16-38-24/error.log +0 -0
- text_generation/bigscience/bloomz-560m/2024-12-04-16-38-24/experiment_config.json +110 -0
- text_generation/bigscience/bloomz-560m/2024-12-04-16-38-24/generate_codecarbon.json +33 -0
- text_generation/bigscience/bloomz-560m/2024-12-04-16-38-24/prefill_codecarbon.json +33 -0
- text_generation/bigscience/bloomz-560m/2024-12-04-16-38-24/preprocess_codecarbon.json +33 -0
- text_generation/facebook/opt-1.3b/2024-12-04-15-35-29/.hydra/config.yaml +96 -0
- text_generation/facebook/opt-1.3b/2024-12-04-15-35-29/.hydra/hydra.yaml +175 -0
- text_generation/facebook/opt-1.3b/2024-12-04-15-35-29/.hydra/overrides.yaml +2 -0
- text_generation/facebook/opt-1.3b/2024-12-04-15-35-29/benchmark_report.json +203 -0
- text_generation/facebook/opt-1.3b/2024-12-04-15-35-29/cli.log +188 -0
- text_generation/facebook/opt-1.3b/2024-12-04-15-35-29/error.log +0 -0
- text_generation/facebook/opt-1.3b/2024-12-04-15-35-29/experiment_config.json +110 -0
- text_generation/facebook/opt-1.3b/2024-12-04-15-35-29/generate_codecarbon.json +33 -0
- text_generation/facebook/opt-1.3b/2024-12-04-15-35-29/prefill_codecarbon.json +33 -0
- text_generation/facebook/opt-1.3b/2024-12-04-15-35-29/preprocess_codecarbon.json +33 -0
- text_generation/openai-community/gpt2/2024-12-04-16-25-55/.hydra/config.yaml +96 -0
- text_generation/openai-community/gpt2/2024-12-04-16-25-55/.hydra/hydra.yaml +175 -0
- text_generation/openai-community/gpt2/2024-12-04-16-25-55/.hydra/overrides.yaml +2 -0
- text_generation/openai-community/gpt2/2024-12-04-16-25-55/benchmark_report.json +203 -0
- text_generation/openai-community/gpt2/2024-12-04-16-25-55/cli.log +188 -0
- text_generation/openai-community/gpt2/2024-12-04-16-25-55/error.log +0 -0
- text_generation/openai-community/gpt2/2024-12-04-16-25-55/experiment_config.json +110 -0
- text_generation/openai-community/gpt2/2024-12-04-16-25-55/generate_codecarbon.json +33 -0
- text_generation/openai-community/gpt2/2024-12-04-16-25-55/prefill_codecarbon.json +33 -0
- text_generation/openai-community/gpt2/2024-12-04-16-25-55/preprocess_codecarbon.json +33 -0
automatic_speech_recognition/jonatasgrosman/wav2vec2-large-xlsr-53-english/2024-12-04-15-35-05/.hydra/config.yaml
ADDED
|
@@ -0,0 +1,94 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
backend:
|
| 2 |
+
name: pytorch
|
| 3 |
+
version: 2.4.0
|
| 4 |
+
_target_: optimum_benchmark.backends.pytorch.backend.PyTorchBackend
|
| 5 |
+
task: automatic-speech-recognition
|
| 6 |
+
model: jonatasgrosman/wav2vec2-large-xlsr-53-english
|
| 7 |
+
processor: jonatasgrosman/wav2vec2-large-xlsr-53-english
|
| 8 |
+
library: null
|
| 9 |
+
device: cuda
|
| 10 |
+
device_ids: '0'
|
| 11 |
+
seed: 42
|
| 12 |
+
inter_op_num_threads: null
|
| 13 |
+
intra_op_num_threads: null
|
| 14 |
+
hub_kwargs: {}
|
| 15 |
+
no_weights: true
|
| 16 |
+
device_map: null
|
| 17 |
+
torch_dtype: null
|
| 18 |
+
amp_autocast: false
|
| 19 |
+
amp_dtype: null
|
| 20 |
+
eval_mode: true
|
| 21 |
+
to_bettertransformer: false
|
| 22 |
+
low_cpu_mem_usage: null
|
| 23 |
+
attn_implementation: null
|
| 24 |
+
cache_implementation: null
|
| 25 |
+
torch_compile: false
|
| 26 |
+
torch_compile_config: {}
|
| 27 |
+
quantization_scheme: null
|
| 28 |
+
quantization_config: {}
|
| 29 |
+
deepspeed_inference: false
|
| 30 |
+
deepspeed_inference_config: {}
|
| 31 |
+
peft_type: null
|
| 32 |
+
peft_config: {}
|
| 33 |
+
launcher:
|
| 34 |
+
name: process
|
| 35 |
+
_target_: optimum_benchmark.launchers.process.launcher.ProcessLauncher
|
| 36 |
+
device_isolation: true
|
| 37 |
+
device_isolation_action: warn
|
| 38 |
+
start_method: spawn
|
| 39 |
+
benchmark:
|
| 40 |
+
name: energy_star
|
| 41 |
+
_target_: optimum_benchmark.benchmarks.energy_star.benchmark.EnergyStarBenchmark
|
| 42 |
+
dataset_name: EnergyStarAI/ASR
|
| 43 |
+
dataset_config: ''
|
| 44 |
+
dataset_split: train
|
| 45 |
+
num_samples: 1000
|
| 46 |
+
input_shapes:
|
| 47 |
+
batch_size: 1
|
| 48 |
+
text_column_name: text
|
| 49 |
+
truncation: true
|
| 50 |
+
max_length: -1
|
| 51 |
+
dataset_prefix1: ''
|
| 52 |
+
dataset_prefix2: ''
|
| 53 |
+
t5_task: ''
|
| 54 |
+
image_column_name: image
|
| 55 |
+
resize: false
|
| 56 |
+
question_column_name: question
|
| 57 |
+
context_column_name: context
|
| 58 |
+
sentence1_column_name: sentence1
|
| 59 |
+
sentence2_column_name: sentence2
|
| 60 |
+
audio_column_name: audio
|
| 61 |
+
iterations: 10
|
| 62 |
+
warmup_runs: 10
|
| 63 |
+
energy: true
|
| 64 |
+
forward_kwargs: {}
|
| 65 |
+
generate_kwargs: {}
|
| 66 |
+
call_kwargs: {}
|
| 67 |
+
experiment_name: automatic_speech_recognition
|
| 68 |
+
environment:
|
| 69 |
+
cpu: ' AMD EPYC 7R32'
|
| 70 |
+
cpu_count: 48
|
| 71 |
+
cpu_ram_mb: 200472.73984
|
| 72 |
+
system: Linux
|
| 73 |
+
machine: x86_64
|
| 74 |
+
platform: Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35
|
| 75 |
+
processor: x86_64
|
| 76 |
+
python_version: 3.9.20
|
| 77 |
+
gpu:
|
| 78 |
+
- NVIDIA A10G
|
| 79 |
+
gpu_count: 1
|
| 80 |
+
gpu_vram_mb: 24146608128
|
| 81 |
+
optimum_benchmark_version: 0.2.0
|
| 82 |
+
optimum_benchmark_commit: null
|
| 83 |
+
transformers_version: 4.44.0
|
| 84 |
+
transformers_commit: null
|
| 85 |
+
accelerate_version: 0.33.0
|
| 86 |
+
accelerate_commit: null
|
| 87 |
+
diffusers_version: 0.30.0
|
| 88 |
+
diffusers_commit: null
|
| 89 |
+
optimum_version: null
|
| 90 |
+
optimum_commit: null
|
| 91 |
+
timm_version: null
|
| 92 |
+
timm_commit: null
|
| 93 |
+
peft_version: null
|
| 94 |
+
peft_commit: null
|
automatic_speech_recognition/jonatasgrosman/wav2vec2-large-xlsr-53-english/2024-12-04-15-35-05/.hydra/hydra.yaml
ADDED
|
@@ -0,0 +1,175 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
hydra:
|
| 2 |
+
run:
|
| 3 |
+
dir: /runs/automatic_speech_recognition/jonatasgrosman/wav2vec2-large-xlsr-53-english/2024-12-04-15-35-05
|
| 4 |
+
sweep:
|
| 5 |
+
dir: runs/${experiment_name}/${backend.model}/${now:%Y-%m-%d-%H-%M-%S}
|
| 6 |
+
subdir: ${hydra.job.num}
|
| 7 |
+
launcher:
|
| 8 |
+
_target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
|
| 9 |
+
sweeper:
|
| 10 |
+
_target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
|
| 11 |
+
max_batch_size: null
|
| 12 |
+
params: null
|
| 13 |
+
help:
|
| 14 |
+
app_name: ${hydra.job.name}
|
| 15 |
+
header: '${hydra.help.app_name} is powered by Hydra.
|
| 16 |
+
|
| 17 |
+
'
|
| 18 |
+
footer: 'Powered by Hydra (https://hydra.cc)
|
| 19 |
+
|
| 20 |
+
Use --hydra-help to view Hydra specific help
|
| 21 |
+
|
| 22 |
+
'
|
| 23 |
+
template: '${hydra.help.header}
|
| 24 |
+
|
| 25 |
+
== Configuration groups ==
|
| 26 |
+
|
| 27 |
+
Compose your configuration from those groups (group=option)
|
| 28 |
+
|
| 29 |
+
|
| 30 |
+
$APP_CONFIG_GROUPS
|
| 31 |
+
|
| 32 |
+
|
| 33 |
+
== Config ==
|
| 34 |
+
|
| 35 |
+
Override anything in the config (foo.bar=value)
|
| 36 |
+
|
| 37 |
+
|
| 38 |
+
$CONFIG
|
| 39 |
+
|
| 40 |
+
|
| 41 |
+
${hydra.help.footer}
|
| 42 |
+
|
| 43 |
+
'
|
| 44 |
+
hydra_help:
|
| 45 |
+
template: 'Hydra (${hydra.runtime.version})
|
| 46 |
+
|
| 47 |
+
See https://hydra.cc for more info.
|
| 48 |
+
|
| 49 |
+
|
| 50 |
+
== Flags ==
|
| 51 |
+
|
| 52 |
+
$FLAGS_HELP
|
| 53 |
+
|
| 54 |
+
|
| 55 |
+
== Configuration groups ==
|
| 56 |
+
|
| 57 |
+
Compose your configuration from those groups (For example, append hydra/job_logging=disabled
|
| 58 |
+
to command line)
|
| 59 |
+
|
| 60 |
+
|
| 61 |
+
$HYDRA_CONFIG_GROUPS
|
| 62 |
+
|
| 63 |
+
|
| 64 |
+
Use ''--cfg hydra'' to Show the Hydra config.
|
| 65 |
+
|
| 66 |
+
'
|
| 67 |
+
hydra_help: ???
|
| 68 |
+
hydra_logging:
|
| 69 |
+
version: 1
|
| 70 |
+
formatters:
|
| 71 |
+
colorlog:
|
| 72 |
+
(): colorlog.ColoredFormatter
|
| 73 |
+
format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s'
|
| 74 |
+
handlers:
|
| 75 |
+
console:
|
| 76 |
+
class: logging.StreamHandler
|
| 77 |
+
formatter: colorlog
|
| 78 |
+
stream: ext://sys.stdout
|
| 79 |
+
root:
|
| 80 |
+
level: INFO
|
| 81 |
+
handlers:
|
| 82 |
+
- console
|
| 83 |
+
disable_existing_loggers: false
|
| 84 |
+
job_logging:
|
| 85 |
+
version: 1
|
| 86 |
+
formatters:
|
| 87 |
+
simple:
|
| 88 |
+
format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
|
| 89 |
+
colorlog:
|
| 90 |
+
(): colorlog.ColoredFormatter
|
| 91 |
+
format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s]
|
| 92 |
+
- %(message)s'
|
| 93 |
+
log_colors:
|
| 94 |
+
DEBUG: purple
|
| 95 |
+
INFO: green
|
| 96 |
+
WARNING: yellow
|
| 97 |
+
ERROR: red
|
| 98 |
+
CRITICAL: red
|
| 99 |
+
handlers:
|
| 100 |
+
console:
|
| 101 |
+
class: logging.StreamHandler
|
| 102 |
+
formatter: colorlog
|
| 103 |
+
stream: ext://sys.stdout
|
| 104 |
+
file:
|
| 105 |
+
class: logging.FileHandler
|
| 106 |
+
formatter: simple
|
| 107 |
+
filename: ${hydra.job.name}.log
|
| 108 |
+
root:
|
| 109 |
+
level: INFO
|
| 110 |
+
handlers:
|
| 111 |
+
- console
|
| 112 |
+
- file
|
| 113 |
+
disable_existing_loggers: false
|
| 114 |
+
env: {}
|
| 115 |
+
mode: RUN
|
| 116 |
+
searchpath: []
|
| 117 |
+
callbacks: {}
|
| 118 |
+
output_subdir: .hydra
|
| 119 |
+
overrides:
|
| 120 |
+
hydra:
|
| 121 |
+
- hydra.run.dir=/runs/automatic_speech_recognition/jonatasgrosman/wav2vec2-large-xlsr-53-english/2024-12-04-15-35-05
|
| 122 |
+
- hydra.mode=RUN
|
| 123 |
+
task:
|
| 124 |
+
- backend.model=jonatasgrosman/wav2vec2-large-xlsr-53-english
|
| 125 |
+
- backend.processor=jonatasgrosman/wav2vec2-large-xlsr-53-english
|
| 126 |
+
job:
|
| 127 |
+
name: cli
|
| 128 |
+
chdir: true
|
| 129 |
+
override_dirname: backend.model=jonatasgrosman/wav2vec2-large-xlsr-53-english,backend.processor=jonatasgrosman/wav2vec2-large-xlsr-53-english
|
| 130 |
+
id: ???
|
| 131 |
+
num: ???
|
| 132 |
+
config_name: automatic_speech_recognition
|
| 133 |
+
env_set:
|
| 134 |
+
OVERRIDE_BENCHMARKS: '1'
|
| 135 |
+
env_copy: []
|
| 136 |
+
config:
|
| 137 |
+
override_dirname:
|
| 138 |
+
kv_sep: '='
|
| 139 |
+
item_sep: ','
|
| 140 |
+
exclude_keys: []
|
| 141 |
+
runtime:
|
| 142 |
+
version: 1.3.2
|
| 143 |
+
version_base: '1.3'
|
| 144 |
+
cwd: /
|
| 145 |
+
config_sources:
|
| 146 |
+
- path: hydra.conf
|
| 147 |
+
schema: pkg
|
| 148 |
+
provider: hydra
|
| 149 |
+
- path: optimum_benchmark
|
| 150 |
+
schema: pkg
|
| 151 |
+
provider: main
|
| 152 |
+
- path: hydra_plugins.hydra_colorlog.conf
|
| 153 |
+
schema: pkg
|
| 154 |
+
provider: hydra-colorlog
|
| 155 |
+
- path: /optimum-benchmark/examples/energy_star
|
| 156 |
+
schema: file
|
| 157 |
+
provider: command-line
|
| 158 |
+
- path: ''
|
| 159 |
+
schema: structured
|
| 160 |
+
provider: schema
|
| 161 |
+
output_dir: /runs/automatic_speech_recognition/jonatasgrosman/wav2vec2-large-xlsr-53-english/2024-12-04-15-35-05
|
| 162 |
+
choices:
|
| 163 |
+
benchmark: energy_star
|
| 164 |
+
launcher: process
|
| 165 |
+
backend: pytorch
|
| 166 |
+
hydra/env: default
|
| 167 |
+
hydra/callbacks: null
|
| 168 |
+
hydra/job_logging: colorlog
|
| 169 |
+
hydra/hydra_logging: colorlog
|
| 170 |
+
hydra/hydra_help: default
|
| 171 |
+
hydra/help: default
|
| 172 |
+
hydra/sweeper: basic
|
| 173 |
+
hydra/launcher: basic
|
| 174 |
+
hydra/output: default
|
| 175 |
+
verbose: false
|
automatic_speech_recognition/jonatasgrosman/wav2vec2-large-xlsr-53-english/2024-12-04-15-35-05/.hydra/overrides.yaml
ADDED
|
@@ -0,0 +1,2 @@
|
|
|
|
|
|
|
|
|
|
| 1 |
+
- backend.model=jonatasgrosman/wav2vec2-large-xlsr-53-english
|
| 2 |
+
- backend.processor=jonatasgrosman/wav2vec2-large-xlsr-53-english
|
automatic_speech_recognition/jonatasgrosman/wav2vec2-large-xlsr-53-english/2024-12-04-15-35-05/cli.log
ADDED
|
@@ -0,0 +1,25 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
[2024-12-04 15:35:08,566][launcher][INFO] - ََAllocating process launcher
|
| 2 |
+
[2024-12-04 15:35:08,566][process][INFO] - + Setting multiprocessing start method to spawn.
|
| 3 |
+
[2024-12-04 15:35:08,576][device-isolation][INFO] - + Launched device(s) isolation process 181
|
| 4 |
+
[2024-12-04 15:35:08,576][device-isolation][INFO] - + Isolating device(s) [0]
|
| 5 |
+
[2024-12-04 15:35:08,581][process][INFO] - + Launched benchmark in isolated process 182.
|
| 6 |
+
[PROC-0][2024-12-04 15:35:11,345][datasets][INFO] - PyTorch version 2.4.0 available.
|
| 7 |
+
[PROC-0][2024-12-04 15:35:12,266][backend][INFO] - َAllocating pytorch backend
|
| 8 |
+
[PROC-0][2024-12-04 15:35:12,266][backend][INFO] - + Setting random seed to 42
|
| 9 |
+
[PROC-0][2024-12-04 15:35:12,688][pytorch][INFO] - + Using AutoModel class AutoModelForCTC
|
| 10 |
+
[PROC-0][2024-12-04 15:35:12,689][pytorch][INFO] - + Creating backend temporary directory
|
| 11 |
+
[PROC-0][2024-12-04 15:35:12,689][pytorch][INFO] - + Loading model with random weights
|
| 12 |
+
[PROC-0][2024-12-04 15:35:12,689][pytorch][INFO] - + Creating no weights model
|
| 13 |
+
[PROC-0][2024-12-04 15:35:12,689][pytorch][INFO] - + Creating no weights model directory
|
| 14 |
+
[PROC-0][2024-12-04 15:35:12,689][pytorch][INFO] - + Creating no weights model state dict
|
| 15 |
+
[PROC-0][2024-12-04 15:35:12,691][pytorch][INFO] - + Saving no weights model safetensors
|
| 16 |
+
[PROC-0][2024-12-04 15:35:12,692][pytorch][INFO] - + Saving no weights model pretrained config
|
| 17 |
+
[PROC-0][2024-12-04 15:35:12,693][pytorch][INFO] - + Loading no weights AutoModel
|
| 18 |
+
[PROC-0][2024-12-04 15:35:12,693][pytorch][INFO] - + Loading model directly on device: cuda
|
| 19 |
+
[PROC-0][2024-12-04 15:35:12,993][pytorch][INFO] - + Turning on model's eval mode
|
| 20 |
+
[PROC-0][2024-12-04 15:35:12,999][benchmark][INFO] - Allocating energy_star benchmark
|
| 21 |
+
[PROC-0][2024-12-04 15:35:13,000][energy_star][INFO] - + Loading raw dataset
|
| 22 |
+
[PROC-0][2024-12-04 15:35:18,026][energy_star][INFO] - + Initializing Inference report
|
| 23 |
+
[PROC-0][2024-12-04 15:35:18,026][energy][INFO] - + Tracking GPU energy on devices [0]
|
| 24 |
+
[PROC-0][2024-12-04 15:35:22,200][energy_star][INFO] - + Preprocessing dataset
|
| 25 |
+
[2024-12-04 15:35:28,653][experiment][ERROR] - Error during experiment
|
automatic_speech_recognition/jonatasgrosman/wav2vec2-large-xlsr-53-english/2024-12-04-15-35-05/error.log
ADDED
|
@@ -0,0 +1,93 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
/opt/conda/lib/python3.9/site-packages/transformers/tokenization_utils_base.py:1601: FutureWarning: `clean_up_tokenization_spaces` was not set. It will be set to `True` by default. This behavior will be depracted in transformers v4.45, and will be then set to `False` by default. For more details check this issue: https://github.com/huggingface/transformers/issues/31884
|
| 2 |
+
warnings.warn(
|
| 3 |
+
|
| 4 |
+
|
| 5 |
+
|
| 6 |
+
|
| 7 |
+
|
| 8 |
+
|
| 9 |
+
|
| 10 |
+
|
| 11 |
+
|
| 12 |
+
|
| 13 |
+
|
| 14 |
+
|
| 15 |
+
|
| 16 |
+
|
| 17 |
+
|
| 18 |
+
|
| 19 |
+
|
| 20 |
+
|
| 21 |
+
|
| 22 |
+
|
| 23 |
+
|
| 24 |
+
|
| 25 |
+
|
| 26 |
+
|
| 27 |
+
|
| 28 |
+
|
| 29 |
+
|
| 30 |
+
|
| 31 |
+
|
| 32 |
+
|
| 33 |
+
|
| 34 |
+
|
| 35 |
+
|
| 36 |
+
[codecarbon INFO @ 15:35:18] [setup] RAM Tracking...
|
| 37 |
+
[codecarbon INFO @ 15:35:18] [setup] GPU Tracking...
|
| 38 |
+
[codecarbon INFO @ 15:35:18] Tracking Nvidia GPU via pynvml
|
| 39 |
+
[codecarbon DEBUG @ 15:35:18] GPU available. Starting setup
|
| 40 |
+
[codecarbon INFO @ 15:35:18] [setup] CPU Tracking...
|
| 41 |
+
[codecarbon DEBUG @ 15:35:18] Not using PowerGadget, an exception occurred while instantiating IntelPowerGadget : Platform not supported by Intel Power Gadget
|
| 42 |
+
[codecarbon DEBUG @ 15:35:18] Not using the RAPL interface, an exception occurred while instantiating IntelRAPL : Intel RAPL files not found at /sys/class/powercap/intel-rapl on linux
|
| 43 |
+
[codecarbon DEBUG @ 15:35:18] Not using PowerMetrics, an exception occurred while instantiating Powermetrics : Platform not supported by Powermetrics
|
| 44 |
+
[codecarbon WARNING @ 15:35:18] No CPU tracking mode found. Falling back on CPU constant mode.
|
| 45 |
+
[codecarbon WARNING @ 15:35:19] We saw that you have a AMD EPYC 7R32 but we don't know it. Please contact us.
|
| 46 |
+
[codecarbon INFO @ 15:35:19] CPU Model on constant consumption mode: AMD EPYC 7R32
|
| 47 |
+
[codecarbon INFO @ 15:35:19] >>> Tracker's metadata:
|
| 48 |
+
[codecarbon INFO @ 15:35:19] Platform system: Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35
|
| 49 |
+
[codecarbon INFO @ 15:35:19] Python version: 3.9.20
|
| 50 |
+
[codecarbon INFO @ 15:35:19] CodeCarbon version: 2.5.1
|
| 51 |
+
[codecarbon INFO @ 15:35:19] Available RAM : 186.705 GB
|
| 52 |
+
[codecarbon INFO @ 15:35:19] CPU count: 48
|
| 53 |
+
[codecarbon INFO @ 15:35:19] CPU model: AMD EPYC 7R32
|
| 54 |
+
[codecarbon INFO @ 15:35:19] GPU count: 1
|
| 55 |
+
[codecarbon INFO @ 15:35:19] GPU model: 1 x NVIDIA A10G
|
| 56 |
+
[codecarbon DEBUG @ 15:35:20] Not running on AWS
|
| 57 |
+
[codecarbon DEBUG @ 15:35:21] Not running on Azure
|
| 58 |
+
[codecarbon DEBUG @ 15:35:22] Not running on GCP
|
| 59 |
+
[codecarbon INFO @ 15:35:22] Saving emissions data to file /runs/automatic_speech_recognition/jonatasgrosman/wav2vec2-large-xlsr-53-english/2024-12-04-15-35-05/codecarbon.csv
|
| 60 |
+
[codecarbon DEBUG @ 15:35:22] EmissionsData(timestamp='2024-12-04T15:35:22', project_name='codecarbon', run_id='4002c2cc-d51d-4713-bce1-ee1ad4edb5ec', duration=0.0021653330040862784, emissions=0.0, emissions_rate=0.0, cpu_power=0.0, gpu_power=0.0, ram_power=0.0, cpu_energy=0, gpu_energy=0, ram_energy=0, energy_consumed=0, country_name='United States', country_iso_code='USA', region='virginia', cloud_provider='', cloud_region='', os='Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35', python_version='3.9.20', codecarbon_version='2.5.1', cpu_count=48, cpu_model='AMD EPYC 7R32', gpu_count=1, gpu_model='1 x NVIDIA A10G', longitude=-77.4903, latitude=39.0469, ram_total_size=186.7047882080078, tracking_mode='process', on_cloud='N', pue=1.0)
|
| 61 |
+
|
| 62 |
+
Error executing job with overrides: ['backend.model=jonatasgrosman/wav2vec2-large-xlsr-53-english', 'backend.processor=jonatasgrosman/wav2vec2-large-xlsr-53-english']
|
| 63 |
+
Traceback (most recent call last):
|
| 64 |
+
File "/optimum-benchmark/optimum_benchmark/cli.py", line 65, in benchmark_cli
|
| 65 |
+
benchmark_report: BenchmarkReport = launch(experiment_config=experiment_config)
|
| 66 |
+
File "/optimum-benchmark/optimum_benchmark/experiment.py", line 102, in launch
|
| 67 |
+
raise error
|
| 68 |
+
File "/optimum-benchmark/optimum_benchmark/experiment.py", line 90, in launch
|
| 69 |
+
report = launcher.launch(run, experiment_config.benchmark, experiment_config.backend)
|
| 70 |
+
File "/optimum-benchmark/optimum_benchmark/launchers/process/launcher.py", line 47, in launch
|
| 71 |
+
while not process_context.join():
|
| 72 |
+
File "/opt/conda/lib/python3.9/site-packages/torch/multiprocessing/spawn.py", line 189, in join
|
| 73 |
+
raise ProcessRaisedException(msg, error_index, failed_process.pid)
|
| 74 |
+
torch.multiprocessing.spawn.ProcessRaisedException:
|
| 75 |
+
|
| 76 |
+
-- Process 0 terminated with the following error:
|
| 77 |
+
Traceback (most recent call last):
|
| 78 |
+
File "/opt/conda/lib/python3.9/site-packages/torch/multiprocessing/spawn.py", line 76, in _wrap
|
| 79 |
+
fn(i, *args)
|
| 80 |
+
File "/optimum-benchmark/optimum_benchmark/launchers/process/launcher.py", line 63, in entrypoint
|
| 81 |
+
worker_output = worker(*worker_args)
|
| 82 |
+
File "/optimum-benchmark/optimum_benchmark/experiment.py", line 62, in run
|
| 83 |
+
benchmark.run(backend)
|
| 84 |
+
File "/optimum-benchmark/optimum_benchmark/benchmarks/energy_star/benchmark.py", line 122, in run
|
| 85 |
+
self.dataset = preprocess(
|
| 86 |
+
File "/optimum-benchmark/optimum_benchmark/benchmarks/energy_star/preprocessing_utils.py", line 28, in preprocess
|
| 87 |
+
return task_to_preprocessing[task](dataset, config, preprocessor, pretrained_config)
|
| 88 |
+
File "/optimum-benchmark/optimum_benchmark/benchmarks/energy_star/preprocessing_utils.py", line 360, in automatic_speech_recognition_preprocessing
|
| 89 |
+
if getattr(processor.tokenizer, "pad_token", None) is None:
|
| 90 |
+
AttributeError: 'Wav2Vec2CTCTokenizer' object has no attribute 'tokenizer'
|
| 91 |
+
|
| 92 |
+
|
| 93 |
+
Set the environment variable HYDRA_FULL_ERROR=1 for a complete stack trace.
|
automatic_speech_recognition/jonatasgrosman/wav2vec2-large-xlsr-53-english/2024-12-04-15-35-05/experiment_config.json
ADDED
|
@@ -0,0 +1,107 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"experiment_name": "automatic_speech_recognition",
|
| 3 |
+
"backend": {
|
| 4 |
+
"name": "pytorch",
|
| 5 |
+
"version": "2.4.0",
|
| 6 |
+
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
|
| 7 |
+
"task": "automatic-speech-recognition",
|
| 8 |
+
"model": "jonatasgrosman/wav2vec2-large-xlsr-53-english",
|
| 9 |
+
"processor": "jonatasgrosman/wav2vec2-large-xlsr-53-english",
|
| 10 |
+
"library": "transformers",
|
| 11 |
+
"device": "cuda",
|
| 12 |
+
"device_ids": "0",
|
| 13 |
+
"seed": 42,
|
| 14 |
+
"inter_op_num_threads": null,
|
| 15 |
+
"intra_op_num_threads": null,
|
| 16 |
+
"hub_kwargs": {
|
| 17 |
+
"revision": "main",
|
| 18 |
+
"force_download": false,
|
| 19 |
+
"local_files_only": false,
|
| 20 |
+
"trust_remote_code": true
|
| 21 |
+
},
|
| 22 |
+
"no_weights": true,
|
| 23 |
+
"device_map": null,
|
| 24 |
+
"torch_dtype": null,
|
| 25 |
+
"amp_autocast": false,
|
| 26 |
+
"amp_dtype": null,
|
| 27 |
+
"eval_mode": true,
|
| 28 |
+
"to_bettertransformer": false,
|
| 29 |
+
"low_cpu_mem_usage": null,
|
| 30 |
+
"attn_implementation": null,
|
| 31 |
+
"cache_implementation": null,
|
| 32 |
+
"torch_compile": false,
|
| 33 |
+
"torch_compile_config": {},
|
| 34 |
+
"quantization_scheme": null,
|
| 35 |
+
"quantization_config": {},
|
| 36 |
+
"deepspeed_inference": false,
|
| 37 |
+
"deepspeed_inference_config": {},
|
| 38 |
+
"peft_type": null,
|
| 39 |
+
"peft_config": {}
|
| 40 |
+
},
|
| 41 |
+
"launcher": {
|
| 42 |
+
"name": "process",
|
| 43 |
+
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
|
| 44 |
+
"device_isolation": true,
|
| 45 |
+
"device_isolation_action": "warn",
|
| 46 |
+
"start_method": "spawn"
|
| 47 |
+
},
|
| 48 |
+
"benchmark": {
|
| 49 |
+
"name": "energy_star",
|
| 50 |
+
"_target_": "optimum_benchmark.benchmarks.energy_star.benchmark.EnergyStarBenchmark",
|
| 51 |
+
"dataset_name": "EnergyStarAI/ASR",
|
| 52 |
+
"dataset_config": "",
|
| 53 |
+
"dataset_split": "train",
|
| 54 |
+
"num_samples": 1000,
|
| 55 |
+
"input_shapes": {
|
| 56 |
+
"batch_size": 1
|
| 57 |
+
},
|
| 58 |
+
"text_column_name": "text",
|
| 59 |
+
"truncation": true,
|
| 60 |
+
"max_length": -1,
|
| 61 |
+
"dataset_prefix1": "",
|
| 62 |
+
"dataset_prefix2": "",
|
| 63 |
+
"t5_task": "",
|
| 64 |
+
"image_column_name": "image",
|
| 65 |
+
"resize": false,
|
| 66 |
+
"question_column_name": "question",
|
| 67 |
+
"context_column_name": "context",
|
| 68 |
+
"sentence1_column_name": "sentence1",
|
| 69 |
+
"sentence2_column_name": "sentence2",
|
| 70 |
+
"audio_column_name": "audio",
|
| 71 |
+
"iterations": 10,
|
| 72 |
+
"warmup_runs": 10,
|
| 73 |
+
"energy": true,
|
| 74 |
+
"forward_kwargs": {},
|
| 75 |
+
"generate_kwargs": {},
|
| 76 |
+
"call_kwargs": {}
|
| 77 |
+
},
|
| 78 |
+
"environment": {
|
| 79 |
+
"cpu": " AMD EPYC 7R32",
|
| 80 |
+
"cpu_count": 48,
|
| 81 |
+
"cpu_ram_mb": 200472.73984,
|
| 82 |
+
"system": "Linux",
|
| 83 |
+
"machine": "x86_64",
|
| 84 |
+
"platform": "Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35",
|
| 85 |
+
"processor": "x86_64",
|
| 86 |
+
"python_version": "3.9.20",
|
| 87 |
+
"gpu": [
|
| 88 |
+
"NVIDIA A10G"
|
| 89 |
+
],
|
| 90 |
+
"gpu_count": 1,
|
| 91 |
+
"gpu_vram_mb": 24146608128,
|
| 92 |
+
"optimum_benchmark_version": "0.2.0",
|
| 93 |
+
"optimum_benchmark_commit": null,
|
| 94 |
+
"transformers_version": "4.44.0",
|
| 95 |
+
"transformers_commit": null,
|
| 96 |
+
"accelerate_version": "0.33.0",
|
| 97 |
+
"accelerate_commit": null,
|
| 98 |
+
"diffusers_version": "0.30.0",
|
| 99 |
+
"diffusers_commit": null,
|
| 100 |
+
"optimum_version": null,
|
| 101 |
+
"optimum_commit": null,
|
| 102 |
+
"timm_version": null,
|
| 103 |
+
"timm_commit": null,
|
| 104 |
+
"peft_version": null,
|
| 105 |
+
"peft_commit": null
|
| 106 |
+
}
|
| 107 |
+
}
|
image_classification/timm/resnet50.a1_in1k/2024-12-04-17-13-03/.hydra/config.yaml
ADDED
|
@@ -0,0 +1,94 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
backend:
|
| 2 |
+
name: pytorch
|
| 3 |
+
version: 2.4.0
|
| 4 |
+
_target_: optimum_benchmark.backends.pytorch.backend.PyTorchBackend
|
| 5 |
+
task: image-classification
|
| 6 |
+
model: timm/resnet50.a1_in1k
|
| 7 |
+
processor: timm/resnet50.a1_in1k
|
| 8 |
+
library: null
|
| 9 |
+
device: cuda
|
| 10 |
+
device_ids: '0'
|
| 11 |
+
seed: 42
|
| 12 |
+
inter_op_num_threads: null
|
| 13 |
+
intra_op_num_threads: null
|
| 14 |
+
hub_kwargs: {}
|
| 15 |
+
no_weights: true
|
| 16 |
+
device_map: null
|
| 17 |
+
torch_dtype: null
|
| 18 |
+
amp_autocast: false
|
| 19 |
+
amp_dtype: null
|
| 20 |
+
eval_mode: true
|
| 21 |
+
to_bettertransformer: false
|
| 22 |
+
low_cpu_mem_usage: null
|
| 23 |
+
attn_implementation: null
|
| 24 |
+
cache_implementation: null
|
| 25 |
+
torch_compile: false
|
| 26 |
+
torch_compile_config: {}
|
| 27 |
+
quantization_scheme: null
|
| 28 |
+
quantization_config: {}
|
| 29 |
+
deepspeed_inference: false
|
| 30 |
+
deepspeed_inference_config: {}
|
| 31 |
+
peft_type: null
|
| 32 |
+
peft_config: {}
|
| 33 |
+
launcher:
|
| 34 |
+
name: process
|
| 35 |
+
_target_: optimum_benchmark.launchers.process.launcher.ProcessLauncher
|
| 36 |
+
device_isolation: true
|
| 37 |
+
device_isolation_action: warn
|
| 38 |
+
start_method: spawn
|
| 39 |
+
benchmark:
|
| 40 |
+
name: energy_star
|
| 41 |
+
_target_: optimum_benchmark.benchmarks.energy_star.benchmark.EnergyStarBenchmark
|
| 42 |
+
dataset_name: EnergyStarAI/image_classification
|
| 43 |
+
dataset_config: ''
|
| 44 |
+
dataset_split: train
|
| 45 |
+
num_samples: 1000
|
| 46 |
+
input_shapes:
|
| 47 |
+
batch_size: 1
|
| 48 |
+
text_column_name: text
|
| 49 |
+
truncation: true
|
| 50 |
+
max_length: -1
|
| 51 |
+
dataset_prefix1: ''
|
| 52 |
+
dataset_prefix2: ''
|
| 53 |
+
t5_task: ''
|
| 54 |
+
image_column_name: image
|
| 55 |
+
resize: false
|
| 56 |
+
question_column_name: question
|
| 57 |
+
context_column_name: context
|
| 58 |
+
sentence1_column_name: sentence1
|
| 59 |
+
sentence2_column_name: sentence2
|
| 60 |
+
audio_column_name: audio
|
| 61 |
+
iterations: 10
|
| 62 |
+
warmup_runs: 10
|
| 63 |
+
energy: true
|
| 64 |
+
forward_kwargs: {}
|
| 65 |
+
generate_kwargs: {}
|
| 66 |
+
call_kwargs: {}
|
| 67 |
+
experiment_name: image_classification
|
| 68 |
+
environment:
|
| 69 |
+
cpu: ' AMD EPYC 7R32'
|
| 70 |
+
cpu_count: 48
|
| 71 |
+
cpu_ram_mb: 200472.73984
|
| 72 |
+
system: Linux
|
| 73 |
+
machine: x86_64
|
| 74 |
+
platform: Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35
|
| 75 |
+
processor: x86_64
|
| 76 |
+
python_version: 3.9.20
|
| 77 |
+
gpu:
|
| 78 |
+
- NVIDIA A10G
|
| 79 |
+
gpu_count: 1
|
| 80 |
+
gpu_vram_mb: 24146608128
|
| 81 |
+
optimum_benchmark_version: 0.2.0
|
| 82 |
+
optimum_benchmark_commit: null
|
| 83 |
+
transformers_version: 4.44.0
|
| 84 |
+
transformers_commit: null
|
| 85 |
+
accelerate_version: 0.33.0
|
| 86 |
+
accelerate_commit: null
|
| 87 |
+
diffusers_version: 0.30.0
|
| 88 |
+
diffusers_commit: null
|
| 89 |
+
optimum_version: null
|
| 90 |
+
optimum_commit: null
|
| 91 |
+
timm_version: null
|
| 92 |
+
timm_commit: null
|
| 93 |
+
peft_version: null
|
| 94 |
+
peft_commit: null
|
image_classification/timm/resnet50.a1_in1k/2024-12-04-17-13-03/.hydra/hydra.yaml
ADDED
|
@@ -0,0 +1,175 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
hydra:
|
| 2 |
+
run:
|
| 3 |
+
dir: /runs/image_classification/timm/resnet50.a1_in1k/2024-12-04-17-13-03
|
| 4 |
+
sweep:
|
| 5 |
+
dir: sweeps/${experiment_name}/${backend.model}/${now:%Y-%m-%d-%H-%M-%S}
|
| 6 |
+
subdir: ${hydra.job.num}
|
| 7 |
+
launcher:
|
| 8 |
+
_target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
|
| 9 |
+
sweeper:
|
| 10 |
+
_target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
|
| 11 |
+
max_batch_size: null
|
| 12 |
+
params: null
|
| 13 |
+
help:
|
| 14 |
+
app_name: ${hydra.job.name}
|
| 15 |
+
header: '${hydra.help.app_name} is powered by Hydra.
|
| 16 |
+
|
| 17 |
+
'
|
| 18 |
+
footer: 'Powered by Hydra (https://hydra.cc)
|
| 19 |
+
|
| 20 |
+
Use --hydra-help to view Hydra specific help
|
| 21 |
+
|
| 22 |
+
'
|
| 23 |
+
template: '${hydra.help.header}
|
| 24 |
+
|
| 25 |
+
== Configuration groups ==
|
| 26 |
+
|
| 27 |
+
Compose your configuration from those groups (group=option)
|
| 28 |
+
|
| 29 |
+
|
| 30 |
+
$APP_CONFIG_GROUPS
|
| 31 |
+
|
| 32 |
+
|
| 33 |
+
== Config ==
|
| 34 |
+
|
| 35 |
+
Override anything in the config (foo.bar=value)
|
| 36 |
+
|
| 37 |
+
|
| 38 |
+
$CONFIG
|
| 39 |
+
|
| 40 |
+
|
| 41 |
+
${hydra.help.footer}
|
| 42 |
+
|
| 43 |
+
'
|
| 44 |
+
hydra_help:
|
| 45 |
+
template: 'Hydra (${hydra.runtime.version})
|
| 46 |
+
|
| 47 |
+
See https://hydra.cc for more info.
|
| 48 |
+
|
| 49 |
+
|
| 50 |
+
== Flags ==
|
| 51 |
+
|
| 52 |
+
$FLAGS_HELP
|
| 53 |
+
|
| 54 |
+
|
| 55 |
+
== Configuration groups ==
|
| 56 |
+
|
| 57 |
+
Compose your configuration from those groups (For example, append hydra/job_logging=disabled
|
| 58 |
+
to command line)
|
| 59 |
+
|
| 60 |
+
|
| 61 |
+
$HYDRA_CONFIG_GROUPS
|
| 62 |
+
|
| 63 |
+
|
| 64 |
+
Use ''--cfg hydra'' to Show the Hydra config.
|
| 65 |
+
|
| 66 |
+
'
|
| 67 |
+
hydra_help: ???
|
| 68 |
+
hydra_logging:
|
| 69 |
+
version: 1
|
| 70 |
+
formatters:
|
| 71 |
+
colorlog:
|
| 72 |
+
(): colorlog.ColoredFormatter
|
| 73 |
+
format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s'
|
| 74 |
+
handlers:
|
| 75 |
+
console:
|
| 76 |
+
class: logging.StreamHandler
|
| 77 |
+
formatter: colorlog
|
| 78 |
+
stream: ext://sys.stdout
|
| 79 |
+
root:
|
| 80 |
+
level: INFO
|
| 81 |
+
handlers:
|
| 82 |
+
- console
|
| 83 |
+
disable_existing_loggers: false
|
| 84 |
+
job_logging:
|
| 85 |
+
version: 1
|
| 86 |
+
formatters:
|
| 87 |
+
simple:
|
| 88 |
+
format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
|
| 89 |
+
colorlog:
|
| 90 |
+
(): colorlog.ColoredFormatter
|
| 91 |
+
format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s]
|
| 92 |
+
- %(message)s'
|
| 93 |
+
log_colors:
|
| 94 |
+
DEBUG: purple
|
| 95 |
+
INFO: green
|
| 96 |
+
WARNING: yellow
|
| 97 |
+
ERROR: red
|
| 98 |
+
CRITICAL: red
|
| 99 |
+
handlers:
|
| 100 |
+
console:
|
| 101 |
+
class: logging.StreamHandler
|
| 102 |
+
formatter: colorlog
|
| 103 |
+
stream: ext://sys.stdout
|
| 104 |
+
file:
|
| 105 |
+
class: logging.FileHandler
|
| 106 |
+
formatter: simple
|
| 107 |
+
filename: ${hydra.job.name}.log
|
| 108 |
+
root:
|
| 109 |
+
level: INFO
|
| 110 |
+
handlers:
|
| 111 |
+
- console
|
| 112 |
+
- file
|
| 113 |
+
disable_existing_loggers: false
|
| 114 |
+
env: {}
|
| 115 |
+
mode: RUN
|
| 116 |
+
searchpath: []
|
| 117 |
+
callbacks: {}
|
| 118 |
+
output_subdir: .hydra
|
| 119 |
+
overrides:
|
| 120 |
+
hydra:
|
| 121 |
+
- hydra.run.dir=/runs/image_classification/timm/resnet50.a1_in1k/2024-12-04-17-13-03
|
| 122 |
+
- hydra.mode=RUN
|
| 123 |
+
task:
|
| 124 |
+
- backend.model=timm/resnet50.a1_in1k
|
| 125 |
+
- backend.processor=timm/resnet50.a1_in1k
|
| 126 |
+
job:
|
| 127 |
+
name: cli
|
| 128 |
+
chdir: true
|
| 129 |
+
override_dirname: backend.model=timm/resnet50.a1_in1k,backend.processor=timm/resnet50.a1_in1k
|
| 130 |
+
id: ???
|
| 131 |
+
num: ???
|
| 132 |
+
config_name: image_classification
|
| 133 |
+
env_set:
|
| 134 |
+
OVERRIDE_BENCHMARKS: '1'
|
| 135 |
+
env_copy: []
|
| 136 |
+
config:
|
| 137 |
+
override_dirname:
|
| 138 |
+
kv_sep: '='
|
| 139 |
+
item_sep: ','
|
| 140 |
+
exclude_keys: []
|
| 141 |
+
runtime:
|
| 142 |
+
version: 1.3.2
|
| 143 |
+
version_base: '1.3'
|
| 144 |
+
cwd: /
|
| 145 |
+
config_sources:
|
| 146 |
+
- path: hydra.conf
|
| 147 |
+
schema: pkg
|
| 148 |
+
provider: hydra
|
| 149 |
+
- path: optimum_benchmark
|
| 150 |
+
schema: pkg
|
| 151 |
+
provider: main
|
| 152 |
+
- path: hydra_plugins.hydra_colorlog.conf
|
| 153 |
+
schema: pkg
|
| 154 |
+
provider: hydra-colorlog
|
| 155 |
+
- path: /optimum-benchmark/examples/energy_star
|
| 156 |
+
schema: file
|
| 157 |
+
provider: command-line
|
| 158 |
+
- path: ''
|
| 159 |
+
schema: structured
|
| 160 |
+
provider: schema
|
| 161 |
+
output_dir: /runs/image_classification/timm/resnet50.a1_in1k/2024-12-04-17-13-03
|
| 162 |
+
choices:
|
| 163 |
+
benchmark: energy_star
|
| 164 |
+
launcher: process
|
| 165 |
+
backend: pytorch
|
| 166 |
+
hydra/env: default
|
| 167 |
+
hydra/callbacks: null
|
| 168 |
+
hydra/job_logging: colorlog
|
| 169 |
+
hydra/hydra_logging: colorlog
|
| 170 |
+
hydra/hydra_help: default
|
| 171 |
+
hydra/help: default
|
| 172 |
+
hydra/sweeper: basic
|
| 173 |
+
hydra/launcher: basic
|
| 174 |
+
hydra/output: default
|
| 175 |
+
verbose: false
|
image_classification/timm/resnet50.a1_in1k/2024-12-04-17-13-03/.hydra/overrides.yaml
ADDED
|
@@ -0,0 +1,2 @@
|
|
|
|
|
|
|
|
|
|
| 1 |
+
- backend.model=timm/resnet50.a1_in1k
|
| 2 |
+
- backend.processor=timm/resnet50.a1_in1k
|
image_classification/timm/resnet50.a1_in1k/2024-12-04-17-13-03/cli.log
ADDED
|
@@ -0,0 +1,9 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
[2024-12-04 17:13:06,484][launcher][INFO] - ََAllocating process launcher
|
| 2 |
+
[2024-12-04 17:13:06,484][process][INFO] - + Setting multiprocessing start method to spawn.
|
| 3 |
+
[2024-12-04 17:13:06,497][device-isolation][INFO] - + Launched device(s) isolation process 986
|
| 4 |
+
[2024-12-04 17:13:06,498][device-isolation][INFO] - + Isolating device(s) [0]
|
| 5 |
+
[2024-12-04 17:13:06,504][process][INFO] - + Launched benchmark in isolated process 987.
|
| 6 |
+
[PROC-0][2024-12-04 17:13:09,182][datasets][INFO] - PyTorch version 2.4.0 available.
|
| 7 |
+
[PROC-0][2024-12-04 17:13:10,172][backend][INFO] - َAllocating pytorch backend
|
| 8 |
+
[PROC-0][2024-12-04 17:13:10,172][backend][INFO] - + Setting random seed to 42
|
| 9 |
+
[2024-12-04 17:13:10,706][experiment][ERROR] - Error during experiment
|
image_classification/timm/resnet50.a1_in1k/2024-12-04-17-13-03/error.log
ADDED
|
@@ -0,0 +1,32 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
Error executing job with overrides: ['backend.model=timm/resnet50.a1_in1k', 'backend.processor=timm/resnet50.a1_in1k']
|
| 2 |
+
Traceback (most recent call last):
|
| 3 |
+
File "/optimum-benchmark/optimum_benchmark/cli.py", line 65, in benchmark_cli
|
| 4 |
+
benchmark_report: BenchmarkReport = launch(experiment_config=experiment_config)
|
| 5 |
+
File "/optimum-benchmark/optimum_benchmark/experiment.py", line 102, in launch
|
| 6 |
+
raise error
|
| 7 |
+
File "/optimum-benchmark/optimum_benchmark/experiment.py", line 90, in launch
|
| 8 |
+
report = launcher.launch(run, experiment_config.benchmark, experiment_config.backend)
|
| 9 |
+
File "/optimum-benchmark/optimum_benchmark/launchers/process/launcher.py", line 47, in launch
|
| 10 |
+
while not process_context.join():
|
| 11 |
+
File "/opt/conda/lib/python3.9/site-packages/torch/multiprocessing/spawn.py", line 189, in join
|
| 12 |
+
raise ProcessRaisedException(msg, error_index, failed_process.pid)
|
| 13 |
+
torch.multiprocessing.spawn.ProcessRaisedException:
|
| 14 |
+
|
| 15 |
+
-- Process 0 terminated with the following error:
|
| 16 |
+
Traceback (most recent call last):
|
| 17 |
+
File "/opt/conda/lib/python3.9/site-packages/torch/multiprocessing/spawn.py", line 76, in _wrap
|
| 18 |
+
fn(i, *args)
|
| 19 |
+
File "/optimum-benchmark/optimum_benchmark/launchers/process/launcher.py", line 63, in entrypoint
|
| 20 |
+
worker_output = worker(*worker_args)
|
| 21 |
+
File "/optimum-benchmark/optimum_benchmark/experiment.py", line 55, in run
|
| 22 |
+
backend: Backend = backend_factory(backend_config)
|
| 23 |
+
File "/optimum-benchmark/optimum_benchmark/backends/pytorch/backend.py", line 45, in __init__
|
| 24 |
+
super().__init__(config)
|
| 25 |
+
File "/optimum-benchmark/optimum_benchmark/backends/base.py", line 54, in __init__
|
| 26 |
+
self.pretrained_config = get_timm_pretrained_config(self.config.model)
|
| 27 |
+
File "/optimum-benchmark/optimum_benchmark/backends/timm_utils.py", line 12, in get_timm_pretrained_config
|
| 28 |
+
model_source, model_name = timm.models.parse_model_name(model_name)
|
| 29 |
+
NameError: name 'timm' is not defined
|
| 30 |
+
|
| 31 |
+
|
| 32 |
+
Set the environment variable HYDRA_FULL_ERROR=1 for a complete stack trace.
|
image_classification/timm/resnet50.a1_in1k/2024-12-04-17-13-03/experiment_config.json
ADDED
|
@@ -0,0 +1,107 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"experiment_name": "image_classification",
|
| 3 |
+
"backend": {
|
| 4 |
+
"name": "pytorch",
|
| 5 |
+
"version": "2.4.0",
|
| 6 |
+
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
|
| 7 |
+
"task": "image-classification",
|
| 8 |
+
"model": "timm/resnet50.a1_in1k",
|
| 9 |
+
"processor": "timm/resnet50.a1_in1k",
|
| 10 |
+
"library": "timm",
|
| 11 |
+
"device": "cuda",
|
| 12 |
+
"device_ids": "0",
|
| 13 |
+
"seed": 42,
|
| 14 |
+
"inter_op_num_threads": null,
|
| 15 |
+
"intra_op_num_threads": null,
|
| 16 |
+
"hub_kwargs": {
|
| 17 |
+
"revision": "main",
|
| 18 |
+
"force_download": false,
|
| 19 |
+
"local_files_only": false,
|
| 20 |
+
"trust_remote_code": true
|
| 21 |
+
},
|
| 22 |
+
"no_weights": true,
|
| 23 |
+
"device_map": null,
|
| 24 |
+
"torch_dtype": null,
|
| 25 |
+
"amp_autocast": false,
|
| 26 |
+
"amp_dtype": null,
|
| 27 |
+
"eval_mode": true,
|
| 28 |
+
"to_bettertransformer": false,
|
| 29 |
+
"low_cpu_mem_usage": null,
|
| 30 |
+
"attn_implementation": null,
|
| 31 |
+
"cache_implementation": null,
|
| 32 |
+
"torch_compile": false,
|
| 33 |
+
"torch_compile_config": {},
|
| 34 |
+
"quantization_scheme": null,
|
| 35 |
+
"quantization_config": {},
|
| 36 |
+
"deepspeed_inference": false,
|
| 37 |
+
"deepspeed_inference_config": {},
|
| 38 |
+
"peft_type": null,
|
| 39 |
+
"peft_config": {}
|
| 40 |
+
},
|
| 41 |
+
"launcher": {
|
| 42 |
+
"name": "process",
|
| 43 |
+
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
|
| 44 |
+
"device_isolation": true,
|
| 45 |
+
"device_isolation_action": "warn",
|
| 46 |
+
"start_method": "spawn"
|
| 47 |
+
},
|
| 48 |
+
"benchmark": {
|
| 49 |
+
"name": "energy_star",
|
| 50 |
+
"_target_": "optimum_benchmark.benchmarks.energy_star.benchmark.EnergyStarBenchmark",
|
| 51 |
+
"dataset_name": "EnergyStarAI/image_classification",
|
| 52 |
+
"dataset_config": "",
|
| 53 |
+
"dataset_split": "train",
|
| 54 |
+
"num_samples": 1000,
|
| 55 |
+
"input_shapes": {
|
| 56 |
+
"batch_size": 1
|
| 57 |
+
},
|
| 58 |
+
"text_column_name": "text",
|
| 59 |
+
"truncation": true,
|
| 60 |
+
"max_length": -1,
|
| 61 |
+
"dataset_prefix1": "",
|
| 62 |
+
"dataset_prefix2": "",
|
| 63 |
+
"t5_task": "",
|
| 64 |
+
"image_column_name": "image",
|
| 65 |
+
"resize": false,
|
| 66 |
+
"question_column_name": "question",
|
| 67 |
+
"context_column_name": "context",
|
| 68 |
+
"sentence1_column_name": "sentence1",
|
| 69 |
+
"sentence2_column_name": "sentence2",
|
| 70 |
+
"audio_column_name": "audio",
|
| 71 |
+
"iterations": 10,
|
| 72 |
+
"warmup_runs": 10,
|
| 73 |
+
"energy": true,
|
| 74 |
+
"forward_kwargs": {},
|
| 75 |
+
"generate_kwargs": {},
|
| 76 |
+
"call_kwargs": {}
|
| 77 |
+
},
|
| 78 |
+
"environment": {
|
| 79 |
+
"cpu": " AMD EPYC 7R32",
|
| 80 |
+
"cpu_count": 48,
|
| 81 |
+
"cpu_ram_mb": 200472.73984,
|
| 82 |
+
"system": "Linux",
|
| 83 |
+
"machine": "x86_64",
|
| 84 |
+
"platform": "Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35",
|
| 85 |
+
"processor": "x86_64",
|
| 86 |
+
"python_version": "3.9.20",
|
| 87 |
+
"gpu": [
|
| 88 |
+
"NVIDIA A10G"
|
| 89 |
+
],
|
| 90 |
+
"gpu_count": 1,
|
| 91 |
+
"gpu_vram_mb": 24146608128,
|
| 92 |
+
"optimum_benchmark_version": "0.2.0",
|
| 93 |
+
"optimum_benchmark_commit": null,
|
| 94 |
+
"transformers_version": "4.44.0",
|
| 95 |
+
"transformers_commit": null,
|
| 96 |
+
"accelerate_version": "0.33.0",
|
| 97 |
+
"accelerate_commit": null,
|
| 98 |
+
"diffusers_version": "0.30.0",
|
| 99 |
+
"diffusers_commit": null,
|
| 100 |
+
"optimum_version": null,
|
| 101 |
+
"optimum_commit": null,
|
| 102 |
+
"timm_version": null,
|
| 103 |
+
"timm_commit": null,
|
| 104 |
+
"peft_version": null,
|
| 105 |
+
"peft_commit": null
|
| 106 |
+
}
|
| 107 |
+
}
|
text_generation/bigscience/bloomz-560m/2024-12-04-16-38-24/.hydra/config.yaml
ADDED
|
@@ -0,0 +1,96 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
backend:
|
| 2 |
+
name: pytorch
|
| 3 |
+
version: 2.4.0
|
| 4 |
+
_target_: optimum_benchmark.backends.pytorch.backend.PyTorchBackend
|
| 5 |
+
task: text-generation
|
| 6 |
+
model: bigscience/bloomz-560m
|
| 7 |
+
processor: bigscience/bloomz-560m
|
| 8 |
+
library: null
|
| 9 |
+
device: cuda
|
| 10 |
+
device_ids: '0'
|
| 11 |
+
seed: 42
|
| 12 |
+
inter_op_num_threads: null
|
| 13 |
+
intra_op_num_threads: null
|
| 14 |
+
hub_kwargs: {}
|
| 15 |
+
no_weights: true
|
| 16 |
+
device_map: null
|
| 17 |
+
torch_dtype: null
|
| 18 |
+
amp_autocast: false
|
| 19 |
+
amp_dtype: null
|
| 20 |
+
eval_mode: true
|
| 21 |
+
to_bettertransformer: false
|
| 22 |
+
low_cpu_mem_usage: null
|
| 23 |
+
attn_implementation: null
|
| 24 |
+
cache_implementation: null
|
| 25 |
+
torch_compile: false
|
| 26 |
+
torch_compile_config: {}
|
| 27 |
+
quantization_scheme: null
|
| 28 |
+
quantization_config: {}
|
| 29 |
+
deepspeed_inference: false
|
| 30 |
+
deepspeed_inference_config: {}
|
| 31 |
+
peft_type: null
|
| 32 |
+
peft_config: {}
|
| 33 |
+
launcher:
|
| 34 |
+
name: process
|
| 35 |
+
_target_: optimum_benchmark.launchers.process.launcher.ProcessLauncher
|
| 36 |
+
device_isolation: false
|
| 37 |
+
device_isolation_action: warn
|
| 38 |
+
start_method: spawn
|
| 39 |
+
benchmark:
|
| 40 |
+
name: energy_star
|
| 41 |
+
_target_: optimum_benchmark.benchmarks.energy_star.benchmark.EnergyStarBenchmark
|
| 42 |
+
dataset_name: EnergyStarAI/text_generation
|
| 43 |
+
dataset_config: ''
|
| 44 |
+
dataset_split: train
|
| 45 |
+
num_samples: 1000
|
| 46 |
+
input_shapes:
|
| 47 |
+
batch_size: 1
|
| 48 |
+
text_column_name: text
|
| 49 |
+
truncation: true
|
| 50 |
+
max_length: -1
|
| 51 |
+
dataset_prefix1: ''
|
| 52 |
+
dataset_prefix2: ''
|
| 53 |
+
t5_task: ''
|
| 54 |
+
image_column_name: image
|
| 55 |
+
resize: false
|
| 56 |
+
question_column_name: question
|
| 57 |
+
context_column_name: context
|
| 58 |
+
sentence1_column_name: sentence1
|
| 59 |
+
sentence2_column_name: sentence2
|
| 60 |
+
audio_column_name: audio
|
| 61 |
+
iterations: 10
|
| 62 |
+
warmup_runs: 10
|
| 63 |
+
energy: true
|
| 64 |
+
forward_kwargs: {}
|
| 65 |
+
generate_kwargs:
|
| 66 |
+
max_new_tokens: 10
|
| 67 |
+
min_new_tokens: 10
|
| 68 |
+
call_kwargs: {}
|
| 69 |
+
experiment_name: text_generation
|
| 70 |
+
environment:
|
| 71 |
+
cpu: ' AMD EPYC 7R32'
|
| 72 |
+
cpu_count: 48
|
| 73 |
+
cpu_ram_mb: 200472.73984
|
| 74 |
+
system: Linux
|
| 75 |
+
machine: x86_64
|
| 76 |
+
platform: Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35
|
| 77 |
+
processor: x86_64
|
| 78 |
+
python_version: 3.9.20
|
| 79 |
+
gpu:
|
| 80 |
+
- NVIDIA A10G
|
| 81 |
+
gpu_count: 1
|
| 82 |
+
gpu_vram_mb: 24146608128
|
| 83 |
+
optimum_benchmark_version: 0.2.0
|
| 84 |
+
optimum_benchmark_commit: null
|
| 85 |
+
transformers_version: 4.44.0
|
| 86 |
+
transformers_commit: null
|
| 87 |
+
accelerate_version: 0.33.0
|
| 88 |
+
accelerate_commit: null
|
| 89 |
+
diffusers_version: 0.30.0
|
| 90 |
+
diffusers_commit: null
|
| 91 |
+
optimum_version: null
|
| 92 |
+
optimum_commit: null
|
| 93 |
+
timm_version: null
|
| 94 |
+
timm_commit: null
|
| 95 |
+
peft_version: null
|
| 96 |
+
peft_commit: null
|
text_generation/bigscience/bloomz-560m/2024-12-04-16-38-24/.hydra/hydra.yaml
ADDED
|
@@ -0,0 +1,175 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
hydra:
|
| 2 |
+
run:
|
| 3 |
+
dir: /runs/text_generation/bigscience/bloomz-560m/2024-12-04-16-38-24
|
| 4 |
+
sweep:
|
| 5 |
+
dir: sweeps/${experiment_name}/${backend.model}/${now:%Y-%m-%d-%H-%M-%S}
|
| 6 |
+
subdir: ${hydra.job.num}
|
| 7 |
+
launcher:
|
| 8 |
+
_target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
|
| 9 |
+
sweeper:
|
| 10 |
+
_target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
|
| 11 |
+
max_batch_size: null
|
| 12 |
+
params: null
|
| 13 |
+
help:
|
| 14 |
+
app_name: ${hydra.job.name}
|
| 15 |
+
header: '${hydra.help.app_name} is powered by Hydra.
|
| 16 |
+
|
| 17 |
+
'
|
| 18 |
+
footer: 'Powered by Hydra (https://hydra.cc)
|
| 19 |
+
|
| 20 |
+
Use --hydra-help to view Hydra specific help
|
| 21 |
+
|
| 22 |
+
'
|
| 23 |
+
template: '${hydra.help.header}
|
| 24 |
+
|
| 25 |
+
== Configuration groups ==
|
| 26 |
+
|
| 27 |
+
Compose your configuration from those groups (group=option)
|
| 28 |
+
|
| 29 |
+
|
| 30 |
+
$APP_CONFIG_GROUPS
|
| 31 |
+
|
| 32 |
+
|
| 33 |
+
== Config ==
|
| 34 |
+
|
| 35 |
+
Override anything in the config (foo.bar=value)
|
| 36 |
+
|
| 37 |
+
|
| 38 |
+
$CONFIG
|
| 39 |
+
|
| 40 |
+
|
| 41 |
+
${hydra.help.footer}
|
| 42 |
+
|
| 43 |
+
'
|
| 44 |
+
hydra_help:
|
| 45 |
+
template: 'Hydra (${hydra.runtime.version})
|
| 46 |
+
|
| 47 |
+
See https://hydra.cc for more info.
|
| 48 |
+
|
| 49 |
+
|
| 50 |
+
== Flags ==
|
| 51 |
+
|
| 52 |
+
$FLAGS_HELP
|
| 53 |
+
|
| 54 |
+
|
| 55 |
+
== Configuration groups ==
|
| 56 |
+
|
| 57 |
+
Compose your configuration from those groups (For example, append hydra/job_logging=disabled
|
| 58 |
+
to command line)
|
| 59 |
+
|
| 60 |
+
|
| 61 |
+
$HYDRA_CONFIG_GROUPS
|
| 62 |
+
|
| 63 |
+
|
| 64 |
+
Use ''--cfg hydra'' to Show the Hydra config.
|
| 65 |
+
|
| 66 |
+
'
|
| 67 |
+
hydra_help: ???
|
| 68 |
+
hydra_logging:
|
| 69 |
+
version: 1
|
| 70 |
+
formatters:
|
| 71 |
+
colorlog:
|
| 72 |
+
(): colorlog.ColoredFormatter
|
| 73 |
+
format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s'
|
| 74 |
+
handlers:
|
| 75 |
+
console:
|
| 76 |
+
class: logging.StreamHandler
|
| 77 |
+
formatter: colorlog
|
| 78 |
+
stream: ext://sys.stdout
|
| 79 |
+
root:
|
| 80 |
+
level: INFO
|
| 81 |
+
handlers:
|
| 82 |
+
- console
|
| 83 |
+
disable_existing_loggers: false
|
| 84 |
+
job_logging:
|
| 85 |
+
version: 1
|
| 86 |
+
formatters:
|
| 87 |
+
simple:
|
| 88 |
+
format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
|
| 89 |
+
colorlog:
|
| 90 |
+
(): colorlog.ColoredFormatter
|
| 91 |
+
format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s]
|
| 92 |
+
- %(message)s'
|
| 93 |
+
log_colors:
|
| 94 |
+
DEBUG: purple
|
| 95 |
+
INFO: green
|
| 96 |
+
WARNING: yellow
|
| 97 |
+
ERROR: red
|
| 98 |
+
CRITICAL: red
|
| 99 |
+
handlers:
|
| 100 |
+
console:
|
| 101 |
+
class: logging.StreamHandler
|
| 102 |
+
formatter: colorlog
|
| 103 |
+
stream: ext://sys.stdout
|
| 104 |
+
file:
|
| 105 |
+
class: logging.FileHandler
|
| 106 |
+
formatter: simple
|
| 107 |
+
filename: ${hydra.job.name}.log
|
| 108 |
+
root:
|
| 109 |
+
level: INFO
|
| 110 |
+
handlers:
|
| 111 |
+
- console
|
| 112 |
+
- file
|
| 113 |
+
disable_existing_loggers: false
|
| 114 |
+
env: {}
|
| 115 |
+
mode: RUN
|
| 116 |
+
searchpath: []
|
| 117 |
+
callbacks: {}
|
| 118 |
+
output_subdir: .hydra
|
| 119 |
+
overrides:
|
| 120 |
+
hydra:
|
| 121 |
+
- hydra.run.dir=/runs/text_generation/bigscience/bloomz-560m/2024-12-04-16-38-24
|
| 122 |
+
- hydra.mode=RUN
|
| 123 |
+
task:
|
| 124 |
+
- backend.model=bigscience/bloomz-560m
|
| 125 |
+
- backend.processor=bigscience/bloomz-560m
|
| 126 |
+
job:
|
| 127 |
+
name: cli
|
| 128 |
+
chdir: true
|
| 129 |
+
override_dirname: backend.model=bigscience/bloomz-560m,backend.processor=bigscience/bloomz-560m
|
| 130 |
+
id: ???
|
| 131 |
+
num: ???
|
| 132 |
+
config_name: text_generation
|
| 133 |
+
env_set:
|
| 134 |
+
OVERRIDE_BENCHMARKS: '1'
|
| 135 |
+
env_copy: []
|
| 136 |
+
config:
|
| 137 |
+
override_dirname:
|
| 138 |
+
kv_sep: '='
|
| 139 |
+
item_sep: ','
|
| 140 |
+
exclude_keys: []
|
| 141 |
+
runtime:
|
| 142 |
+
version: 1.3.2
|
| 143 |
+
version_base: '1.3'
|
| 144 |
+
cwd: /
|
| 145 |
+
config_sources:
|
| 146 |
+
- path: hydra.conf
|
| 147 |
+
schema: pkg
|
| 148 |
+
provider: hydra
|
| 149 |
+
- path: optimum_benchmark
|
| 150 |
+
schema: pkg
|
| 151 |
+
provider: main
|
| 152 |
+
- path: hydra_plugins.hydra_colorlog.conf
|
| 153 |
+
schema: pkg
|
| 154 |
+
provider: hydra-colorlog
|
| 155 |
+
- path: /optimum-benchmark/examples/energy_star
|
| 156 |
+
schema: file
|
| 157 |
+
provider: command-line
|
| 158 |
+
- path: ''
|
| 159 |
+
schema: structured
|
| 160 |
+
provider: schema
|
| 161 |
+
output_dir: /runs/text_generation/bigscience/bloomz-560m/2024-12-04-16-38-24
|
| 162 |
+
choices:
|
| 163 |
+
benchmark: energy_star
|
| 164 |
+
launcher: process
|
| 165 |
+
backend: pytorch
|
| 166 |
+
hydra/env: default
|
| 167 |
+
hydra/callbacks: null
|
| 168 |
+
hydra/job_logging: colorlog
|
| 169 |
+
hydra/hydra_logging: colorlog
|
| 170 |
+
hydra/hydra_help: default
|
| 171 |
+
hydra/help: default
|
| 172 |
+
hydra/sweeper: basic
|
| 173 |
+
hydra/launcher: basic
|
| 174 |
+
hydra/output: default
|
| 175 |
+
verbose: false
|
text_generation/bigscience/bloomz-560m/2024-12-04-16-38-24/.hydra/overrides.yaml
ADDED
|
@@ -0,0 +1,2 @@
|
|
|
|
|
|
|
|
|
|
| 1 |
+
- backend.model=bigscience/bloomz-560m
|
| 2 |
+
- backend.processor=bigscience/bloomz-560m
|
text_generation/bigscience/bloomz-560m/2024-12-04-16-38-24/benchmark_report.json
ADDED
|
@@ -0,0 +1,203 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"prefill": {
|
| 3 |
+
"memory": null,
|
| 4 |
+
"latency": null,
|
| 5 |
+
"throughput": null,
|
| 6 |
+
"energy": {
|
| 7 |
+
"unit": "kWh",
|
| 8 |
+
"cpu": 0.0003082951182833939,
|
| 9 |
+
"ram": 3.1850969165582464e-06,
|
| 10 |
+
"gpu": 0.001811276337908796,
|
| 11 |
+
"total": 0.0021227565531087484
|
| 12 |
+
},
|
| 13 |
+
"efficiency": {
|
| 14 |
+
"unit": "tokens/kWh",
|
| 15 |
+
"value": 91065082.2002653
|
| 16 |
+
},
|
| 17 |
+
"measures": [
|
| 18 |
+
{
|
| 19 |
+
"unit": "kWh",
|
| 20 |
+
"cpu": 0.0003429987985006948,
|
| 21 |
+
"ram": 3.542954712246048e-06,
|
| 22 |
+
"gpu": 0.001997280486711994,
|
| 23 |
+
"total": 0.002343822239924935
|
| 24 |
+
},
|
| 25 |
+
{
|
| 26 |
+
"unit": "kWh",
|
| 27 |
+
"cpu": 0.00034366121815134854,
|
| 28 |
+
"ram": 3.5501402373248306e-06,
|
| 29 |
+
"gpu": 0.002026376065544011,
|
| 30 |
+
"total": 0.002373587423932685
|
| 31 |
+
},
|
| 32 |
+
{
|
| 33 |
+
"unit": "kWh",
|
| 34 |
+
"cpu": 0.00034208863734233,
|
| 35 |
+
"ram": 3.5343963865605078e-06,
|
| 36 |
+
"gpu": 0.0020103096638020013,
|
| 37 |
+
"total": 0.0023559326975308914
|
| 38 |
+
},
|
| 39 |
+
{
|
| 40 |
+
"unit": "kWh",
|
| 41 |
+
"cpu": 0.00034247366563189535,
|
| 42 |
+
"ram": 3.538371880745545e-06,
|
| 43 |
+
"gpu": 0.002014361333710024,
|
| 44 |
+
"total": 0.0023603733712226653
|
| 45 |
+
},
|
| 46 |
+
{
|
| 47 |
+
"unit": "kWh",
|
| 48 |
+
"cpu": 0.00034193711197430185,
|
| 49 |
+
"ram": 3.5326911277494935e-06,
|
| 50 |
+
"gpu": 0.0020110554977319506,
|
| 51 |
+
"total": 0.0023565253008340017
|
| 52 |
+
},
|
| 53 |
+
{
|
| 54 |
+
"unit": "kWh",
|
| 55 |
+
"cpu": 0.0003427284664014071,
|
| 56 |
+
"ram": 3.541041700480612e-06,
|
| 57 |
+
"gpu": 0.0020142532780679234,
|
| 58 |
+
"total": 0.002360522786169812
|
| 59 |
+
},
|
| 60 |
+
{
|
| 61 |
+
"unit": "kWh",
|
| 62 |
+
"cpu": 0.0,
|
| 63 |
+
"ram": 0.0,
|
| 64 |
+
"gpu": 0.0,
|
| 65 |
+
"total": 0.0
|
| 66 |
+
},
|
| 67 |
+
{
|
| 68 |
+
"unit": "kWh",
|
| 69 |
+
"cpu": 0.00034183093565275146,
|
| 70 |
+
"ram": 3.531765329011117e-06,
|
| 71 |
+
"gpu": 0.0020089163293540713,
|
| 72 |
+
"total": 0.002354279030335832
|
| 73 |
+
},
|
| 74 |
+
{
|
| 75 |
+
"unit": "kWh",
|
| 76 |
+
"cpu": 0.00034218637264447774,
|
| 77 |
+
"ram": 3.5352872224273327e-06,
|
| 78 |
+
"gpu": 0.002010571608455969,
|
| 79 |
+
"total": 0.0023562932683228767
|
| 80 |
+
},
|
| 81 |
+
{
|
| 82 |
+
"unit": "kWh",
|
| 83 |
+
"cpu": 0.00034304597653473214,
|
| 84 |
+
"ram": 3.5443205690369786e-06,
|
| 85 |
+
"gpu": 0.0020196391157100146,
|
| 86 |
+
"total": 0.002366229412813784
|
| 87 |
+
}
|
| 88 |
+
]
|
| 89 |
+
},
|
| 90 |
+
"decode": {
|
| 91 |
+
"memory": null,
|
| 92 |
+
"latency": null,
|
| 93 |
+
"throughput": null,
|
| 94 |
+
"energy": {
|
| 95 |
+
"unit": "kWh",
|
| 96 |
+
"cpu": 0.0015752735088163188,
|
| 97 |
+
"ram": 1.6288348539985184e-05,
|
| 98 |
+
"gpu": 0.003964184199122589,
|
| 99 |
+
"total": 0.005555746056478895
|
| 100 |
+
},
|
| 101 |
+
"efficiency": {
|
| 102 |
+
"unit": "tokens/kWh",
|
| 103 |
+
"value": 1619944.45183551
|
| 104 |
+
},
|
| 105 |
+
"measures": [
|
| 106 |
+
{
|
| 107 |
+
"unit": "kWh",
|
| 108 |
+
"cpu": 0.0017516841623555145,
|
| 109 |
+
"ram": 1.8112295395436646e-05,
|
| 110 |
+
"gpu": 0.004435401603874012,
|
| 111 |
+
"total": 0.006205198061624966
|
| 112 |
+
},
|
| 113 |
+
{
|
| 114 |
+
"unit": "kWh",
|
| 115 |
+
"cpu": 0.0017452471201590872,
|
| 116 |
+
"ram": 1.8046290018650998e-05,
|
| 117 |
+
"gpu": 0.004403079633571894,
|
| 118 |
+
"total": 0.006166373043749631
|
| 119 |
+
},
|
| 120 |
+
{
|
| 121 |
+
"unit": "kWh",
|
| 122 |
+
"cpu": 0.0017454764441145717,
|
| 123 |
+
"ram": 1.8048288935674893e-05,
|
| 124 |
+
"gpu": 0.004411841585026166,
|
| 125 |
+
"total": 0.006175366318076415
|
| 126 |
+
},
|
| 127 |
+
{
|
| 128 |
+
"unit": "kWh",
|
| 129 |
+
"cpu": 0.0017546621838340397,
|
| 130 |
+
"ram": 1.8143266360687642e-05,
|
| 131 |
+
"gpu": 0.004420565203115845,
|
| 132 |
+
"total": 0.006193370653310573
|
| 133 |
+
},
|
| 134 |
+
{
|
| 135 |
+
"unit": "kWh",
|
| 136 |
+
"cpu": -0.00034193711197430185,
|
| 137 |
+
"ram": -3.5326911277494935e-06,
|
| 138 |
+
"gpu": -0.0020110554977319506,
|
| 139 |
+
"total": -0.0023565253008340017
|
| 140 |
+
},
|
| 141 |
+
{
|
| 142 |
+
"unit": "kWh",
|
| 143 |
+
"cpu": 0.0017451369265368472,
|
| 144 |
+
"ram": 1.804475479899136e-05,
|
| 145 |
+
"gpu": 0.00441268325236599,
|
| 146 |
+
"total": 0.006175864933701829
|
| 147 |
+
},
|
| 148 |
+
{
|
| 149 |
+
"unit": "kWh",
|
| 150 |
+
"cpu": 0.0020869030950346965,
|
| 151 |
+
"ram": 2.157583443595883e-05,
|
| 152 |
+
"gpu": 0.0064130976304739296,
|
| 153 |
+
"total": 0.008521576559944594
|
| 154 |
+
},
|
| 155 |
+
{
|
| 156 |
+
"unit": "kWh",
|
| 157 |
+
"cpu": 0.001747071537846492,
|
| 158 |
+
"ram": 1.8064560154956028e-05,
|
| 159 |
+
"gpu": 0.004382076283436009,
|
| 160 |
+
"total": 0.006147212381437464
|
| 161 |
+
},
|
| 162 |
+
{
|
| 163 |
+
"unit": "kWh",
|
| 164 |
+
"cpu": 0.001746269038160391,
|
| 165 |
+
"ram": 1.805659913703533e-05,
|
| 166 |
+
"gpu": 0.0043722779422640246,
|
| 167 |
+
"total": 0.006136603579561441
|
| 168 |
+
},
|
| 169 |
+
{
|
| 170 |
+
"unit": "kWh",
|
| 171 |
+
"cpu": 0.0017722216920958499,
|
| 172 |
+
"ram": 1.8324287290209603e-05,
|
| 173 |
+
"gpu": 0.004401874354829971,
|
| 174 |
+
"total": 0.00619242033421603
|
| 175 |
+
}
|
| 176 |
+
]
|
| 177 |
+
},
|
| 178 |
+
"per_token": {
|
| 179 |
+
"memory": null,
|
| 180 |
+
"latency": null,
|
| 181 |
+
"throughput": null,
|
| 182 |
+
"energy": null,
|
| 183 |
+
"efficiency": null,
|
| 184 |
+
"measures": null
|
| 185 |
+
},
|
| 186 |
+
"preprocess": {
|
| 187 |
+
"memory": null,
|
| 188 |
+
"latency": null,
|
| 189 |
+
"throughput": null,
|
| 190 |
+
"energy": {
|
| 191 |
+
"unit": "kWh",
|
| 192 |
+
"cpu": 1.189836183752454e-05,
|
| 193 |
+
"ram": 8.778073359449114e-08,
|
| 194 |
+
"gpu": 1.8812515049981826e-05,
|
| 195 |
+
"total": 3.079865762110086e-05
|
| 196 |
+
},
|
| 197 |
+
"efficiency": {
|
| 198 |
+
"unit": "samples/kWh",
|
| 199 |
+
"value": 32468947.585393377
|
| 200 |
+
},
|
| 201 |
+
"measures": null
|
| 202 |
+
}
|
| 203 |
+
}
|
text_generation/bigscience/bloomz-560m/2024-12-04-16-38-24/cli.log
ADDED
|
@@ -0,0 +1,188 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
[2024-12-04 16:38:27,209][launcher][INFO] - ََAllocating process launcher
|
| 2 |
+
[2024-12-04 16:38:27,209][process][INFO] - + Setting multiprocessing start method to spawn.
|
| 3 |
+
[2024-12-04 16:38:27,223][process][INFO] - + Launched benchmark in isolated process 800.
|
| 4 |
+
[PROC-0][2024-12-04 16:38:29,734][datasets][INFO] - PyTorch version 2.4.0 available.
|
| 5 |
+
[PROC-0][2024-12-04 16:38:30,650][backend][INFO] - َAllocating pytorch backend
|
| 6 |
+
[PROC-0][2024-12-04 16:38:30,651][backend][INFO] - + Setting random seed to 42
|
| 7 |
+
[PROC-0][2024-12-04 16:38:32,113][pytorch][INFO] - + Using AutoModel class AutoModelForCausalLM
|
| 8 |
+
[PROC-0][2024-12-04 16:38:32,113][pytorch][INFO] - + Creating backend temporary directory
|
| 9 |
+
[PROC-0][2024-12-04 16:38:32,114][pytorch][INFO] - + Loading model with random weights
|
| 10 |
+
[PROC-0][2024-12-04 16:38:32,114][pytorch][INFO] - + Creating no weights model
|
| 11 |
+
[PROC-0][2024-12-04 16:38:32,114][pytorch][INFO] - + Creating no weights model directory
|
| 12 |
+
[PROC-0][2024-12-04 16:38:32,114][pytorch][INFO] - + Creating no weights model state dict
|
| 13 |
+
[PROC-0][2024-12-04 16:38:32,133][pytorch][INFO] - + Saving no weights model safetensors
|
| 14 |
+
[PROC-0][2024-12-04 16:38:32,134][pytorch][INFO] - + Saving no weights model pretrained config
|
| 15 |
+
[PROC-0][2024-12-04 16:38:32,135][pytorch][INFO] - + Loading no weights AutoModel
|
| 16 |
+
[PROC-0][2024-12-04 16:38:32,135][pytorch][INFO] - + Loading model directly on device: cuda
|
| 17 |
+
[PROC-0][2024-12-04 16:38:32,417][pytorch][INFO] - + Turning on model's eval mode
|
| 18 |
+
[PROC-0][2024-12-04 16:38:32,424][benchmark][INFO] - Allocating energy_star benchmark
|
| 19 |
+
[PROC-0][2024-12-04 16:38:32,424][energy_star][INFO] - + Loading raw dataset
|
| 20 |
+
[PROC-0][2024-12-04 16:38:33,147][energy_star][INFO] - + Updating Text Generation kwargs with default values
|
| 21 |
+
[PROC-0][2024-12-04 16:38:33,147][energy_star][INFO] - + Initializing Text Generation report
|
| 22 |
+
[PROC-0][2024-12-04 16:38:33,147][energy][INFO] - + Tracking GPU energy on devices [0]
|
| 23 |
+
[PROC-0][2024-12-04 16:38:37,332][energy_star][INFO] - + Preprocessing dataset
|
| 24 |
+
[PROC-0][2024-12-04 16:38:38,341][energy][INFO] - + Saving codecarbon emission data to preprocess_codecarbon.json
|
| 25 |
+
[PROC-0][2024-12-04 16:38:38,341][energy_star][INFO] - + Preparing backend for Inference
|
| 26 |
+
[PROC-0][2024-12-04 16:38:38,341][energy_star][INFO] - + Initialising dataloader
|
| 27 |
+
[PROC-0][2024-12-04 16:38:38,341][energy_star][INFO] - + Warming up backend for Inference
|
| 28 |
+
[PROC-0][2024-12-04 16:38:39,395][energy_star][INFO] - + Additional warmup for Text Generation
|
| 29 |
+
[PROC-0][2024-12-04 16:38:39,576][energy_star][INFO] - + Running Text Generation energy tracking for 10 iterations
|
| 30 |
+
[PROC-0][2024-12-04 16:38:39,576][energy_star][INFO] - + Prefill iteration 1/10
|
| 31 |
+
[PROC-0][2024-12-04 16:39:08,631][energy][INFO] - + Saving codecarbon emission data to prefill_codecarbon.json
|
| 32 |
+
[PROC-0][2024-12-04 16:39:08,631][energy_star][INFO] - + Prefill iteration 2/10
|
| 33 |
+
[PROC-0][2024-12-04 16:39:37,742][energy][INFO] - + Saving codecarbon emission data to prefill_codecarbon.json
|
| 34 |
+
[PROC-0][2024-12-04 16:39:37,742][energy_star][INFO] - + Prefill iteration 3/10
|
| 35 |
+
[PROC-0][2024-12-04 16:40:06,720][energy][INFO] - + Saving codecarbon emission data to prefill_codecarbon.json
|
| 36 |
+
[PROC-0][2024-12-04 16:40:06,720][energy_star][INFO] - + Prefill iteration 4/10
|
| 37 |
+
[PROC-0][2024-12-04 16:40:35,730][energy][INFO] - + Saving codecarbon emission data to prefill_codecarbon.json
|
| 38 |
+
[PROC-0][2024-12-04 16:40:35,731][energy_star][INFO] - + Prefill iteration 5/10
|
| 39 |
+
[PROC-0][2024-12-04 16:41:04,696][energy][INFO] - + Saving codecarbon emission data to prefill_codecarbon.json
|
| 40 |
+
[PROC-0][2024-12-04 16:41:04,696][energy_star][INFO] - + Prefill iteration 6/10
|
| 41 |
+
[PROC-0][2024-12-04 16:41:33,728][energy][INFO] - + Saving codecarbon emission data to prefill_codecarbon.json
|
| 42 |
+
[PROC-0][2024-12-04 16:41:33,728][energy_star][INFO] - + Prefill iteration 7/10
|
| 43 |
+
[PROC-0][2024-12-04 16:42:02,696][energy][INFO] - + Saving codecarbon emission data to prefill_codecarbon.json
|
| 44 |
+
[PROC-0][2024-12-04 16:42:02,696][energy_star][INFO] - + Prefill iteration 8/10
|
| 45 |
+
[PROC-0][2024-12-04 16:42:31,652][energy][INFO] - + Saving codecarbon emission data to prefill_codecarbon.json
|
| 46 |
+
[PROC-0][2024-12-04 16:42:31,652][energy_star][INFO] - + Prefill iteration 9/10
|
| 47 |
+
[PROC-0][2024-12-04 16:43:00,638][energy][INFO] - + Saving codecarbon emission data to prefill_codecarbon.json
|
| 48 |
+
[PROC-0][2024-12-04 16:43:00,639][energy_star][INFO] - + Prefill iteration 10/10
|
| 49 |
+
[PROC-0][2024-12-04 16:43:29,697][energy][INFO] - + Saving codecarbon emission data to prefill_codecarbon.json
|
| 50 |
+
[PROC-0][2024-12-04 16:43:29,698][energy_star][INFO] - + Decoding iteration 1/10
|
| 51 |
+
[PROC-0][2024-12-04 16:46:27,130][energy][INFO] - + Saving codecarbon emission data to generate_codecarbon.json
|
| 52 |
+
[PROC-0][2024-12-04 16:46:27,131][energy_star][INFO] - + Decoding iteration 2/10
|
| 53 |
+
[PROC-0][2024-12-04 16:49:24,074][energy][INFO] - + Saving codecarbon emission data to generate_codecarbon.json
|
| 54 |
+
[PROC-0][2024-12-04 16:49:24,075][energy_star][INFO] - + Decoding iteration 3/10
|
| 55 |
+
[PROC-0][2024-12-04 16:52:20,904][energy][INFO] - + Saving codecarbon emission data to generate_codecarbon.json
|
| 56 |
+
[PROC-0][2024-12-04 16:52:20,905][energy_star][INFO] - + Decoding iteration 4/10
|
| 57 |
+
[PROC-0][2024-12-04 16:55:18,545][energy][INFO] - + Saving codecarbon emission data to generate_codecarbon.json
|
| 58 |
+
[PROC-0][2024-12-04 16:55:18,545][energy_star][INFO] - + Decoding iteration 5/10
|
| 59 |
+
[PROC-0][2024-12-04 16:58:15,309][energy][INFO] - + Saving codecarbon emission data to generate_codecarbon.json
|
| 60 |
+
[PROC-0][2024-12-04 16:58:15,309][energy_star][INFO] - + Decoding iteration 6/10
|
| 61 |
+
[PROC-0][2024-12-04 17:01:12,164][energy][INFO] - + Saving codecarbon emission data to generate_codecarbon.json
|
| 62 |
+
[PROC-0][2024-12-04 17:01:12,165][energy_star][INFO] - + Decoding iteration 7/10
|
| 63 |
+
[PROC-0][2024-12-04 17:04:08,938][energy][INFO] - + Saving codecarbon emission data to generate_codecarbon.json
|
| 64 |
+
[PROC-0][2024-12-04 17:04:08,939][energy_star][INFO] - + Decoding iteration 8/10
|
| 65 |
+
[PROC-0][2024-12-04 17:07:05,881][energy][INFO] - + Saving codecarbon emission data to generate_codecarbon.json
|
| 66 |
+
[PROC-0][2024-12-04 17:07:05,882][energy_star][INFO] - + Decoding iteration 9/10
|
| 67 |
+
[PROC-0][2024-12-04 17:10:02,787][energy][INFO] - + Saving codecarbon emission data to generate_codecarbon.json
|
| 68 |
+
[PROC-0][2024-12-04 17:10:02,787][energy_star][INFO] - + Decoding iteration 10/10
|
| 69 |
+
[PROC-0][2024-12-04 17:13:01,964][energy][INFO] - + Saving codecarbon emission data to generate_codecarbon.json
|
| 70 |
+
[PROC-0][2024-12-04 17:13:01,964][energy][INFO] - + prefill energy consumption:
|
| 71 |
+
[PROC-0][2024-12-04 17:13:01,964][energy][INFO] - + CPU: 0.000308 (kWh)
|
| 72 |
+
[PROC-0][2024-12-04 17:13:01,965][energy][INFO] - + GPU: 0.001811 (kWh)
|
| 73 |
+
[PROC-0][2024-12-04 17:13:01,965][energy][INFO] - + RAM: 0.000003 (kWh)
|
| 74 |
+
[PROC-0][2024-12-04 17:13:01,965][energy][INFO] - + total: 0.002123 (kWh)
|
| 75 |
+
[PROC-0][2024-12-04 17:13:01,965][energy][INFO] - + prefill_iteration_1 energy consumption:
|
| 76 |
+
[PROC-0][2024-12-04 17:13:01,965][energy][INFO] - + CPU: 0.000343 (kWh)
|
| 77 |
+
[PROC-0][2024-12-04 17:13:01,965][energy][INFO] - + GPU: 0.001997 (kWh)
|
| 78 |
+
[PROC-0][2024-12-04 17:13:01,965][energy][INFO] - + RAM: 0.000004 (kWh)
|
| 79 |
+
[PROC-0][2024-12-04 17:13:01,965][energy][INFO] - + total: 0.002344 (kWh)
|
| 80 |
+
[PROC-0][2024-12-04 17:13:01,965][energy][INFO] - + prefill_iteration_2 energy consumption:
|
| 81 |
+
[PROC-0][2024-12-04 17:13:01,965][energy][INFO] - + CPU: 0.000344 (kWh)
|
| 82 |
+
[PROC-0][2024-12-04 17:13:01,965][energy][INFO] - + GPU: 0.002026 (kWh)
|
| 83 |
+
[PROC-0][2024-12-04 17:13:01,965][energy][INFO] - + RAM: 0.000004 (kWh)
|
| 84 |
+
[PROC-0][2024-12-04 17:13:01,966][energy][INFO] - + total: 0.002374 (kWh)
|
| 85 |
+
[PROC-0][2024-12-04 17:13:01,966][energy][INFO] - + prefill_iteration_3 energy consumption:
|
| 86 |
+
[PROC-0][2024-12-04 17:13:01,966][energy][INFO] - + CPU: 0.000342 (kWh)
|
| 87 |
+
[PROC-0][2024-12-04 17:13:01,966][energy][INFO] - + GPU: 0.002010 (kWh)
|
| 88 |
+
[PROC-0][2024-12-04 17:13:01,966][energy][INFO] - + RAM: 0.000004 (kWh)
|
| 89 |
+
[PROC-0][2024-12-04 17:13:01,966][energy][INFO] - + total: 0.002356 (kWh)
|
| 90 |
+
[PROC-0][2024-12-04 17:13:01,966][energy][INFO] - + prefill_iteration_4 energy consumption:
|
| 91 |
+
[PROC-0][2024-12-04 17:13:01,966][energy][INFO] - + CPU: 0.000342 (kWh)
|
| 92 |
+
[PROC-0][2024-12-04 17:13:01,966][energy][INFO] - + GPU: 0.002014 (kWh)
|
| 93 |
+
[PROC-0][2024-12-04 17:13:01,966][energy][INFO] - + RAM: 0.000004 (kWh)
|
| 94 |
+
[PROC-0][2024-12-04 17:13:01,966][energy][INFO] - + total: 0.002360 (kWh)
|
| 95 |
+
[PROC-0][2024-12-04 17:13:01,966][energy][INFO] - + prefill_iteration_5 energy consumption:
|
| 96 |
+
[PROC-0][2024-12-04 17:13:01,967][energy][INFO] - + CPU: 0.000342 (kWh)
|
| 97 |
+
[PROC-0][2024-12-04 17:13:01,967][energy][INFO] - + GPU: 0.002011 (kWh)
|
| 98 |
+
[PROC-0][2024-12-04 17:13:01,967][energy][INFO] - + RAM: 0.000004 (kWh)
|
| 99 |
+
[PROC-0][2024-12-04 17:13:01,967][energy][INFO] - + total: 0.002357 (kWh)
|
| 100 |
+
[PROC-0][2024-12-04 17:13:01,967][energy][INFO] - + prefill_iteration_6 energy consumption:
|
| 101 |
+
[PROC-0][2024-12-04 17:13:01,967][energy][INFO] - + CPU: 0.000343 (kWh)
|
| 102 |
+
[PROC-0][2024-12-04 17:13:01,967][energy][INFO] - + GPU: 0.002014 (kWh)
|
| 103 |
+
[PROC-0][2024-12-04 17:13:01,967][energy][INFO] - + RAM: 0.000004 (kWh)
|
| 104 |
+
[PROC-0][2024-12-04 17:13:01,967][energy][INFO] - + total: 0.002361 (kWh)
|
| 105 |
+
[PROC-0][2024-12-04 17:13:01,967][energy][INFO] - + prefill_iteration_7 energy consumption:
|
| 106 |
+
[PROC-0][2024-12-04 17:13:01,967][energy][INFO] - + CPU: 0.000000 (kWh)
|
| 107 |
+
[PROC-0][2024-12-04 17:13:01,967][energy][INFO] - + GPU: 0.000000 (kWh)
|
| 108 |
+
[PROC-0][2024-12-04 17:13:01,967][energy][INFO] - + RAM: 0.000000 (kWh)
|
| 109 |
+
[PROC-0][2024-12-04 17:13:01,968][energy][INFO] - + total: 0.000000 (kWh)
|
| 110 |
+
[PROC-0][2024-12-04 17:13:01,968][energy][INFO] - + prefill_iteration_8 energy consumption:
|
| 111 |
+
[PROC-0][2024-12-04 17:13:01,968][energy][INFO] - + CPU: 0.000342 (kWh)
|
| 112 |
+
[PROC-0][2024-12-04 17:13:01,968][energy][INFO] - + GPU: 0.002009 (kWh)
|
| 113 |
+
[PROC-0][2024-12-04 17:13:01,968][energy][INFO] - + RAM: 0.000004 (kWh)
|
| 114 |
+
[PROC-0][2024-12-04 17:13:01,968][energy][INFO] - + total: 0.002354 (kWh)
|
| 115 |
+
[PROC-0][2024-12-04 17:13:01,968][energy][INFO] - + prefill_iteration_9 energy consumption:
|
| 116 |
+
[PROC-0][2024-12-04 17:13:01,968][energy][INFO] - + CPU: 0.000342 (kWh)
|
| 117 |
+
[PROC-0][2024-12-04 17:13:01,968][energy][INFO] - + GPU: 0.002011 (kWh)
|
| 118 |
+
[PROC-0][2024-12-04 17:13:01,968][energy][INFO] - + RAM: 0.000004 (kWh)
|
| 119 |
+
[PROC-0][2024-12-04 17:13:01,968][energy][INFO] - + total: 0.002356 (kWh)
|
| 120 |
+
[PROC-0][2024-12-04 17:13:01,968][energy][INFO] - + prefill_iteration_10 energy consumption:
|
| 121 |
+
[PROC-0][2024-12-04 17:13:01,969][energy][INFO] - + CPU: 0.000343 (kWh)
|
| 122 |
+
[PROC-0][2024-12-04 17:13:01,969][energy][INFO] - + GPU: 0.002020 (kWh)
|
| 123 |
+
[PROC-0][2024-12-04 17:13:01,969][energy][INFO] - + RAM: 0.000004 (kWh)
|
| 124 |
+
[PROC-0][2024-12-04 17:13:01,969][energy][INFO] - + total: 0.002366 (kWh)
|
| 125 |
+
[PROC-0][2024-12-04 17:13:01,969][energy][INFO] - + decode energy consumption:
|
| 126 |
+
[PROC-0][2024-12-04 17:13:01,969][energy][INFO] - + CPU: 0.001575 (kWh)
|
| 127 |
+
[PROC-0][2024-12-04 17:13:01,969][energy][INFO] - + GPU: 0.003964 (kWh)
|
| 128 |
+
[PROC-0][2024-12-04 17:13:01,969][energy][INFO] - + RAM: 0.000016 (kWh)
|
| 129 |
+
[PROC-0][2024-12-04 17:13:01,969][energy][INFO] - + total: 0.005556 (kWh)
|
| 130 |
+
[PROC-0][2024-12-04 17:13:01,969][energy][INFO] - + decode_iteration_1 energy consumption:
|
| 131 |
+
[PROC-0][2024-12-04 17:13:01,969][energy][INFO] - + CPU: 0.001752 (kWh)
|
| 132 |
+
[PROC-0][2024-12-04 17:13:01,969][energy][INFO] - + GPU: 0.004435 (kWh)
|
| 133 |
+
[PROC-0][2024-12-04 17:13:01,969][energy][INFO] - + RAM: 0.000018 (kWh)
|
| 134 |
+
[PROC-0][2024-12-04 17:13:01,970][energy][INFO] - + total: 0.006205 (kWh)
|
| 135 |
+
[PROC-0][2024-12-04 17:13:01,970][energy][INFO] - + decode_iteration_2 energy consumption:
|
| 136 |
+
[PROC-0][2024-12-04 17:13:01,970][energy][INFO] - + CPU: 0.001745 (kWh)
|
| 137 |
+
[PROC-0][2024-12-04 17:13:01,970][energy][INFO] - + GPU: 0.004403 (kWh)
|
| 138 |
+
[PROC-0][2024-12-04 17:13:01,970][energy][INFO] - + RAM: 0.000018 (kWh)
|
| 139 |
+
[PROC-0][2024-12-04 17:13:01,970][energy][INFO] - + total: 0.006166 (kWh)
|
| 140 |
+
[PROC-0][2024-12-04 17:13:01,970][energy][INFO] - + decode_iteration_3 energy consumption:
|
| 141 |
+
[PROC-0][2024-12-04 17:13:01,970][energy][INFO] - + CPU: 0.001745 (kWh)
|
| 142 |
+
[PROC-0][2024-12-04 17:13:01,970][energy][INFO] - + GPU: 0.004412 (kWh)
|
| 143 |
+
[PROC-0][2024-12-04 17:13:01,970][energy][INFO] - + RAM: 0.000018 (kWh)
|
| 144 |
+
[PROC-0][2024-12-04 17:13:01,970][energy][INFO] - + total: 0.006175 (kWh)
|
| 145 |
+
[PROC-0][2024-12-04 17:13:01,970][energy][INFO] - + decode_iteration_4 energy consumption:
|
| 146 |
+
[PROC-0][2024-12-04 17:13:01,971][energy][INFO] - + CPU: 0.001755 (kWh)
|
| 147 |
+
[PROC-0][2024-12-04 17:13:01,971][energy][INFO] - + GPU: 0.004421 (kWh)
|
| 148 |
+
[PROC-0][2024-12-04 17:13:01,971][energy][INFO] - + RAM: 0.000018 (kWh)
|
| 149 |
+
[PROC-0][2024-12-04 17:13:01,971][energy][INFO] - + total: 0.006193 (kWh)
|
| 150 |
+
[PROC-0][2024-12-04 17:13:01,971][energy][INFO] - + decode_iteration_5 energy consumption:
|
| 151 |
+
[PROC-0][2024-12-04 17:13:01,971][energy][INFO] - + CPU: -0.000342 (kWh)
|
| 152 |
+
[PROC-0][2024-12-04 17:13:01,971][energy][INFO] - + GPU: -0.002011 (kWh)
|
| 153 |
+
[PROC-0][2024-12-04 17:13:01,971][energy][INFO] - + RAM: -0.000004 (kWh)
|
| 154 |
+
[PROC-0][2024-12-04 17:13:01,971][energy][INFO] - + total: -0.002357 (kWh)
|
| 155 |
+
[PROC-0][2024-12-04 17:13:01,971][energy][INFO] - + decode_iteration_6 energy consumption:
|
| 156 |
+
[PROC-0][2024-12-04 17:13:01,971][energy][INFO] - + CPU: 0.001745 (kWh)
|
| 157 |
+
[PROC-0][2024-12-04 17:13:01,971][energy][INFO] - + GPU: 0.004413 (kWh)
|
| 158 |
+
[PROC-0][2024-12-04 17:13:01,972][energy][INFO] - + RAM: 0.000018 (kWh)
|
| 159 |
+
[PROC-0][2024-12-04 17:13:01,972][energy][INFO] - + total: 0.006176 (kWh)
|
| 160 |
+
[PROC-0][2024-12-04 17:13:01,972][energy][INFO] - + decode_iteration_7 energy consumption:
|
| 161 |
+
[PROC-0][2024-12-04 17:13:01,972][energy][INFO] - + CPU: 0.002087 (kWh)
|
| 162 |
+
[PROC-0][2024-12-04 17:13:01,972][energy][INFO] - + GPU: 0.006413 (kWh)
|
| 163 |
+
[PROC-0][2024-12-04 17:13:01,972][energy][INFO] - + RAM: 0.000022 (kWh)
|
| 164 |
+
[PROC-0][2024-12-04 17:13:01,972][energy][INFO] - + total: 0.008522 (kWh)
|
| 165 |
+
[PROC-0][2024-12-04 17:13:01,972][energy][INFO] - + decode_iteration_8 energy consumption:
|
| 166 |
+
[PROC-0][2024-12-04 17:13:01,972][energy][INFO] - + CPU: 0.001747 (kWh)
|
| 167 |
+
[PROC-0][2024-12-04 17:13:01,972][energy][INFO] - + GPU: 0.004382 (kWh)
|
| 168 |
+
[PROC-0][2024-12-04 17:13:01,972][energy][INFO] - + RAM: 0.000018 (kWh)
|
| 169 |
+
[PROC-0][2024-12-04 17:13:01,972][energy][INFO] - + total: 0.006147 (kWh)
|
| 170 |
+
[PROC-0][2024-12-04 17:13:01,973][energy][INFO] - + decode_iteration_9 energy consumption:
|
| 171 |
+
[PROC-0][2024-12-04 17:13:01,973][energy][INFO] - + CPU: 0.001746 (kWh)
|
| 172 |
+
[PROC-0][2024-12-04 17:13:01,973][energy][INFO] - + GPU: 0.004372 (kWh)
|
| 173 |
+
[PROC-0][2024-12-04 17:13:01,973][energy][INFO] - + RAM: 0.000018 (kWh)
|
| 174 |
+
[PROC-0][2024-12-04 17:13:01,973][energy][INFO] - + total: 0.006137 (kWh)
|
| 175 |
+
[PROC-0][2024-12-04 17:13:01,973][energy][INFO] - + decode_iteration_10 energy consumption:
|
| 176 |
+
[PROC-0][2024-12-04 17:13:01,973][energy][INFO] - + CPU: 0.001772 (kWh)
|
| 177 |
+
[PROC-0][2024-12-04 17:13:01,973][energy][INFO] - + GPU: 0.004402 (kWh)
|
| 178 |
+
[PROC-0][2024-12-04 17:13:01,973][energy][INFO] - + RAM: 0.000018 (kWh)
|
| 179 |
+
[PROC-0][2024-12-04 17:13:01,973][energy][INFO] - + total: 0.006192 (kWh)
|
| 180 |
+
[PROC-0][2024-12-04 17:13:01,973][energy][INFO] - + preprocess energy consumption:
|
| 181 |
+
[PROC-0][2024-12-04 17:13:01,973][energy][INFO] - + CPU: 0.000012 (kWh)
|
| 182 |
+
[PROC-0][2024-12-04 17:13:01,974][energy][INFO] - + GPU: 0.000019 (kWh)
|
| 183 |
+
[PROC-0][2024-12-04 17:13:01,974][energy][INFO] - + RAM: 0.000000 (kWh)
|
| 184 |
+
[PROC-0][2024-12-04 17:13:01,974][energy][INFO] - + total: 0.000031 (kWh)
|
| 185 |
+
[PROC-0][2024-12-04 17:13:01,974][energy][INFO] - + prefill energy efficiency: 91065082.200265 (tokens/kWh)
|
| 186 |
+
[PROC-0][2024-12-04 17:13:01,974][energy][INFO] - + decode energy efficiency: 1619944.451836 (tokens/kWh)
|
| 187 |
+
[PROC-0][2024-12-04 17:13:01,974][energy][INFO] - + preprocess energy efficiency: 32468947.585393 (samples/kWh)
|
| 188 |
+
[2024-12-04 17:13:02,662][datasets][INFO] - PyTorch version 2.4.0 available.
|
text_generation/bigscience/bloomz-560m/2024-12-04-16-38-24/error.log
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
text_generation/bigscience/bloomz-560m/2024-12-04-16-38-24/experiment_config.json
ADDED
|
@@ -0,0 +1,110 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"experiment_name": "text_generation",
|
| 3 |
+
"backend": {
|
| 4 |
+
"name": "pytorch",
|
| 5 |
+
"version": "2.4.0",
|
| 6 |
+
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
|
| 7 |
+
"task": "text-generation",
|
| 8 |
+
"model": "bigscience/bloomz-560m",
|
| 9 |
+
"processor": "bigscience/bloomz-560m",
|
| 10 |
+
"library": "transformers",
|
| 11 |
+
"device": "cuda",
|
| 12 |
+
"device_ids": "0",
|
| 13 |
+
"seed": 42,
|
| 14 |
+
"inter_op_num_threads": null,
|
| 15 |
+
"intra_op_num_threads": null,
|
| 16 |
+
"hub_kwargs": {
|
| 17 |
+
"revision": "main",
|
| 18 |
+
"force_download": false,
|
| 19 |
+
"local_files_only": false,
|
| 20 |
+
"trust_remote_code": true
|
| 21 |
+
},
|
| 22 |
+
"no_weights": true,
|
| 23 |
+
"device_map": null,
|
| 24 |
+
"torch_dtype": null,
|
| 25 |
+
"amp_autocast": false,
|
| 26 |
+
"amp_dtype": null,
|
| 27 |
+
"eval_mode": true,
|
| 28 |
+
"to_bettertransformer": false,
|
| 29 |
+
"low_cpu_mem_usage": null,
|
| 30 |
+
"attn_implementation": null,
|
| 31 |
+
"cache_implementation": null,
|
| 32 |
+
"torch_compile": false,
|
| 33 |
+
"torch_compile_config": {},
|
| 34 |
+
"quantization_scheme": null,
|
| 35 |
+
"quantization_config": {},
|
| 36 |
+
"deepspeed_inference": false,
|
| 37 |
+
"deepspeed_inference_config": {},
|
| 38 |
+
"peft_type": null,
|
| 39 |
+
"peft_config": {}
|
| 40 |
+
},
|
| 41 |
+
"launcher": {
|
| 42 |
+
"name": "process",
|
| 43 |
+
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
|
| 44 |
+
"device_isolation": false,
|
| 45 |
+
"device_isolation_action": "warn",
|
| 46 |
+
"start_method": "spawn"
|
| 47 |
+
},
|
| 48 |
+
"benchmark": {
|
| 49 |
+
"name": "energy_star",
|
| 50 |
+
"_target_": "optimum_benchmark.benchmarks.energy_star.benchmark.EnergyStarBenchmark",
|
| 51 |
+
"dataset_name": "EnergyStarAI/text_generation",
|
| 52 |
+
"dataset_config": "",
|
| 53 |
+
"dataset_split": "train",
|
| 54 |
+
"num_samples": 1000,
|
| 55 |
+
"input_shapes": {
|
| 56 |
+
"batch_size": 1
|
| 57 |
+
},
|
| 58 |
+
"text_column_name": "text",
|
| 59 |
+
"truncation": true,
|
| 60 |
+
"max_length": -1,
|
| 61 |
+
"dataset_prefix1": "",
|
| 62 |
+
"dataset_prefix2": "",
|
| 63 |
+
"t5_task": "",
|
| 64 |
+
"image_column_name": "image",
|
| 65 |
+
"resize": false,
|
| 66 |
+
"question_column_name": "question",
|
| 67 |
+
"context_column_name": "context",
|
| 68 |
+
"sentence1_column_name": "sentence1",
|
| 69 |
+
"sentence2_column_name": "sentence2",
|
| 70 |
+
"audio_column_name": "audio",
|
| 71 |
+
"iterations": 10,
|
| 72 |
+
"warmup_runs": 10,
|
| 73 |
+
"energy": true,
|
| 74 |
+
"forward_kwargs": {},
|
| 75 |
+
"generate_kwargs": {
|
| 76 |
+
"max_new_tokens": 10,
|
| 77 |
+
"min_new_tokens": 10
|
| 78 |
+
},
|
| 79 |
+
"call_kwargs": {}
|
| 80 |
+
},
|
| 81 |
+
"environment": {
|
| 82 |
+
"cpu": " AMD EPYC 7R32",
|
| 83 |
+
"cpu_count": 48,
|
| 84 |
+
"cpu_ram_mb": 200472.73984,
|
| 85 |
+
"system": "Linux",
|
| 86 |
+
"machine": "x86_64",
|
| 87 |
+
"platform": "Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35",
|
| 88 |
+
"processor": "x86_64",
|
| 89 |
+
"python_version": "3.9.20",
|
| 90 |
+
"gpu": [
|
| 91 |
+
"NVIDIA A10G"
|
| 92 |
+
],
|
| 93 |
+
"gpu_count": 1,
|
| 94 |
+
"gpu_vram_mb": 24146608128,
|
| 95 |
+
"optimum_benchmark_version": "0.2.0",
|
| 96 |
+
"optimum_benchmark_commit": null,
|
| 97 |
+
"transformers_version": "4.44.0",
|
| 98 |
+
"transformers_commit": null,
|
| 99 |
+
"accelerate_version": "0.33.0",
|
| 100 |
+
"accelerate_commit": null,
|
| 101 |
+
"diffusers_version": "0.30.0",
|
| 102 |
+
"diffusers_commit": null,
|
| 103 |
+
"optimum_version": null,
|
| 104 |
+
"optimum_commit": null,
|
| 105 |
+
"timm_version": null,
|
| 106 |
+
"timm_commit": null,
|
| 107 |
+
"peft_version": null,
|
| 108 |
+
"peft_commit": null
|
| 109 |
+
}
|
| 110 |
+
}
|
text_generation/bigscience/bloomz-560m/2024-12-04-16-38-24/generate_codecarbon.json
ADDED
|
@@ -0,0 +1,33 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"timestamp": "2024-12-04T17:13:01",
|
| 3 |
+
"project_name": "codecarbon",
|
| 4 |
+
"run_id": "5b242a91-3197-41b3-9253-d34861798546",
|
| 5 |
+
"duration": -1733284414.1146047,
|
| 6 |
+
"emissions": 0.0031592918499545944,
|
| 7 |
+
"emissions_rate": 1.763256415979793e-05,
|
| 8 |
+
"cpu_power": 42.5,
|
| 9 |
+
"gpu_power": 129.02572657389157,
|
| 10 |
+
"ram_power": 0.4394016265869141,
|
| 11 |
+
"cpu_energy": 0.002115267668630582,
|
| 12 |
+
"gpu_energy": 0.006421513470539986,
|
| 13 |
+
"ram_energy": 2.186860785924658e-05,
|
| 14 |
+
"energy_consumed": 0.008558649747029815,
|
| 15 |
+
"country_name": "United States",
|
| 16 |
+
"country_iso_code": "USA",
|
| 17 |
+
"region": "virginia",
|
| 18 |
+
"cloud_provider": "",
|
| 19 |
+
"cloud_region": "",
|
| 20 |
+
"os": "Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35",
|
| 21 |
+
"python_version": "3.9.20",
|
| 22 |
+
"codecarbon_version": "2.5.1",
|
| 23 |
+
"cpu_count": 48,
|
| 24 |
+
"cpu_model": "AMD EPYC 7R32",
|
| 25 |
+
"gpu_count": 1,
|
| 26 |
+
"gpu_model": "1 x NVIDIA A10G",
|
| 27 |
+
"longitude": -77.4903,
|
| 28 |
+
"latitude": 39.0469,
|
| 29 |
+
"ram_total_size": 186.7047882080078,
|
| 30 |
+
"tracking_mode": "process",
|
| 31 |
+
"on_cloud": "N",
|
| 32 |
+
"pue": 1.0
|
| 33 |
+
}
|
text_generation/bigscience/bloomz-560m/2024-12-04-16-38-24/prefill_codecarbon.json
ADDED
|
@@ -0,0 +1,33 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"timestamp": "2024-12-04T16:43:29",
|
| 3 |
+
"project_name": "codecarbon",
|
| 4 |
+
"run_id": "5b242a91-3197-41b3-9253-d34861798546",
|
| 5 |
+
"duration": -1733284564.2325568,
|
| 6 |
+
"emissions": 0.0008734566222457878,
|
| 7 |
+
"emissions_rate": 3.006139127695831e-05,
|
| 8 |
+
"cpu_power": 42.5,
|
| 9 |
+
"gpu_power": 250.2334203013918,
|
| 10 |
+
"ram_power": 0.4391512870788574,
|
| 11 |
+
"cpu_energy": 0.00034304597653473214,
|
| 12 |
+
"gpu_energy": 0.0020196391157100146,
|
| 13 |
+
"ram_energy": 3.5443205690369786e-06,
|
| 14 |
+
"energy_consumed": 0.002366229412813784,
|
| 15 |
+
"country_name": "United States",
|
| 16 |
+
"country_iso_code": "USA",
|
| 17 |
+
"region": "virginia",
|
| 18 |
+
"cloud_provider": "",
|
| 19 |
+
"cloud_region": "",
|
| 20 |
+
"os": "Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35",
|
| 21 |
+
"python_version": "3.9.20",
|
| 22 |
+
"codecarbon_version": "2.5.1",
|
| 23 |
+
"cpu_count": 48,
|
| 24 |
+
"cpu_model": "AMD EPYC 7R32",
|
| 25 |
+
"gpu_count": 1,
|
| 26 |
+
"gpu_model": "1 x NVIDIA A10G",
|
| 27 |
+
"longitude": -77.4903,
|
| 28 |
+
"latitude": 39.0469,
|
| 29 |
+
"ram_total_size": 186.7047882080078,
|
| 30 |
+
"tracking_mode": "process",
|
| 31 |
+
"on_cloud": "N",
|
| 32 |
+
"pue": 1.0
|
| 33 |
+
}
|
text_generation/bigscience/bloomz-560m/2024-12-04-16-38-24/preprocess_codecarbon.json
ADDED
|
@@ -0,0 +1,33 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"timestamp": "2024-12-04T16:38:38",
|
| 3 |
+
"project_name": "codecarbon",
|
| 4 |
+
"run_id": "5b242a91-3197-41b3-9253-d34861798546",
|
| 5 |
+
"duration": -1733284592.2829382,
|
| 6 |
+
"emissions": 1.1368843320834963e-05,
|
| 7 |
+
"emissions_rate": 1.1307034794637876e-05,
|
| 8 |
+
"cpu_power": 42.5,
|
| 9 |
+
"gpu_power": 67.3536754430518,
|
| 10 |
+
"ram_power": 0.31451940536499023,
|
| 11 |
+
"cpu_energy": 1.189836183752454e-05,
|
| 12 |
+
"gpu_energy": 1.8812515049981826e-05,
|
| 13 |
+
"ram_energy": 8.778073359449114e-08,
|
| 14 |
+
"energy_consumed": 3.079865762110086e-05,
|
| 15 |
+
"country_name": "United States",
|
| 16 |
+
"country_iso_code": "USA",
|
| 17 |
+
"region": "virginia",
|
| 18 |
+
"cloud_provider": "",
|
| 19 |
+
"cloud_region": "",
|
| 20 |
+
"os": "Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35",
|
| 21 |
+
"python_version": "3.9.20",
|
| 22 |
+
"codecarbon_version": "2.5.1",
|
| 23 |
+
"cpu_count": 48,
|
| 24 |
+
"cpu_model": "AMD EPYC 7R32",
|
| 25 |
+
"gpu_count": 1,
|
| 26 |
+
"gpu_model": "1 x NVIDIA A10G",
|
| 27 |
+
"longitude": -77.4903,
|
| 28 |
+
"latitude": 39.0469,
|
| 29 |
+
"ram_total_size": 186.7047882080078,
|
| 30 |
+
"tracking_mode": "process",
|
| 31 |
+
"on_cloud": "N",
|
| 32 |
+
"pue": 1.0
|
| 33 |
+
}
|
text_generation/facebook/opt-1.3b/2024-12-04-15-35-29/.hydra/config.yaml
ADDED
|
@@ -0,0 +1,96 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
backend:
|
| 2 |
+
name: pytorch
|
| 3 |
+
version: 2.4.0
|
| 4 |
+
_target_: optimum_benchmark.backends.pytorch.backend.PyTorchBackend
|
| 5 |
+
task: text-generation
|
| 6 |
+
model: facebook/opt-1.3b
|
| 7 |
+
processor: facebook/opt-1.3b
|
| 8 |
+
library: null
|
| 9 |
+
device: cuda
|
| 10 |
+
device_ids: '0'
|
| 11 |
+
seed: 42
|
| 12 |
+
inter_op_num_threads: null
|
| 13 |
+
intra_op_num_threads: null
|
| 14 |
+
hub_kwargs: {}
|
| 15 |
+
no_weights: true
|
| 16 |
+
device_map: null
|
| 17 |
+
torch_dtype: null
|
| 18 |
+
amp_autocast: false
|
| 19 |
+
amp_dtype: null
|
| 20 |
+
eval_mode: true
|
| 21 |
+
to_bettertransformer: false
|
| 22 |
+
low_cpu_mem_usage: null
|
| 23 |
+
attn_implementation: null
|
| 24 |
+
cache_implementation: null
|
| 25 |
+
torch_compile: false
|
| 26 |
+
torch_compile_config: {}
|
| 27 |
+
quantization_scheme: null
|
| 28 |
+
quantization_config: {}
|
| 29 |
+
deepspeed_inference: false
|
| 30 |
+
deepspeed_inference_config: {}
|
| 31 |
+
peft_type: null
|
| 32 |
+
peft_config: {}
|
| 33 |
+
launcher:
|
| 34 |
+
name: process
|
| 35 |
+
_target_: optimum_benchmark.launchers.process.launcher.ProcessLauncher
|
| 36 |
+
device_isolation: false
|
| 37 |
+
device_isolation_action: warn
|
| 38 |
+
start_method: spawn
|
| 39 |
+
benchmark:
|
| 40 |
+
name: energy_star
|
| 41 |
+
_target_: optimum_benchmark.benchmarks.energy_star.benchmark.EnergyStarBenchmark
|
| 42 |
+
dataset_name: EnergyStarAI/text_generation
|
| 43 |
+
dataset_config: ''
|
| 44 |
+
dataset_split: train
|
| 45 |
+
num_samples: 1000
|
| 46 |
+
input_shapes:
|
| 47 |
+
batch_size: 1
|
| 48 |
+
text_column_name: text
|
| 49 |
+
truncation: true
|
| 50 |
+
max_length: -1
|
| 51 |
+
dataset_prefix1: ''
|
| 52 |
+
dataset_prefix2: ''
|
| 53 |
+
t5_task: ''
|
| 54 |
+
image_column_name: image
|
| 55 |
+
resize: false
|
| 56 |
+
question_column_name: question
|
| 57 |
+
context_column_name: context
|
| 58 |
+
sentence1_column_name: sentence1
|
| 59 |
+
sentence2_column_name: sentence2
|
| 60 |
+
audio_column_name: audio
|
| 61 |
+
iterations: 10
|
| 62 |
+
warmup_runs: 10
|
| 63 |
+
energy: true
|
| 64 |
+
forward_kwargs: {}
|
| 65 |
+
generate_kwargs:
|
| 66 |
+
max_new_tokens: 10
|
| 67 |
+
min_new_tokens: 10
|
| 68 |
+
call_kwargs: {}
|
| 69 |
+
experiment_name: text_generation
|
| 70 |
+
environment:
|
| 71 |
+
cpu: ' AMD EPYC 7R32'
|
| 72 |
+
cpu_count: 48
|
| 73 |
+
cpu_ram_mb: 200472.73984
|
| 74 |
+
system: Linux
|
| 75 |
+
machine: x86_64
|
| 76 |
+
platform: Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35
|
| 77 |
+
processor: x86_64
|
| 78 |
+
python_version: 3.9.20
|
| 79 |
+
gpu:
|
| 80 |
+
- NVIDIA A10G
|
| 81 |
+
gpu_count: 1
|
| 82 |
+
gpu_vram_mb: 24146608128
|
| 83 |
+
optimum_benchmark_version: 0.2.0
|
| 84 |
+
optimum_benchmark_commit: null
|
| 85 |
+
transformers_version: 4.44.0
|
| 86 |
+
transformers_commit: null
|
| 87 |
+
accelerate_version: 0.33.0
|
| 88 |
+
accelerate_commit: null
|
| 89 |
+
diffusers_version: 0.30.0
|
| 90 |
+
diffusers_commit: null
|
| 91 |
+
optimum_version: null
|
| 92 |
+
optimum_commit: null
|
| 93 |
+
timm_version: null
|
| 94 |
+
timm_commit: null
|
| 95 |
+
peft_version: null
|
| 96 |
+
peft_commit: null
|
text_generation/facebook/opt-1.3b/2024-12-04-15-35-29/.hydra/hydra.yaml
ADDED
|
@@ -0,0 +1,175 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
hydra:
|
| 2 |
+
run:
|
| 3 |
+
dir: /runs/text_generation/facebook/opt-1.3b/2024-12-04-15-35-29
|
| 4 |
+
sweep:
|
| 5 |
+
dir: sweeps/${experiment_name}/${backend.model}/${now:%Y-%m-%d-%H-%M-%S}
|
| 6 |
+
subdir: ${hydra.job.num}
|
| 7 |
+
launcher:
|
| 8 |
+
_target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
|
| 9 |
+
sweeper:
|
| 10 |
+
_target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
|
| 11 |
+
max_batch_size: null
|
| 12 |
+
params: null
|
| 13 |
+
help:
|
| 14 |
+
app_name: ${hydra.job.name}
|
| 15 |
+
header: '${hydra.help.app_name} is powered by Hydra.
|
| 16 |
+
|
| 17 |
+
'
|
| 18 |
+
footer: 'Powered by Hydra (https://hydra.cc)
|
| 19 |
+
|
| 20 |
+
Use --hydra-help to view Hydra specific help
|
| 21 |
+
|
| 22 |
+
'
|
| 23 |
+
template: '${hydra.help.header}
|
| 24 |
+
|
| 25 |
+
== Configuration groups ==
|
| 26 |
+
|
| 27 |
+
Compose your configuration from those groups (group=option)
|
| 28 |
+
|
| 29 |
+
|
| 30 |
+
$APP_CONFIG_GROUPS
|
| 31 |
+
|
| 32 |
+
|
| 33 |
+
== Config ==
|
| 34 |
+
|
| 35 |
+
Override anything in the config (foo.bar=value)
|
| 36 |
+
|
| 37 |
+
|
| 38 |
+
$CONFIG
|
| 39 |
+
|
| 40 |
+
|
| 41 |
+
${hydra.help.footer}
|
| 42 |
+
|
| 43 |
+
'
|
| 44 |
+
hydra_help:
|
| 45 |
+
template: 'Hydra (${hydra.runtime.version})
|
| 46 |
+
|
| 47 |
+
See https://hydra.cc for more info.
|
| 48 |
+
|
| 49 |
+
|
| 50 |
+
== Flags ==
|
| 51 |
+
|
| 52 |
+
$FLAGS_HELP
|
| 53 |
+
|
| 54 |
+
|
| 55 |
+
== Configuration groups ==
|
| 56 |
+
|
| 57 |
+
Compose your configuration from those groups (For example, append hydra/job_logging=disabled
|
| 58 |
+
to command line)
|
| 59 |
+
|
| 60 |
+
|
| 61 |
+
$HYDRA_CONFIG_GROUPS
|
| 62 |
+
|
| 63 |
+
|
| 64 |
+
Use ''--cfg hydra'' to Show the Hydra config.
|
| 65 |
+
|
| 66 |
+
'
|
| 67 |
+
hydra_help: ???
|
| 68 |
+
hydra_logging:
|
| 69 |
+
version: 1
|
| 70 |
+
formatters:
|
| 71 |
+
colorlog:
|
| 72 |
+
(): colorlog.ColoredFormatter
|
| 73 |
+
format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s'
|
| 74 |
+
handlers:
|
| 75 |
+
console:
|
| 76 |
+
class: logging.StreamHandler
|
| 77 |
+
formatter: colorlog
|
| 78 |
+
stream: ext://sys.stdout
|
| 79 |
+
root:
|
| 80 |
+
level: INFO
|
| 81 |
+
handlers:
|
| 82 |
+
- console
|
| 83 |
+
disable_existing_loggers: false
|
| 84 |
+
job_logging:
|
| 85 |
+
version: 1
|
| 86 |
+
formatters:
|
| 87 |
+
simple:
|
| 88 |
+
format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
|
| 89 |
+
colorlog:
|
| 90 |
+
(): colorlog.ColoredFormatter
|
| 91 |
+
format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s]
|
| 92 |
+
- %(message)s'
|
| 93 |
+
log_colors:
|
| 94 |
+
DEBUG: purple
|
| 95 |
+
INFO: green
|
| 96 |
+
WARNING: yellow
|
| 97 |
+
ERROR: red
|
| 98 |
+
CRITICAL: red
|
| 99 |
+
handlers:
|
| 100 |
+
console:
|
| 101 |
+
class: logging.StreamHandler
|
| 102 |
+
formatter: colorlog
|
| 103 |
+
stream: ext://sys.stdout
|
| 104 |
+
file:
|
| 105 |
+
class: logging.FileHandler
|
| 106 |
+
formatter: simple
|
| 107 |
+
filename: ${hydra.job.name}.log
|
| 108 |
+
root:
|
| 109 |
+
level: INFO
|
| 110 |
+
handlers:
|
| 111 |
+
- console
|
| 112 |
+
- file
|
| 113 |
+
disable_existing_loggers: false
|
| 114 |
+
env: {}
|
| 115 |
+
mode: RUN
|
| 116 |
+
searchpath: []
|
| 117 |
+
callbacks: {}
|
| 118 |
+
output_subdir: .hydra
|
| 119 |
+
overrides:
|
| 120 |
+
hydra:
|
| 121 |
+
- hydra.run.dir=/runs/text_generation/facebook/opt-1.3b/2024-12-04-15-35-29
|
| 122 |
+
- hydra.mode=RUN
|
| 123 |
+
task:
|
| 124 |
+
- backend.model=facebook/opt-1.3b
|
| 125 |
+
- backend.processor=facebook/opt-1.3b
|
| 126 |
+
job:
|
| 127 |
+
name: cli
|
| 128 |
+
chdir: true
|
| 129 |
+
override_dirname: backend.model=facebook/opt-1.3b,backend.processor=facebook/opt-1.3b
|
| 130 |
+
id: ???
|
| 131 |
+
num: ???
|
| 132 |
+
config_name: text_generation
|
| 133 |
+
env_set:
|
| 134 |
+
OVERRIDE_BENCHMARKS: '1'
|
| 135 |
+
env_copy: []
|
| 136 |
+
config:
|
| 137 |
+
override_dirname:
|
| 138 |
+
kv_sep: '='
|
| 139 |
+
item_sep: ','
|
| 140 |
+
exclude_keys: []
|
| 141 |
+
runtime:
|
| 142 |
+
version: 1.3.2
|
| 143 |
+
version_base: '1.3'
|
| 144 |
+
cwd: /
|
| 145 |
+
config_sources:
|
| 146 |
+
- path: hydra.conf
|
| 147 |
+
schema: pkg
|
| 148 |
+
provider: hydra
|
| 149 |
+
- path: optimum_benchmark
|
| 150 |
+
schema: pkg
|
| 151 |
+
provider: main
|
| 152 |
+
- path: hydra_plugins.hydra_colorlog.conf
|
| 153 |
+
schema: pkg
|
| 154 |
+
provider: hydra-colorlog
|
| 155 |
+
- path: /optimum-benchmark/examples/energy_star
|
| 156 |
+
schema: file
|
| 157 |
+
provider: command-line
|
| 158 |
+
- path: ''
|
| 159 |
+
schema: structured
|
| 160 |
+
provider: schema
|
| 161 |
+
output_dir: /runs/text_generation/facebook/opt-1.3b/2024-12-04-15-35-29
|
| 162 |
+
choices:
|
| 163 |
+
benchmark: energy_star
|
| 164 |
+
launcher: process
|
| 165 |
+
backend: pytorch
|
| 166 |
+
hydra/env: default
|
| 167 |
+
hydra/callbacks: null
|
| 168 |
+
hydra/job_logging: colorlog
|
| 169 |
+
hydra/hydra_logging: colorlog
|
| 170 |
+
hydra/hydra_help: default
|
| 171 |
+
hydra/help: default
|
| 172 |
+
hydra/sweeper: basic
|
| 173 |
+
hydra/launcher: basic
|
| 174 |
+
hydra/output: default
|
| 175 |
+
verbose: false
|
text_generation/facebook/opt-1.3b/2024-12-04-15-35-29/.hydra/overrides.yaml
ADDED
|
@@ -0,0 +1,2 @@
|
|
|
|
|
|
|
|
|
|
| 1 |
+
- backend.model=facebook/opt-1.3b
|
| 2 |
+
- backend.processor=facebook/opt-1.3b
|
text_generation/facebook/opt-1.3b/2024-12-04-15-35-29/benchmark_report.json
ADDED
|
@@ -0,0 +1,203 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"prefill": {
|
| 3 |
+
"memory": null,
|
| 4 |
+
"latency": null,
|
| 5 |
+
"throughput": null,
|
| 6 |
+
"energy": {
|
| 7 |
+
"unit": "kWh",
|
| 8 |
+
"cpu": 0.0008514253416934658,
|
| 9 |
+
"ram": 7.867963273444184e-06,
|
| 10 |
+
"gpu": 0.0057380391182053955,
|
| 11 |
+
"total": 0.006597332423172309
|
| 12 |
+
},
|
| 13 |
+
"efficiency": {
|
| 14 |
+
"unit": "tokens/kWh",
|
| 15 |
+
"value": 45599945.66036175
|
| 16 |
+
},
|
| 17 |
+
"measures": [
|
| 18 |
+
{
|
| 19 |
+
"unit": "kWh",
|
| 20 |
+
"cpu": 0.0009420541374506937,
|
| 21 |
+
"ram": 8.704100682712942e-06,
|
| 22 |
+
"gpu": 0.0063278192289179525,
|
| 23 |
+
"total": 0.007278577467051359
|
| 24 |
+
},
|
| 25 |
+
{
|
| 26 |
+
"unit": "kWh",
|
| 27 |
+
"cpu": 0.000944713068940953,
|
| 28 |
+
"ram": 8.730216552006652e-06,
|
| 29 |
+
"gpu": 0.006389016222320076,
|
| 30 |
+
"total": 0.007342459507813036
|
| 31 |
+
},
|
| 32 |
+
{
|
| 33 |
+
"unit": "kWh",
|
| 34 |
+
"cpu": 0.000945915997960396,
|
| 35 |
+
"ram": 8.74128147129218e-06,
|
| 36 |
+
"gpu": 0.0063994151195279625,
|
| 37 |
+
"total": 0.00735407239895965
|
| 38 |
+
},
|
| 39 |
+
{
|
| 40 |
+
"unit": "kWh",
|
| 41 |
+
"cpu": 0.0009464871018867345,
|
| 42 |
+
"ram": 8.745973796001767e-06,
|
| 43 |
+
"gpu": 0.006413208463895981,
|
| 44 |
+
"total": 0.007368441539578718
|
| 45 |
+
},
|
| 46 |
+
{
|
| 47 |
+
"unit": "kWh",
|
| 48 |
+
"cpu": 0.0009467335027201289,
|
| 49 |
+
"ram": 8.748670999265155e-06,
|
| 50 |
+
"gpu": 0.006408340404445989,
|
| 51 |
+
"total": 0.007363822578165386
|
| 52 |
+
},
|
| 53 |
+
{
|
| 54 |
+
"unit": "kWh",
|
| 55 |
+
"cpu": 0.0009467267385145986,
|
| 56 |
+
"ram": 8.748809905354579e-06,
|
| 57 |
+
"gpu": 0.006411111517774015,
|
| 58 |
+
"total": 0.007366587066193968
|
| 59 |
+
},
|
| 60 |
+
{
|
| 61 |
+
"unit": "kWh",
|
| 62 |
+
"cpu": 0.0,
|
| 63 |
+
"ram": 0.0,
|
| 64 |
+
"gpu": 0.0,
|
| 65 |
+
"total": 0.0
|
| 66 |
+
},
|
| 67 |
+
{
|
| 68 |
+
"unit": "kWh",
|
| 69 |
+
"cpu": 0.0009472711554750681,
|
| 70 |
+
"ram": 8.754110904909515e-06,
|
| 71 |
+
"gpu": 0.006341506184311951,
|
| 72 |
+
"total": 0.007297531450691931
|
| 73 |
+
},
|
| 74 |
+
{
|
| 75 |
+
"unit": "kWh",
|
| 76 |
+
"cpu": 0.0009472746114687711,
|
| 77 |
+
"ram": 8.754147384568123e-06,
|
| 78 |
+
"gpu": 0.0063449789648680754,
|
| 79 |
+
"total": 0.007301007723721421
|
| 80 |
+
},
|
| 81 |
+
{
|
| 82 |
+
"unit": "kWh",
|
| 83 |
+
"cpu": 0.0009470771025173134,
|
| 84 |
+
"ram": 8.75232103833093e-06,
|
| 85 |
+
"gpu": 0.006344995075991955,
|
| 86 |
+
"total": 0.007300824499547609
|
| 87 |
+
}
|
| 88 |
+
]
|
| 89 |
+
},
|
| 90 |
+
"decode": {
|
| 91 |
+
"memory": null,
|
| 92 |
+
"latency": null,
|
| 93 |
+
"throughput": null,
|
| 94 |
+
"energy": {
|
| 95 |
+
"unit": "kWh",
|
| 96 |
+
"cpu": 0.0014949766351097563,
|
| 97 |
+
"ram": 1.3839823376050235e-05,
|
| 98 |
+
"gpu": 0.005730950168089817,
|
| 99 |
+
"total": 0.007239766626575616
|
| 100 |
+
},
|
| 101 |
+
"efficiency": {
|
| 102 |
+
"unit": "tokens/kWh",
|
| 103 |
+
"value": 1243133.9937067789
|
| 104 |
+
},
|
| 105 |
+
"measures": [
|
| 106 |
+
{
|
| 107 |
+
"unit": "kWh",
|
| 108 |
+
"cpu": 0.0016573730981167152,
|
| 109 |
+
"ram": 1.5343569632938047e-05,
|
| 110 |
+
"gpu": 0.006264276122528112,
|
| 111 |
+
"total": 0.007936992790277769
|
| 112 |
+
},
|
| 113 |
+
{
|
| 114 |
+
"unit": "kWh",
|
| 115 |
+
"cpu": 0.0016591519767944645,
|
| 116 |
+
"ram": 1.535895363465579e-05,
|
| 117 |
+
"gpu": 0.00633461312324185,
|
| 118 |
+
"total": 0.008009124053670963
|
| 119 |
+
},
|
| 120 |
+
{
|
| 121 |
+
"unit": "kWh",
|
| 122 |
+
"cpu": 0.0016632256255590767,
|
| 123 |
+
"ram": 1.5396708280115472e-05,
|
| 124 |
+
"gpu": 0.006332735899518049,
|
| 125 |
+
"total": 0.008011358233357243
|
| 126 |
+
},
|
| 127 |
+
{
|
| 128 |
+
"unit": "kWh",
|
| 129 |
+
"cpu": 0.0016589466871556845,
|
| 130 |
+
"ram": 1.535750031033357e-05,
|
| 131 |
+
"gpu": 0.006323610892218057,
|
| 132 |
+
"total": 0.007997915079684073
|
| 133 |
+
},
|
| 134 |
+
{
|
| 135 |
+
"unit": "kWh",
|
| 136 |
+
"cpu": -0.0009467335027201289,
|
| 137 |
+
"ram": -8.748670999265155e-06,
|
| 138 |
+
"gpu": -0.006408340404445989,
|
| 139 |
+
"total": -0.007363822578165386
|
| 140 |
+
},
|
| 141 |
+
{
|
| 142 |
+
"unit": "kWh",
|
| 143 |
+
"cpu": 0.0016625658083840242,
|
| 144 |
+
"ram": 1.5391710184190226e-05,
|
| 145 |
+
"gpu": 0.006337731181291939,
|
| 146 |
+
"total": 0.008015688699860142
|
| 147 |
+
},
|
| 148 |
+
{
|
| 149 |
+
"unit": "kWh",
|
| 150 |
+
"cpu": 0.0026131794655145715,
|
| 151 |
+
"ram": 2.417648965006555e-05,
|
| 152 |
+
"gpu": 0.012788611619769963,
|
| 153 |
+
"total": 0.015425967574934585
|
| 154 |
+
},
|
| 155 |
+
{
|
| 156 |
+
"unit": "kWh",
|
| 157 |
+
"cpu": 0.0016600494504749718,
|
| 158 |
+
"ram": 1.536817153908214e-05,
|
| 159 |
+
"gpu": 0.0064490648814701235,
|
| 160 |
+
"total": 0.00812448250348418
|
| 161 |
+
},
|
| 162 |
+
{
|
| 163 |
+
"unit": "kWh",
|
| 164 |
+
"cpu": 0.001662433071380592,
|
| 165 |
+
"ram": 1.5390029146946412e-05,
|
| 166 |
+
"gpu": 0.006438650428693848,
|
| 167 |
+
"total": 0.008116473529221367
|
| 168 |
+
},
|
| 169 |
+
{
|
| 170 |
+
"unit": "kWh",
|
| 171 |
+
"cpu": 0.0016595746704375896,
|
| 172 |
+
"ram": 1.536377238144031e-05,
|
| 173 |
+
"gpu": 0.006448547936612203,
|
| 174 |
+
"total": 0.008123486379431222
|
| 175 |
+
}
|
| 176 |
+
]
|
| 177 |
+
},
|
| 178 |
+
"per_token": {
|
| 179 |
+
"memory": null,
|
| 180 |
+
"latency": null,
|
| 181 |
+
"throughput": null,
|
| 182 |
+
"energy": null,
|
| 183 |
+
"efficiency": null,
|
| 184 |
+
"measures": null
|
| 185 |
+
},
|
| 186 |
+
"preprocess": {
|
| 187 |
+
"memory": null,
|
| 188 |
+
"latency": null,
|
| 189 |
+
"throughput": null,
|
| 190 |
+
"energy": {
|
| 191 |
+
"unit": "kWh",
|
| 192 |
+
"cpu": 1.1007535735416243e-05,
|
| 193 |
+
"ram": 7.318336868205421e-08,
|
| 194 |
+
"gpu": 1.6262513009956958e-05,
|
| 195 |
+
"total": 2.7343232114055257e-05
|
| 196 |
+
},
|
| 197 |
+
"efficiency": {
|
| 198 |
+
"unit": "samples/kWh",
|
| 199 |
+
"value": 36572121.24114506
|
| 200 |
+
},
|
| 201 |
+
"measures": null
|
| 202 |
+
}
|
| 203 |
+
}
|
text_generation/facebook/opt-1.3b/2024-12-04-15-35-29/cli.log
ADDED
|
@@ -0,0 +1,188 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
[2024-12-04 15:35:31,908][launcher][INFO] - ََAllocating process launcher
|
| 2 |
+
[2024-12-04 15:35:31,908][process][INFO] - + Setting multiprocessing start method to spawn.
|
| 3 |
+
[2024-12-04 15:35:31,921][process][INFO] - + Launched benchmark in isolated process 426.
|
| 4 |
+
[PROC-0][2024-12-04 15:35:34,474][datasets][INFO] - PyTorch version 2.4.0 available.
|
| 5 |
+
[PROC-0][2024-12-04 15:35:35,364][backend][INFO] - َAllocating pytorch backend
|
| 6 |
+
[PROC-0][2024-12-04 15:35:35,365][backend][INFO] - + Setting random seed to 42
|
| 7 |
+
[PROC-0][2024-12-04 15:35:36,180][pytorch][INFO] - + Using AutoModel class AutoModelForCausalLM
|
| 8 |
+
[PROC-0][2024-12-04 15:35:36,181][pytorch][INFO] - + Creating backend temporary directory
|
| 9 |
+
[PROC-0][2024-12-04 15:35:36,181][pytorch][INFO] - + Loading model with random weights
|
| 10 |
+
[PROC-0][2024-12-04 15:35:36,181][pytorch][INFO] - + Creating no weights model
|
| 11 |
+
[PROC-0][2024-12-04 15:35:36,181][pytorch][INFO] - + Creating no weights model directory
|
| 12 |
+
[PROC-0][2024-12-04 15:35:36,181][pytorch][INFO] - + Creating no weights model state dict
|
| 13 |
+
[PROC-0][2024-12-04 15:35:36,201][pytorch][INFO] - + Saving no weights model safetensors
|
| 14 |
+
[PROC-0][2024-12-04 15:35:36,202][pytorch][INFO] - + Saving no weights model pretrained config
|
| 15 |
+
[PROC-0][2024-12-04 15:35:36,203][pytorch][INFO] - + Loading no weights AutoModel
|
| 16 |
+
[PROC-0][2024-12-04 15:35:36,203][pytorch][INFO] - + Loading model directly on device: cuda
|
| 17 |
+
[PROC-0][2024-12-04 15:35:36,402][pytorch][INFO] - + Turning on model's eval mode
|
| 18 |
+
[PROC-0][2024-12-04 15:35:36,409][benchmark][INFO] - Allocating energy_star benchmark
|
| 19 |
+
[PROC-0][2024-12-04 15:35:36,409][energy_star][INFO] - + Loading raw dataset
|
| 20 |
+
[PROC-0][2024-12-04 15:35:37,584][energy_star][INFO] - + Updating Text Generation kwargs with default values
|
| 21 |
+
[PROC-0][2024-12-04 15:35:37,584][energy_star][INFO] - + Initializing Text Generation report
|
| 22 |
+
[PROC-0][2024-12-04 15:35:37,584][energy][INFO] - + Tracking GPU energy on devices [0]
|
| 23 |
+
[PROC-0][2024-12-04 15:35:41,777][energy_star][INFO] - + Preprocessing dataset
|
| 24 |
+
[PROC-0][2024-12-04 15:35:42,710][energy][INFO] - + Saving codecarbon emission data to preprocess_codecarbon.json
|
| 25 |
+
[PROC-0][2024-12-04 15:35:42,711][energy_star][INFO] - + Preparing backend for Inference
|
| 26 |
+
[PROC-0][2024-12-04 15:35:42,711][energy_star][INFO] - + Initialising dataloader
|
| 27 |
+
[PROC-0][2024-12-04 15:35:42,711][energy_star][INFO] - + Warming up backend for Inference
|
| 28 |
+
[PROC-0][2024-12-04 15:35:43,974][energy_star][INFO] - + Additional warmup for Text Generation
|
| 29 |
+
[PROC-0][2024-12-04 15:35:44,172][energy_star][INFO] - + Running Text Generation energy tracking for 10 iterations
|
| 30 |
+
[PROC-0][2024-12-04 15:35:44,172][energy_star][INFO] - + Prefill iteration 1/10
|
| 31 |
+
[PROC-0][2024-12-04 15:37:03,971][energy][INFO] - + Saving codecarbon emission data to prefill_codecarbon.json
|
| 32 |
+
[PROC-0][2024-12-04 15:37:03,971][energy_star][INFO] - + Prefill iteration 2/10
|
| 33 |
+
[PROC-0][2024-12-04 15:38:23,994][energy][INFO] - + Saving codecarbon emission data to prefill_codecarbon.json
|
| 34 |
+
[PROC-0][2024-12-04 15:38:23,995][energy_star][INFO] - + Prefill iteration 3/10
|
| 35 |
+
[PROC-0][2024-12-04 15:39:44,120][energy][INFO] - + Saving codecarbon emission data to prefill_codecarbon.json
|
| 36 |
+
[PROC-0][2024-12-04 15:39:44,120][energy_star][INFO] - + Prefill iteration 4/10
|
| 37 |
+
[PROC-0][2024-12-04 15:41:04,294][energy][INFO] - + Saving codecarbon emission data to prefill_codecarbon.json
|
| 38 |
+
[PROC-0][2024-12-04 15:41:04,294][energy_star][INFO] - + Prefill iteration 5/10
|
| 39 |
+
[PROC-0][2024-12-04 15:42:24,489][energy][INFO] - + Saving codecarbon emission data to prefill_codecarbon.json
|
| 40 |
+
[PROC-0][2024-12-04 15:42:24,489][energy_star][INFO] - + Prefill iteration 6/10
|
| 41 |
+
[PROC-0][2024-12-04 15:43:44,683][energy][INFO] - + Saving codecarbon emission data to prefill_codecarbon.json
|
| 42 |
+
[PROC-0][2024-12-04 15:43:44,684][energy_star][INFO] - + Prefill iteration 7/10
|
| 43 |
+
[PROC-0][2024-12-04 15:45:04,899][energy][INFO] - + Saving codecarbon emission data to prefill_codecarbon.json
|
| 44 |
+
[PROC-0][2024-12-04 15:45:04,899][energy_star][INFO] - + Prefill iteration 8/10
|
| 45 |
+
[PROC-0][2024-12-04 15:46:25,139][energy][INFO] - + Saving codecarbon emission data to prefill_codecarbon.json
|
| 46 |
+
[PROC-0][2024-12-04 15:46:25,139][energy_star][INFO] - + Prefill iteration 9/10
|
| 47 |
+
[PROC-0][2024-12-04 15:47:45,380][energy][INFO] - + Saving codecarbon emission data to prefill_codecarbon.json
|
| 48 |
+
[PROC-0][2024-12-04 15:47:45,380][energy_star][INFO] - + Prefill iteration 10/10
|
| 49 |
+
[PROC-0][2024-12-04 15:49:05,604][energy][INFO] - + Saving codecarbon emission data to prefill_codecarbon.json
|
| 50 |
+
[PROC-0][2024-12-04 15:49:05,604][energy_star][INFO] - + Decoding iteration 1/10
|
| 51 |
+
[PROC-0][2024-12-04 15:52:45,792][energy][INFO] - + Saving codecarbon emission data to generate_codecarbon.json
|
| 52 |
+
[PROC-0][2024-12-04 15:52:45,792][energy_star][INFO] - + Decoding iteration 2/10
|
| 53 |
+
[PROC-0][2024-12-04 15:56:26,355][energy][INFO] - + Saving codecarbon emission data to generate_codecarbon.json
|
| 54 |
+
[PROC-0][2024-12-04 15:56:26,356][energy_star][INFO] - + Decoding iteration 3/10
|
| 55 |
+
[PROC-0][2024-12-04 16:00:07,366][energy][INFO] - + Saving codecarbon emission data to generate_codecarbon.json
|
| 56 |
+
[PROC-0][2024-12-04 16:00:07,366][energy_star][INFO] - + Decoding iteration 4/10
|
| 57 |
+
[PROC-0][2024-12-04 16:03:48,063][energy][INFO] - + Saving codecarbon emission data to generate_codecarbon.json
|
| 58 |
+
[PROC-0][2024-12-04 16:03:48,063][energy_star][INFO] - + Decoding iteration 5/10
|
| 59 |
+
[PROC-0][2024-12-04 16:07:28,762][energy][INFO] - + Saving codecarbon emission data to generate_codecarbon.json
|
| 60 |
+
[PROC-0][2024-12-04 16:07:28,762][energy_star][INFO] - + Decoding iteration 6/10
|
| 61 |
+
[PROC-0][2024-12-04 16:11:09,785][energy][INFO] - + Saving codecarbon emission data to generate_codecarbon.json
|
| 62 |
+
[PROC-0][2024-12-04 16:11:09,786][energy_star][INFO] - + Decoding iteration 7/10
|
| 63 |
+
[PROC-0][2024-12-04 16:14:51,138][energy][INFO] - + Saving codecarbon emission data to generate_codecarbon.json
|
| 64 |
+
[PROC-0][2024-12-04 16:14:51,138][energy_star][INFO] - + Decoding iteration 8/10
|
| 65 |
+
[PROC-0][2024-12-04 16:18:31,994][energy][INFO] - + Saving codecarbon emission data to generate_codecarbon.json
|
| 66 |
+
[PROC-0][2024-12-04 16:18:31,995][energy_star][INFO] - + Decoding iteration 9/10
|
| 67 |
+
[PROC-0][2024-12-04 16:22:13,053][energy][INFO] - + Saving codecarbon emission data to generate_codecarbon.json
|
| 68 |
+
[PROC-0][2024-12-04 16:22:13,053][energy_star][INFO] - + Decoding iteration 10/10
|
| 69 |
+
[PROC-0][2024-12-04 16:25:53,853][energy][INFO] - + Saving codecarbon emission data to generate_codecarbon.json
|
| 70 |
+
[PROC-0][2024-12-04 16:25:53,853][energy][INFO] - + prefill energy consumption:
|
| 71 |
+
[PROC-0][2024-12-04 16:25:53,854][energy][INFO] - + CPU: 0.000851 (kWh)
|
| 72 |
+
[PROC-0][2024-12-04 16:25:53,854][energy][INFO] - + GPU: 0.005738 (kWh)
|
| 73 |
+
[PROC-0][2024-12-04 16:25:53,854][energy][INFO] - + RAM: 0.000008 (kWh)
|
| 74 |
+
[PROC-0][2024-12-04 16:25:53,854][energy][INFO] - + total: 0.006597 (kWh)
|
| 75 |
+
[PROC-0][2024-12-04 16:25:53,854][energy][INFO] - + prefill_iteration_1 energy consumption:
|
| 76 |
+
[PROC-0][2024-12-04 16:25:53,854][energy][INFO] - + CPU: 0.000942 (kWh)
|
| 77 |
+
[PROC-0][2024-12-04 16:25:53,854][energy][INFO] - + GPU: 0.006328 (kWh)
|
| 78 |
+
[PROC-0][2024-12-04 16:25:53,854][energy][INFO] - + RAM: 0.000009 (kWh)
|
| 79 |
+
[PROC-0][2024-12-04 16:25:53,854][energy][INFO] - + total: 0.007279 (kWh)
|
| 80 |
+
[PROC-0][2024-12-04 16:25:53,854][energy][INFO] - + prefill_iteration_2 energy consumption:
|
| 81 |
+
[PROC-0][2024-12-04 16:25:53,854][energy][INFO] - + CPU: 0.000945 (kWh)
|
| 82 |
+
[PROC-0][2024-12-04 16:25:53,854][energy][INFO] - + GPU: 0.006389 (kWh)
|
| 83 |
+
[PROC-0][2024-12-04 16:25:53,855][energy][INFO] - + RAM: 0.000009 (kWh)
|
| 84 |
+
[PROC-0][2024-12-04 16:25:53,855][energy][INFO] - + total: 0.007342 (kWh)
|
| 85 |
+
[PROC-0][2024-12-04 16:25:53,855][energy][INFO] - + prefill_iteration_3 energy consumption:
|
| 86 |
+
[PROC-0][2024-12-04 16:25:53,855][energy][INFO] - + CPU: 0.000946 (kWh)
|
| 87 |
+
[PROC-0][2024-12-04 16:25:53,855][energy][INFO] - + GPU: 0.006399 (kWh)
|
| 88 |
+
[PROC-0][2024-12-04 16:25:53,855][energy][INFO] - + RAM: 0.000009 (kWh)
|
| 89 |
+
[PROC-0][2024-12-04 16:25:53,855][energy][INFO] - + total: 0.007354 (kWh)
|
| 90 |
+
[PROC-0][2024-12-04 16:25:53,855][energy][INFO] - + prefill_iteration_4 energy consumption:
|
| 91 |
+
[PROC-0][2024-12-04 16:25:53,855][energy][INFO] - + CPU: 0.000946 (kWh)
|
| 92 |
+
[PROC-0][2024-12-04 16:25:53,855][energy][INFO] - + GPU: 0.006413 (kWh)
|
| 93 |
+
[PROC-0][2024-12-04 16:25:53,855][energy][INFO] - + RAM: 0.000009 (kWh)
|
| 94 |
+
[PROC-0][2024-12-04 16:25:53,855][energy][INFO] - + total: 0.007368 (kWh)
|
| 95 |
+
[PROC-0][2024-12-04 16:25:53,856][energy][INFO] - + prefill_iteration_5 energy consumption:
|
| 96 |
+
[PROC-0][2024-12-04 16:25:53,856][energy][INFO] - + CPU: 0.000947 (kWh)
|
| 97 |
+
[PROC-0][2024-12-04 16:25:53,856][energy][INFO] - + GPU: 0.006408 (kWh)
|
| 98 |
+
[PROC-0][2024-12-04 16:25:53,856][energy][INFO] - + RAM: 0.000009 (kWh)
|
| 99 |
+
[PROC-0][2024-12-04 16:25:53,856][energy][INFO] - + total: 0.007364 (kWh)
|
| 100 |
+
[PROC-0][2024-12-04 16:25:53,856][energy][INFO] - + prefill_iteration_6 energy consumption:
|
| 101 |
+
[PROC-0][2024-12-04 16:25:53,856][energy][INFO] - + CPU: 0.000947 (kWh)
|
| 102 |
+
[PROC-0][2024-12-04 16:25:53,856][energy][INFO] - + GPU: 0.006411 (kWh)
|
| 103 |
+
[PROC-0][2024-12-04 16:25:53,856][energy][INFO] - + RAM: 0.000009 (kWh)
|
| 104 |
+
[PROC-0][2024-12-04 16:25:53,856][energy][INFO] - + total: 0.007367 (kWh)
|
| 105 |
+
[PROC-0][2024-12-04 16:25:53,856][energy][INFO] - + prefill_iteration_7 energy consumption:
|
| 106 |
+
[PROC-0][2024-12-04 16:25:53,856][energy][INFO] - + CPU: 0.000000 (kWh)
|
| 107 |
+
[PROC-0][2024-12-04 16:25:53,857][energy][INFO] - + GPU: 0.000000 (kWh)
|
| 108 |
+
[PROC-0][2024-12-04 16:25:53,857][energy][INFO] - + RAM: 0.000000 (kWh)
|
| 109 |
+
[PROC-0][2024-12-04 16:25:53,857][energy][INFO] - + total: 0.000000 (kWh)
|
| 110 |
+
[PROC-0][2024-12-04 16:25:53,857][energy][INFO] - + prefill_iteration_8 energy consumption:
|
| 111 |
+
[PROC-0][2024-12-04 16:25:53,857][energy][INFO] - + CPU: 0.000947 (kWh)
|
| 112 |
+
[PROC-0][2024-12-04 16:25:53,857][energy][INFO] - + GPU: 0.006342 (kWh)
|
| 113 |
+
[PROC-0][2024-12-04 16:25:53,857][energy][INFO] - + RAM: 0.000009 (kWh)
|
| 114 |
+
[PROC-0][2024-12-04 16:25:53,857][energy][INFO] - + total: 0.007298 (kWh)
|
| 115 |
+
[PROC-0][2024-12-04 16:25:53,857][energy][INFO] - + prefill_iteration_9 energy consumption:
|
| 116 |
+
[PROC-0][2024-12-04 16:25:53,857][energy][INFO] - + CPU: 0.000947 (kWh)
|
| 117 |
+
[PROC-0][2024-12-04 16:25:53,857][energy][INFO] - + GPU: 0.006345 (kWh)
|
| 118 |
+
[PROC-0][2024-12-04 16:25:53,857][energy][INFO] - + RAM: 0.000009 (kWh)
|
| 119 |
+
[PROC-0][2024-12-04 16:25:53,858][energy][INFO] - + total: 0.007301 (kWh)
|
| 120 |
+
[PROC-0][2024-12-04 16:25:53,858][energy][INFO] - + prefill_iteration_10 energy consumption:
|
| 121 |
+
[PROC-0][2024-12-04 16:25:53,858][energy][INFO] - + CPU: 0.000947 (kWh)
|
| 122 |
+
[PROC-0][2024-12-04 16:25:53,858][energy][INFO] - + GPU: 0.006345 (kWh)
|
| 123 |
+
[PROC-0][2024-12-04 16:25:53,858][energy][INFO] - + RAM: 0.000009 (kWh)
|
| 124 |
+
[PROC-0][2024-12-04 16:25:53,858][energy][INFO] - + total: 0.007301 (kWh)
|
| 125 |
+
[PROC-0][2024-12-04 16:25:53,858][energy][INFO] - + decode energy consumption:
|
| 126 |
+
[PROC-0][2024-12-04 16:25:53,858][energy][INFO] - + CPU: 0.001495 (kWh)
|
| 127 |
+
[PROC-0][2024-12-04 16:25:53,858][energy][INFO] - + GPU: 0.005731 (kWh)
|
| 128 |
+
[PROC-0][2024-12-04 16:25:53,858][energy][INFO] - + RAM: 0.000014 (kWh)
|
| 129 |
+
[PROC-0][2024-12-04 16:25:53,858][energy][INFO] - + total: 0.007240 (kWh)
|
| 130 |
+
[PROC-0][2024-12-04 16:25:53,858][energy][INFO] - + decode_iteration_1 energy consumption:
|
| 131 |
+
[PROC-0][2024-12-04 16:25:53,859][energy][INFO] - + CPU: 0.001657 (kWh)
|
| 132 |
+
[PROC-0][2024-12-04 16:25:53,859][energy][INFO] - + GPU: 0.006264 (kWh)
|
| 133 |
+
[PROC-0][2024-12-04 16:25:53,859][energy][INFO] - + RAM: 0.000015 (kWh)
|
| 134 |
+
[PROC-0][2024-12-04 16:25:53,859][energy][INFO] - + total: 0.007937 (kWh)
|
| 135 |
+
[PROC-0][2024-12-04 16:25:53,859][energy][INFO] - + decode_iteration_2 energy consumption:
|
| 136 |
+
[PROC-0][2024-12-04 16:25:53,859][energy][INFO] - + CPU: 0.001659 (kWh)
|
| 137 |
+
[PROC-0][2024-12-04 16:25:53,859][energy][INFO] - + GPU: 0.006335 (kWh)
|
| 138 |
+
[PROC-0][2024-12-04 16:25:53,859][energy][INFO] - + RAM: 0.000015 (kWh)
|
| 139 |
+
[PROC-0][2024-12-04 16:25:53,859][energy][INFO] - + total: 0.008009 (kWh)
|
| 140 |
+
[PROC-0][2024-12-04 16:25:53,859][energy][INFO] - + decode_iteration_3 energy consumption:
|
| 141 |
+
[PROC-0][2024-12-04 16:25:53,859][energy][INFO] - + CPU: 0.001663 (kWh)
|
| 142 |
+
[PROC-0][2024-12-04 16:25:53,859][energy][INFO] - + GPU: 0.006333 (kWh)
|
| 143 |
+
[PROC-0][2024-12-04 16:25:53,860][energy][INFO] - + RAM: 0.000015 (kWh)
|
| 144 |
+
[PROC-0][2024-12-04 16:25:53,860][energy][INFO] - + total: 0.008011 (kWh)
|
| 145 |
+
[PROC-0][2024-12-04 16:25:53,860][energy][INFO] - + decode_iteration_4 energy consumption:
|
| 146 |
+
[PROC-0][2024-12-04 16:25:53,860][energy][INFO] - + CPU: 0.001659 (kWh)
|
| 147 |
+
[PROC-0][2024-12-04 16:25:53,860][energy][INFO] - + GPU: 0.006324 (kWh)
|
| 148 |
+
[PROC-0][2024-12-04 16:25:53,860][energy][INFO] - + RAM: 0.000015 (kWh)
|
| 149 |
+
[PROC-0][2024-12-04 16:25:53,860][energy][INFO] - + total: 0.007998 (kWh)
|
| 150 |
+
[PROC-0][2024-12-04 16:25:53,860][energy][INFO] - + decode_iteration_5 energy consumption:
|
| 151 |
+
[PROC-0][2024-12-04 16:25:53,860][energy][INFO] - + CPU: -0.000947 (kWh)
|
| 152 |
+
[PROC-0][2024-12-04 16:25:53,860][energy][INFO] - + GPU: -0.006408 (kWh)
|
| 153 |
+
[PROC-0][2024-12-04 16:25:53,860][energy][INFO] - + RAM: -0.000009 (kWh)
|
| 154 |
+
[PROC-0][2024-12-04 16:25:53,860][energy][INFO] - + total: -0.007364 (kWh)
|
| 155 |
+
[PROC-0][2024-12-04 16:25:53,861][energy][INFO] - + decode_iteration_6 energy consumption:
|
| 156 |
+
[PROC-0][2024-12-04 16:25:53,861][energy][INFO] - + CPU: 0.001663 (kWh)
|
| 157 |
+
[PROC-0][2024-12-04 16:25:53,861][energy][INFO] - + GPU: 0.006338 (kWh)
|
| 158 |
+
[PROC-0][2024-12-04 16:25:53,861][energy][INFO] - + RAM: 0.000015 (kWh)
|
| 159 |
+
[PROC-0][2024-12-04 16:25:53,861][energy][INFO] - + total: 0.008016 (kWh)
|
| 160 |
+
[PROC-0][2024-12-04 16:25:53,861][energy][INFO] - + decode_iteration_7 energy consumption:
|
| 161 |
+
[PROC-0][2024-12-04 16:25:53,861][energy][INFO] - + CPU: 0.002613 (kWh)
|
| 162 |
+
[PROC-0][2024-12-04 16:25:53,861][energy][INFO] - + GPU: 0.012789 (kWh)
|
| 163 |
+
[PROC-0][2024-12-04 16:25:53,861][energy][INFO] - + RAM: 0.000024 (kWh)
|
| 164 |
+
[PROC-0][2024-12-04 16:25:53,861][energy][INFO] - + total: 0.015426 (kWh)
|
| 165 |
+
[PROC-0][2024-12-04 16:25:53,861][energy][INFO] - + decode_iteration_8 energy consumption:
|
| 166 |
+
[PROC-0][2024-12-04 16:25:53,861][energy][INFO] - + CPU: 0.001660 (kWh)
|
| 167 |
+
[PROC-0][2024-12-04 16:25:53,862][energy][INFO] - + GPU: 0.006449 (kWh)
|
| 168 |
+
[PROC-0][2024-12-04 16:25:53,862][energy][INFO] - + RAM: 0.000015 (kWh)
|
| 169 |
+
[PROC-0][2024-12-04 16:25:53,862][energy][INFO] - + total: 0.008124 (kWh)
|
| 170 |
+
[PROC-0][2024-12-04 16:25:53,862][energy][INFO] - + decode_iteration_9 energy consumption:
|
| 171 |
+
[PROC-0][2024-12-04 16:25:53,862][energy][INFO] - + CPU: 0.001662 (kWh)
|
| 172 |
+
[PROC-0][2024-12-04 16:25:53,862][energy][INFO] - + GPU: 0.006439 (kWh)
|
| 173 |
+
[PROC-0][2024-12-04 16:25:53,862][energy][INFO] - + RAM: 0.000015 (kWh)
|
| 174 |
+
[PROC-0][2024-12-04 16:25:53,862][energy][INFO] - + total: 0.008116 (kWh)
|
| 175 |
+
[PROC-0][2024-12-04 16:25:53,862][energy][INFO] - + decode_iteration_10 energy consumption:
|
| 176 |
+
[PROC-0][2024-12-04 16:25:53,862][energy][INFO] - + CPU: 0.001660 (kWh)
|
| 177 |
+
[PROC-0][2024-12-04 16:25:53,862][energy][INFO] - + GPU: 0.006449 (kWh)
|
| 178 |
+
[PROC-0][2024-12-04 16:25:53,862][energy][INFO] - + RAM: 0.000015 (kWh)
|
| 179 |
+
[PROC-0][2024-12-04 16:25:53,862][energy][INFO] - + total: 0.008123 (kWh)
|
| 180 |
+
[PROC-0][2024-12-04 16:25:53,863][energy][INFO] - + preprocess energy consumption:
|
| 181 |
+
[PROC-0][2024-12-04 16:25:53,863][energy][INFO] - + CPU: 0.000011 (kWh)
|
| 182 |
+
[PROC-0][2024-12-04 16:25:53,863][energy][INFO] - + GPU: 0.000016 (kWh)
|
| 183 |
+
[PROC-0][2024-12-04 16:25:53,863][energy][INFO] - + RAM: 0.000000 (kWh)
|
| 184 |
+
[PROC-0][2024-12-04 16:25:53,863][energy][INFO] - + total: 0.000027 (kWh)
|
| 185 |
+
[PROC-0][2024-12-04 16:25:53,863][energy][INFO] - + prefill energy efficiency: 45599945.660362 (tokens/kWh)
|
| 186 |
+
[PROC-0][2024-12-04 16:25:53,863][energy][INFO] - + decode energy efficiency: 1243133.993707 (tokens/kWh)
|
| 187 |
+
[PROC-0][2024-12-04 16:25:53,863][energy][INFO] - + preprocess energy efficiency: 36572121.241145 (samples/kWh)
|
| 188 |
+
[2024-12-04 16:25:54,519][datasets][INFO] - PyTorch version 2.4.0 available.
|
text_generation/facebook/opt-1.3b/2024-12-04-15-35-29/error.log
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
text_generation/facebook/opt-1.3b/2024-12-04-15-35-29/experiment_config.json
ADDED
|
@@ -0,0 +1,110 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"experiment_name": "text_generation",
|
| 3 |
+
"backend": {
|
| 4 |
+
"name": "pytorch",
|
| 5 |
+
"version": "2.4.0",
|
| 6 |
+
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
|
| 7 |
+
"task": "text-generation",
|
| 8 |
+
"model": "facebook/opt-1.3b",
|
| 9 |
+
"processor": "facebook/opt-1.3b",
|
| 10 |
+
"library": "transformers",
|
| 11 |
+
"device": "cuda",
|
| 12 |
+
"device_ids": "0",
|
| 13 |
+
"seed": 42,
|
| 14 |
+
"inter_op_num_threads": null,
|
| 15 |
+
"intra_op_num_threads": null,
|
| 16 |
+
"hub_kwargs": {
|
| 17 |
+
"revision": "main",
|
| 18 |
+
"force_download": false,
|
| 19 |
+
"local_files_only": false,
|
| 20 |
+
"trust_remote_code": true
|
| 21 |
+
},
|
| 22 |
+
"no_weights": true,
|
| 23 |
+
"device_map": null,
|
| 24 |
+
"torch_dtype": null,
|
| 25 |
+
"amp_autocast": false,
|
| 26 |
+
"amp_dtype": null,
|
| 27 |
+
"eval_mode": true,
|
| 28 |
+
"to_bettertransformer": false,
|
| 29 |
+
"low_cpu_mem_usage": null,
|
| 30 |
+
"attn_implementation": null,
|
| 31 |
+
"cache_implementation": null,
|
| 32 |
+
"torch_compile": false,
|
| 33 |
+
"torch_compile_config": {},
|
| 34 |
+
"quantization_scheme": null,
|
| 35 |
+
"quantization_config": {},
|
| 36 |
+
"deepspeed_inference": false,
|
| 37 |
+
"deepspeed_inference_config": {},
|
| 38 |
+
"peft_type": null,
|
| 39 |
+
"peft_config": {}
|
| 40 |
+
},
|
| 41 |
+
"launcher": {
|
| 42 |
+
"name": "process",
|
| 43 |
+
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
|
| 44 |
+
"device_isolation": false,
|
| 45 |
+
"device_isolation_action": "warn",
|
| 46 |
+
"start_method": "spawn"
|
| 47 |
+
},
|
| 48 |
+
"benchmark": {
|
| 49 |
+
"name": "energy_star",
|
| 50 |
+
"_target_": "optimum_benchmark.benchmarks.energy_star.benchmark.EnergyStarBenchmark",
|
| 51 |
+
"dataset_name": "EnergyStarAI/text_generation",
|
| 52 |
+
"dataset_config": "",
|
| 53 |
+
"dataset_split": "train",
|
| 54 |
+
"num_samples": 1000,
|
| 55 |
+
"input_shapes": {
|
| 56 |
+
"batch_size": 1
|
| 57 |
+
},
|
| 58 |
+
"text_column_name": "text",
|
| 59 |
+
"truncation": true,
|
| 60 |
+
"max_length": -1,
|
| 61 |
+
"dataset_prefix1": "",
|
| 62 |
+
"dataset_prefix2": "",
|
| 63 |
+
"t5_task": "",
|
| 64 |
+
"image_column_name": "image",
|
| 65 |
+
"resize": false,
|
| 66 |
+
"question_column_name": "question",
|
| 67 |
+
"context_column_name": "context",
|
| 68 |
+
"sentence1_column_name": "sentence1",
|
| 69 |
+
"sentence2_column_name": "sentence2",
|
| 70 |
+
"audio_column_name": "audio",
|
| 71 |
+
"iterations": 10,
|
| 72 |
+
"warmup_runs": 10,
|
| 73 |
+
"energy": true,
|
| 74 |
+
"forward_kwargs": {},
|
| 75 |
+
"generate_kwargs": {
|
| 76 |
+
"max_new_tokens": 10,
|
| 77 |
+
"min_new_tokens": 10
|
| 78 |
+
},
|
| 79 |
+
"call_kwargs": {}
|
| 80 |
+
},
|
| 81 |
+
"environment": {
|
| 82 |
+
"cpu": " AMD EPYC 7R32",
|
| 83 |
+
"cpu_count": 48,
|
| 84 |
+
"cpu_ram_mb": 200472.73984,
|
| 85 |
+
"system": "Linux",
|
| 86 |
+
"machine": "x86_64",
|
| 87 |
+
"platform": "Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35",
|
| 88 |
+
"processor": "x86_64",
|
| 89 |
+
"python_version": "3.9.20",
|
| 90 |
+
"gpu": [
|
| 91 |
+
"NVIDIA A10G"
|
| 92 |
+
],
|
| 93 |
+
"gpu_count": 1,
|
| 94 |
+
"gpu_vram_mb": 24146608128,
|
| 95 |
+
"optimum_benchmark_version": "0.2.0",
|
| 96 |
+
"optimum_benchmark_commit": null,
|
| 97 |
+
"transformers_version": "4.44.0",
|
| 98 |
+
"transformers_commit": null,
|
| 99 |
+
"accelerate_version": "0.33.0",
|
| 100 |
+
"accelerate_commit": null,
|
| 101 |
+
"diffusers_version": "0.30.0",
|
| 102 |
+
"diffusers_commit": null,
|
| 103 |
+
"optimum_version": null,
|
| 104 |
+
"optimum_commit": null,
|
| 105 |
+
"timm_version": null,
|
| 106 |
+
"timm_commit": null,
|
| 107 |
+
"peft_version": null,
|
| 108 |
+
"peft_commit": null
|
| 109 |
+
}
|
| 110 |
+
}
|
text_generation/facebook/opt-1.3b/2024-12-04-15-35-29/generate_codecarbon.json
ADDED
|
@@ -0,0 +1,33 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"timestamp": "2024-12-04T16:25:53",
|
| 3 |
+
"project_name": "codecarbon",
|
| 4 |
+
"run_id": "11b39f4a-102b-4ccd-a4e2-10220e0cac0d",
|
| 5 |
+
"duration": -1733284372.491492,
|
| 6 |
+
"emissions": 0.005693643400705231,
|
| 7 |
+
"emissions_rate": 2.5786799656597985e-05,
|
| 8 |
+
"cpu_power": 42.5,
|
| 9 |
+
"gpu_power": 208.5937464054236,
|
| 10 |
+
"ram_power": 0.3932046890258789,
|
| 11 |
+
"cpu_energy": 0.002606651772954903,
|
| 12 |
+
"gpu_energy": 0.012793543012604158,
|
| 13 |
+
"ram_energy": 2.4116093419771238e-05,
|
| 14 |
+
"energy_consumed": 0.01542431087897883,
|
| 15 |
+
"country_name": "United States",
|
| 16 |
+
"country_iso_code": "USA",
|
| 17 |
+
"region": "virginia",
|
| 18 |
+
"cloud_provider": "",
|
| 19 |
+
"cloud_region": "",
|
| 20 |
+
"os": "Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35",
|
| 21 |
+
"python_version": "3.9.20",
|
| 22 |
+
"codecarbon_version": "2.5.1",
|
| 23 |
+
"cpu_count": 48,
|
| 24 |
+
"cpu_model": "AMD EPYC 7R32",
|
| 25 |
+
"gpu_count": 1,
|
| 26 |
+
"gpu_model": "1 x NVIDIA A10G",
|
| 27 |
+
"longitude": -77.4903,
|
| 28 |
+
"latitude": 39.0469,
|
| 29 |
+
"ram_total_size": 186.7047882080078,
|
| 30 |
+
"tracking_mode": "process",
|
| 31 |
+
"on_cloud": "N",
|
| 32 |
+
"pue": 1.0
|
| 33 |
+
}
|
text_generation/facebook/opt-1.3b/2024-12-04-15-35-29/prefill_codecarbon.json
ADDED
|
@@ -0,0 +1,33 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"timestamp": "2024-12-04T15:49:05",
|
| 3 |
+
"project_name": "codecarbon",
|
| 4 |
+
"run_id": "11b39f4a-102b-4ccd-a4e2-10220e0cac0d",
|
| 5 |
+
"duration": -1733284513.0719962,
|
| 6 |
+
"emissions": 0.002694985309729978,
|
| 7 |
+
"emissions_rate": 3.3596470904941724e-05,
|
| 8 |
+
"cpu_power": 42.5,
|
| 9 |
+
"gpu_power": 284.73932067551607,
|
| 10 |
+
"ram_power": 0.3927741050720215,
|
| 11 |
+
"cpu_energy": 0.0009470771025173134,
|
| 12 |
+
"gpu_energy": 0.006344995075991955,
|
| 13 |
+
"ram_energy": 8.75232103833093e-06,
|
| 14 |
+
"energy_consumed": 0.007300824499547609,
|
| 15 |
+
"country_name": "United States",
|
| 16 |
+
"country_iso_code": "USA",
|
| 17 |
+
"region": "virginia",
|
| 18 |
+
"cloud_provider": "",
|
| 19 |
+
"cloud_region": "",
|
| 20 |
+
"os": "Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35",
|
| 21 |
+
"python_version": "3.9.20",
|
| 22 |
+
"codecarbon_version": "2.5.1",
|
| 23 |
+
"cpu_count": 48,
|
| 24 |
+
"cpu_model": "AMD EPYC 7R32",
|
| 25 |
+
"gpu_count": 1,
|
| 26 |
+
"gpu_model": "1 x NVIDIA A10G",
|
| 27 |
+
"longitude": -77.4903,
|
| 28 |
+
"latitude": 39.0469,
|
| 29 |
+
"ram_total_size": 186.7047882080078,
|
| 30 |
+
"tracking_mode": "process",
|
| 31 |
+
"on_cloud": "N",
|
| 32 |
+
"pue": 1.0
|
| 33 |
+
}
|
text_generation/facebook/opt-1.3b/2024-12-04-15-35-29/preprocess_codecarbon.json
ADDED
|
@@ -0,0 +1,33 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"timestamp": "2024-12-04T15:35:42",
|
| 3 |
+
"project_name": "codecarbon",
|
| 4 |
+
"run_id": "11b39f4a-102b-4ccd-a4e2-10220e0cac0d",
|
| 5 |
+
"duration": -1733284592.358094,
|
| 6 |
+
"emissions": 1.0093326975943891e-05,
|
| 7 |
+
"emissions_rate": 1.0849407214254869e-05,
|
| 8 |
+
"cpu_power": 42.5,
|
| 9 |
+
"gpu_power": 62.94082925267726,
|
| 10 |
+
"ram_power": 0.283480167388916,
|
| 11 |
+
"cpu_energy": 1.1007535735416243e-05,
|
| 12 |
+
"gpu_energy": 1.6262513009956958e-05,
|
| 13 |
+
"ram_energy": 7.318336868205421e-08,
|
| 14 |
+
"energy_consumed": 2.7343232114055257e-05,
|
| 15 |
+
"country_name": "United States",
|
| 16 |
+
"country_iso_code": "USA",
|
| 17 |
+
"region": "virginia",
|
| 18 |
+
"cloud_provider": "",
|
| 19 |
+
"cloud_region": "",
|
| 20 |
+
"os": "Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35",
|
| 21 |
+
"python_version": "3.9.20",
|
| 22 |
+
"codecarbon_version": "2.5.1",
|
| 23 |
+
"cpu_count": 48,
|
| 24 |
+
"cpu_model": "AMD EPYC 7R32",
|
| 25 |
+
"gpu_count": 1,
|
| 26 |
+
"gpu_model": "1 x NVIDIA A10G",
|
| 27 |
+
"longitude": -77.4903,
|
| 28 |
+
"latitude": 39.0469,
|
| 29 |
+
"ram_total_size": 186.7047882080078,
|
| 30 |
+
"tracking_mode": "process",
|
| 31 |
+
"on_cloud": "N",
|
| 32 |
+
"pue": 1.0
|
| 33 |
+
}
|
text_generation/openai-community/gpt2/2024-12-04-16-25-55/.hydra/config.yaml
ADDED
|
@@ -0,0 +1,96 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
backend:
|
| 2 |
+
name: pytorch
|
| 3 |
+
version: 2.4.0
|
| 4 |
+
_target_: optimum_benchmark.backends.pytorch.backend.PyTorchBackend
|
| 5 |
+
task: text-generation
|
| 6 |
+
model: openai-community/gpt2
|
| 7 |
+
processor: openai-community/gpt2
|
| 8 |
+
library: null
|
| 9 |
+
device: cuda
|
| 10 |
+
device_ids: '0'
|
| 11 |
+
seed: 42
|
| 12 |
+
inter_op_num_threads: null
|
| 13 |
+
intra_op_num_threads: null
|
| 14 |
+
hub_kwargs: {}
|
| 15 |
+
no_weights: true
|
| 16 |
+
device_map: null
|
| 17 |
+
torch_dtype: null
|
| 18 |
+
amp_autocast: false
|
| 19 |
+
amp_dtype: null
|
| 20 |
+
eval_mode: true
|
| 21 |
+
to_bettertransformer: false
|
| 22 |
+
low_cpu_mem_usage: null
|
| 23 |
+
attn_implementation: null
|
| 24 |
+
cache_implementation: null
|
| 25 |
+
torch_compile: false
|
| 26 |
+
torch_compile_config: {}
|
| 27 |
+
quantization_scheme: null
|
| 28 |
+
quantization_config: {}
|
| 29 |
+
deepspeed_inference: false
|
| 30 |
+
deepspeed_inference_config: {}
|
| 31 |
+
peft_type: null
|
| 32 |
+
peft_config: {}
|
| 33 |
+
launcher:
|
| 34 |
+
name: process
|
| 35 |
+
_target_: optimum_benchmark.launchers.process.launcher.ProcessLauncher
|
| 36 |
+
device_isolation: false
|
| 37 |
+
device_isolation_action: warn
|
| 38 |
+
start_method: spawn
|
| 39 |
+
benchmark:
|
| 40 |
+
name: energy_star
|
| 41 |
+
_target_: optimum_benchmark.benchmarks.energy_star.benchmark.EnergyStarBenchmark
|
| 42 |
+
dataset_name: EnergyStarAI/text_generation
|
| 43 |
+
dataset_config: ''
|
| 44 |
+
dataset_split: train
|
| 45 |
+
num_samples: 1000
|
| 46 |
+
input_shapes:
|
| 47 |
+
batch_size: 1
|
| 48 |
+
text_column_name: text
|
| 49 |
+
truncation: true
|
| 50 |
+
max_length: -1
|
| 51 |
+
dataset_prefix1: ''
|
| 52 |
+
dataset_prefix2: ''
|
| 53 |
+
t5_task: ''
|
| 54 |
+
image_column_name: image
|
| 55 |
+
resize: false
|
| 56 |
+
question_column_name: question
|
| 57 |
+
context_column_name: context
|
| 58 |
+
sentence1_column_name: sentence1
|
| 59 |
+
sentence2_column_name: sentence2
|
| 60 |
+
audio_column_name: audio
|
| 61 |
+
iterations: 10
|
| 62 |
+
warmup_runs: 10
|
| 63 |
+
energy: true
|
| 64 |
+
forward_kwargs: {}
|
| 65 |
+
generate_kwargs:
|
| 66 |
+
max_new_tokens: 10
|
| 67 |
+
min_new_tokens: 10
|
| 68 |
+
call_kwargs: {}
|
| 69 |
+
experiment_name: text_generation
|
| 70 |
+
environment:
|
| 71 |
+
cpu: ' AMD EPYC 7R32'
|
| 72 |
+
cpu_count: 48
|
| 73 |
+
cpu_ram_mb: 200472.73984
|
| 74 |
+
system: Linux
|
| 75 |
+
machine: x86_64
|
| 76 |
+
platform: Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35
|
| 77 |
+
processor: x86_64
|
| 78 |
+
python_version: 3.9.20
|
| 79 |
+
gpu:
|
| 80 |
+
- NVIDIA A10G
|
| 81 |
+
gpu_count: 1
|
| 82 |
+
gpu_vram_mb: 24146608128
|
| 83 |
+
optimum_benchmark_version: 0.2.0
|
| 84 |
+
optimum_benchmark_commit: null
|
| 85 |
+
transformers_version: 4.44.0
|
| 86 |
+
transformers_commit: null
|
| 87 |
+
accelerate_version: 0.33.0
|
| 88 |
+
accelerate_commit: null
|
| 89 |
+
diffusers_version: 0.30.0
|
| 90 |
+
diffusers_commit: null
|
| 91 |
+
optimum_version: null
|
| 92 |
+
optimum_commit: null
|
| 93 |
+
timm_version: null
|
| 94 |
+
timm_commit: null
|
| 95 |
+
peft_version: null
|
| 96 |
+
peft_commit: null
|
text_generation/openai-community/gpt2/2024-12-04-16-25-55/.hydra/hydra.yaml
ADDED
|
@@ -0,0 +1,175 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
hydra:
|
| 2 |
+
run:
|
| 3 |
+
dir: /runs/text_generation/openai-community/gpt2/2024-12-04-16-25-55
|
| 4 |
+
sweep:
|
| 5 |
+
dir: sweeps/${experiment_name}/${backend.model}/${now:%Y-%m-%d-%H-%M-%S}
|
| 6 |
+
subdir: ${hydra.job.num}
|
| 7 |
+
launcher:
|
| 8 |
+
_target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
|
| 9 |
+
sweeper:
|
| 10 |
+
_target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
|
| 11 |
+
max_batch_size: null
|
| 12 |
+
params: null
|
| 13 |
+
help:
|
| 14 |
+
app_name: ${hydra.job.name}
|
| 15 |
+
header: '${hydra.help.app_name} is powered by Hydra.
|
| 16 |
+
|
| 17 |
+
'
|
| 18 |
+
footer: 'Powered by Hydra (https://hydra.cc)
|
| 19 |
+
|
| 20 |
+
Use --hydra-help to view Hydra specific help
|
| 21 |
+
|
| 22 |
+
'
|
| 23 |
+
template: '${hydra.help.header}
|
| 24 |
+
|
| 25 |
+
== Configuration groups ==
|
| 26 |
+
|
| 27 |
+
Compose your configuration from those groups (group=option)
|
| 28 |
+
|
| 29 |
+
|
| 30 |
+
$APP_CONFIG_GROUPS
|
| 31 |
+
|
| 32 |
+
|
| 33 |
+
== Config ==
|
| 34 |
+
|
| 35 |
+
Override anything in the config (foo.bar=value)
|
| 36 |
+
|
| 37 |
+
|
| 38 |
+
$CONFIG
|
| 39 |
+
|
| 40 |
+
|
| 41 |
+
${hydra.help.footer}
|
| 42 |
+
|
| 43 |
+
'
|
| 44 |
+
hydra_help:
|
| 45 |
+
template: 'Hydra (${hydra.runtime.version})
|
| 46 |
+
|
| 47 |
+
See https://hydra.cc for more info.
|
| 48 |
+
|
| 49 |
+
|
| 50 |
+
== Flags ==
|
| 51 |
+
|
| 52 |
+
$FLAGS_HELP
|
| 53 |
+
|
| 54 |
+
|
| 55 |
+
== Configuration groups ==
|
| 56 |
+
|
| 57 |
+
Compose your configuration from those groups (For example, append hydra/job_logging=disabled
|
| 58 |
+
to command line)
|
| 59 |
+
|
| 60 |
+
|
| 61 |
+
$HYDRA_CONFIG_GROUPS
|
| 62 |
+
|
| 63 |
+
|
| 64 |
+
Use ''--cfg hydra'' to Show the Hydra config.
|
| 65 |
+
|
| 66 |
+
'
|
| 67 |
+
hydra_help: ???
|
| 68 |
+
hydra_logging:
|
| 69 |
+
version: 1
|
| 70 |
+
formatters:
|
| 71 |
+
colorlog:
|
| 72 |
+
(): colorlog.ColoredFormatter
|
| 73 |
+
format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s'
|
| 74 |
+
handlers:
|
| 75 |
+
console:
|
| 76 |
+
class: logging.StreamHandler
|
| 77 |
+
formatter: colorlog
|
| 78 |
+
stream: ext://sys.stdout
|
| 79 |
+
root:
|
| 80 |
+
level: INFO
|
| 81 |
+
handlers:
|
| 82 |
+
- console
|
| 83 |
+
disable_existing_loggers: false
|
| 84 |
+
job_logging:
|
| 85 |
+
version: 1
|
| 86 |
+
formatters:
|
| 87 |
+
simple:
|
| 88 |
+
format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
|
| 89 |
+
colorlog:
|
| 90 |
+
(): colorlog.ColoredFormatter
|
| 91 |
+
format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s]
|
| 92 |
+
- %(message)s'
|
| 93 |
+
log_colors:
|
| 94 |
+
DEBUG: purple
|
| 95 |
+
INFO: green
|
| 96 |
+
WARNING: yellow
|
| 97 |
+
ERROR: red
|
| 98 |
+
CRITICAL: red
|
| 99 |
+
handlers:
|
| 100 |
+
console:
|
| 101 |
+
class: logging.StreamHandler
|
| 102 |
+
formatter: colorlog
|
| 103 |
+
stream: ext://sys.stdout
|
| 104 |
+
file:
|
| 105 |
+
class: logging.FileHandler
|
| 106 |
+
formatter: simple
|
| 107 |
+
filename: ${hydra.job.name}.log
|
| 108 |
+
root:
|
| 109 |
+
level: INFO
|
| 110 |
+
handlers:
|
| 111 |
+
- console
|
| 112 |
+
- file
|
| 113 |
+
disable_existing_loggers: false
|
| 114 |
+
env: {}
|
| 115 |
+
mode: RUN
|
| 116 |
+
searchpath: []
|
| 117 |
+
callbacks: {}
|
| 118 |
+
output_subdir: .hydra
|
| 119 |
+
overrides:
|
| 120 |
+
hydra:
|
| 121 |
+
- hydra.run.dir=/runs/text_generation/openai-community/gpt2/2024-12-04-16-25-55
|
| 122 |
+
- hydra.mode=RUN
|
| 123 |
+
task:
|
| 124 |
+
- backend.model=openai-community/gpt2
|
| 125 |
+
- backend.processor=openai-community/gpt2
|
| 126 |
+
job:
|
| 127 |
+
name: cli
|
| 128 |
+
chdir: true
|
| 129 |
+
override_dirname: backend.model=openai-community/gpt2,backend.processor=openai-community/gpt2
|
| 130 |
+
id: ???
|
| 131 |
+
num: ???
|
| 132 |
+
config_name: text_generation
|
| 133 |
+
env_set:
|
| 134 |
+
OVERRIDE_BENCHMARKS: '1'
|
| 135 |
+
env_copy: []
|
| 136 |
+
config:
|
| 137 |
+
override_dirname:
|
| 138 |
+
kv_sep: '='
|
| 139 |
+
item_sep: ','
|
| 140 |
+
exclude_keys: []
|
| 141 |
+
runtime:
|
| 142 |
+
version: 1.3.2
|
| 143 |
+
version_base: '1.3'
|
| 144 |
+
cwd: /
|
| 145 |
+
config_sources:
|
| 146 |
+
- path: hydra.conf
|
| 147 |
+
schema: pkg
|
| 148 |
+
provider: hydra
|
| 149 |
+
- path: optimum_benchmark
|
| 150 |
+
schema: pkg
|
| 151 |
+
provider: main
|
| 152 |
+
- path: hydra_plugins.hydra_colorlog.conf
|
| 153 |
+
schema: pkg
|
| 154 |
+
provider: hydra-colorlog
|
| 155 |
+
- path: /optimum-benchmark/examples/energy_star
|
| 156 |
+
schema: file
|
| 157 |
+
provider: command-line
|
| 158 |
+
- path: ''
|
| 159 |
+
schema: structured
|
| 160 |
+
provider: schema
|
| 161 |
+
output_dir: /runs/text_generation/openai-community/gpt2/2024-12-04-16-25-55
|
| 162 |
+
choices:
|
| 163 |
+
benchmark: energy_star
|
| 164 |
+
launcher: process
|
| 165 |
+
backend: pytorch
|
| 166 |
+
hydra/env: default
|
| 167 |
+
hydra/callbacks: null
|
| 168 |
+
hydra/job_logging: colorlog
|
| 169 |
+
hydra/hydra_logging: colorlog
|
| 170 |
+
hydra/hydra_help: default
|
| 171 |
+
hydra/help: default
|
| 172 |
+
hydra/sweeper: basic
|
| 173 |
+
hydra/launcher: basic
|
| 174 |
+
hydra/output: default
|
| 175 |
+
verbose: false
|
text_generation/openai-community/gpt2/2024-12-04-16-25-55/.hydra/overrides.yaml
ADDED
|
@@ -0,0 +1,2 @@
|
|
|
|
|
|
|
|
|
|
| 1 |
+
- backend.model=openai-community/gpt2
|
| 2 |
+
- backend.processor=openai-community/gpt2
|
text_generation/openai-community/gpt2/2024-12-04-16-25-55/benchmark_report.json
ADDED
|
@@ -0,0 +1,203 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"prefill": {
|
| 3 |
+
"memory": null,
|
| 4 |
+
"latency": null,
|
| 5 |
+
"throughput": null,
|
| 6 |
+
"energy": {
|
| 7 |
+
"unit": "kWh",
|
| 8 |
+
"cpu": 0.00010681085390979409,
|
| 9 |
+
"ram": 1.0297819029295639e-06,
|
| 10 |
+
"gpu": 0.0005602429481940008,
|
| 11 |
+
"total": 0.0006680835840067245
|
| 12 |
+
},
|
| 13 |
+
"efficiency": {
|
| 14 |
+
"unit": "tokens/kWh",
|
| 15 |
+
"value": 376223882.78510684
|
| 16 |
+
},
|
| 17 |
+
"measures": [
|
| 18 |
+
{
|
| 19 |
+
"unit": "kWh",
|
| 20 |
+
"cpu": 0.00011892353653820971,
|
| 21 |
+
"ram": 1.146182340068226e-06,
|
| 22 |
+
"gpu": 0.000611663822663977,
|
| 23 |
+
"total": 0.000731733541542255
|
| 24 |
+
},
|
| 25 |
+
{
|
| 26 |
+
"unit": "kWh",
|
| 27 |
+
"cpu": 0.00011834202891877291,
|
| 28 |
+
"ram": 1.1408746650941389e-06,
|
| 29 |
+
"gpu": 0.0006136460464720717,
|
| 30 |
+
"total": 0.0007331289500559388
|
| 31 |
+
},
|
| 32 |
+
{
|
| 33 |
+
"unit": "kWh",
|
| 34 |
+
"cpu": 0.00011868103943603878,
|
| 35 |
+
"ram": 1.1442912318482203e-06,
|
| 36 |
+
"gpu": 0.0006243707772740636,
|
| 37 |
+
"total": 0.0007441961079419509
|
| 38 |
+
},
|
| 39 |
+
{
|
| 40 |
+
"unit": "kWh",
|
| 41 |
+
"cpu": 0.00011849597172222274,
|
| 42 |
+
"ram": 1.1424885480380128e-06,
|
| 43 |
+
"gpu": 0.0006210666079640159,
|
| 44 |
+
"total": 0.0007407050682342767
|
| 45 |
+
},
|
| 46 |
+
{
|
| 47 |
+
"unit": "kWh",
|
| 48 |
+
"cpu": 0.00011872881676738167,
|
| 49 |
+
"ram": 1.1447441446528026e-06,
|
| 50 |
+
"gpu": 0.0006302155041718605,
|
| 51 |
+
"total": 0.0007500890650838951
|
| 52 |
+
},
|
| 53 |
+
{
|
| 54 |
+
"unit": "kWh",
|
| 55 |
+
"cpu": 0.00011865041948475562,
|
| 56 |
+
"ram": 1.1439717885187163e-06,
|
| 57 |
+
"gpu": 0.0006199085514820846,
|
| 58 |
+
"total": 0.0007397029427553595
|
| 59 |
+
},
|
| 60 |
+
{
|
| 61 |
+
"unit": "kWh",
|
| 62 |
+
"cpu": 0.0,
|
| 63 |
+
"ram": 0.0,
|
| 64 |
+
"gpu": 0.0,
|
| 65 |
+
"total": 0.0
|
| 66 |
+
},
|
| 67 |
+
{
|
| 68 |
+
"unit": "kWh",
|
| 69 |
+
"cpu": 0.00011862063215556493,
|
| 70 |
+
"ram": 1.143727828368652e-06,
|
| 71 |
+
"gpu": 0.000625526055975989,
|
| 72 |
+
"total": 0.0007452904159599228
|
| 73 |
+
},
|
| 74 |
+
{
|
| 75 |
+
"unit": "kWh",
|
| 76 |
+
"cpu": 0.00011873292297564662,
|
| 77 |
+
"ram": 1.1448061300214643e-06,
|
| 78 |
+
"gpu": 0.0006323510614358874,
|
| 79 |
+
"total": 0.0007522287905415547
|
| 80 |
+
},
|
| 81 |
+
{
|
| 82 |
+
"unit": "kWh",
|
| 83 |
+
"cpu": 0.0001189331710993479,
|
| 84 |
+
"ram": 1.1467323526854064e-06,
|
| 85 |
+
"gpu": 0.0006236810545000582,
|
| 86 |
+
"total": 0.0007437609579520913
|
| 87 |
+
}
|
| 88 |
+
]
|
| 89 |
+
},
|
| 90 |
+
"decode": {
|
| 91 |
+
"memory": null,
|
| 92 |
+
"latency": null,
|
| 93 |
+
"throughput": null,
|
| 94 |
+
"energy": {
|
| 95 |
+
"unit": "kWh",
|
| 96 |
+
"cpu": 0.0005650565983572258,
|
| 97 |
+
"ram": 5.450526864520907e-06,
|
| 98 |
+
"gpu": 0.0012151225554306033,
|
| 99 |
+
"total": 0.0017856296806523508
|
| 100 |
+
},
|
| 101 |
+
"efficiency": {
|
| 102 |
+
"unit": "tokens/kWh",
|
| 103 |
+
"value": 5040238.800640902
|
| 104 |
+
},
|
| 105 |
+
"measures": [
|
| 106 |
+
{
|
| 107 |
+
"unit": "kWh",
|
| 108 |
+
"cpu": 0.0006368191074999889,
|
| 109 |
+
"ram": 6.142777259528393e-06,
|
| 110 |
+
"gpu": 0.0014249425288420658,
|
| 111 |
+
"total": 0.002067904413601583
|
| 112 |
+
},
|
| 113 |
+
{
|
| 114 |
+
"unit": "kWh",
|
| 115 |
+
"cpu": 0.0006249990986270203,
|
| 116 |
+
"ram": 6.028900507407003e-06,
|
| 117 |
+
"gpu": 0.0013758658229138243,
|
| 118 |
+
"total": 0.0020068938220482507
|
| 119 |
+
},
|
| 120 |
+
{
|
| 121 |
+
"unit": "kWh",
|
| 122 |
+
"cpu": 0.0006283337551250499,
|
| 123 |
+
"ram": 6.060942308998227e-06,
|
| 124 |
+
"gpu": 0.0013661177595599705,
|
| 125 |
+
"total": 0.002000512456994019
|
| 126 |
+
},
|
| 127 |
+
{
|
| 128 |
+
"unit": "kWh",
|
| 129 |
+
"cpu": 0.0006280304214923487,
|
| 130 |
+
"ram": 6.057771480830845e-06,
|
| 131 |
+
"gpu": 0.0013490144125440917,
|
| 132 |
+
"total": 0.0019831026055172726
|
| 133 |
+
},
|
| 134 |
+
{
|
| 135 |
+
"unit": "kWh",
|
| 136 |
+
"cpu": -0.00011872881676738167,
|
| 137 |
+
"ram": -1.1447441446528026e-06,
|
| 138 |
+
"gpu": -0.0006302155041718605,
|
| 139 |
+
"total": -0.0007500890650838951
|
| 140 |
+
},
|
| 141 |
+
{
|
| 142 |
+
"unit": "kWh",
|
| 143 |
+
"cpu": 0.0006244352388493348,
|
| 144 |
+
"ram": 6.023390549229008e-06,
|
| 145 |
+
"gpu": 0.001313566606407801,
|
| 146 |
+
"total": 0.0019440252358063647
|
| 147 |
+
},
|
| 148 |
+
{
|
| 149 |
+
"unit": "kWh",
|
| 150 |
+
"cpu": 0.0007459197505402847,
|
| 151 |
+
"ram": 7.1942984685488865e-06,
|
| 152 |
+
"gpu": 0.0019516298946361532,
|
| 153 |
+
"total": 0.0027047439436449876
|
| 154 |
+
},
|
| 155 |
+
{
|
| 156 |
+
"unit": "kWh",
|
| 157 |
+
"cpu": 0.0006257280835250589,
|
| 158 |
+
"ram": 6.035812423349825e-06,
|
| 159 |
+
"gpu": 0.0013290538410198893,
|
| 160 |
+
"total": 0.001960817736968298
|
| 161 |
+
},
|
| 162 |
+
{
|
| 163 |
+
"unit": "kWh",
|
| 164 |
+
"cpu": 0.0006308752698229713,
|
| 165 |
+
"ram": 6.085473057879209e-06,
|
| 166 |
+
"gpu": 0.001338571626412155,
|
| 167 |
+
"total": 0.0019755323692930097
|
| 168 |
+
},
|
| 169 |
+
{
|
| 170 |
+
"unit": "kWh",
|
| 171 |
+
"cpu": 0.0006241540748575822,
|
| 172 |
+
"ram": 6.020646734090474e-06,
|
| 173 |
+
"gpu": 0.0013326785661419427,
|
| 174 |
+
"total": 0.0019628532877336196
|
| 175 |
+
}
|
| 176 |
+
]
|
| 177 |
+
},
|
| 178 |
+
"per_token": {
|
| 179 |
+
"memory": null,
|
| 180 |
+
"latency": null,
|
| 181 |
+
"throughput": null,
|
| 182 |
+
"energy": null,
|
| 183 |
+
"efficiency": null,
|
| 184 |
+
"measures": null
|
| 185 |
+
},
|
| 186 |
+
"preprocess": {
|
| 187 |
+
"memory": null,
|
| 188 |
+
"latency": null,
|
| 189 |
+
"throughput": null,
|
| 190 |
+
"energy": {
|
| 191 |
+
"unit": "kWh",
|
| 192 |
+
"cpu": 1.0390234946579667e-05,
|
| 193 |
+
"ram": 6.987208991124449e-08,
|
| 194 |
+
"gpu": 1.754223625605178e-05,
|
| 195 |
+
"total": 2.800234329254269e-05
|
| 196 |
+
},
|
| 197 |
+
"efficiency": {
|
| 198 |
+
"unit": "samples/kWh",
|
| 199 |
+
"value": 35711297.070853
|
| 200 |
+
},
|
| 201 |
+
"measures": null
|
| 202 |
+
}
|
| 203 |
+
}
|
text_generation/openai-community/gpt2/2024-12-04-16-25-55/cli.log
ADDED
|
@@ -0,0 +1,188 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
[2024-12-04 16:25:58,357][launcher][INFO] - ََAllocating process launcher
|
| 2 |
+
[2024-12-04 16:25:58,357][process][INFO] - + Setting multiprocessing start method to spawn.
|
| 3 |
+
[2024-12-04 16:25:58,371][process][INFO] - + Launched benchmark in isolated process 614.
|
| 4 |
+
[PROC-0][2024-12-04 16:26:00,920][datasets][INFO] - PyTorch version 2.4.0 available.
|
| 5 |
+
[PROC-0][2024-12-04 16:26:01,844][backend][INFO] - َAllocating pytorch backend
|
| 6 |
+
[PROC-0][2024-12-04 16:26:01,844][backend][INFO] - + Setting random seed to 42
|
| 7 |
+
[PROC-0][2024-12-04 16:26:02,751][pytorch][INFO] - + Using AutoModel class AutoModelForCausalLM
|
| 8 |
+
[PROC-0][2024-12-04 16:26:02,751][pytorch][INFO] - + Creating backend temporary directory
|
| 9 |
+
[PROC-0][2024-12-04 16:26:02,752][pytorch][INFO] - + Loading model with random weights
|
| 10 |
+
[PROC-0][2024-12-04 16:26:02,752][pytorch][INFO] - + Creating no weights model
|
| 11 |
+
[PROC-0][2024-12-04 16:26:02,752][pytorch][INFO] - + Creating no weights model directory
|
| 12 |
+
[PROC-0][2024-12-04 16:26:02,752][pytorch][INFO] - + Creating no weights model state dict
|
| 13 |
+
[PROC-0][2024-12-04 16:26:02,773][pytorch][INFO] - + Saving no weights model safetensors
|
| 14 |
+
[PROC-0][2024-12-04 16:26:02,774][pytorch][INFO] - + Saving no weights model pretrained config
|
| 15 |
+
[PROC-0][2024-12-04 16:26:02,775][pytorch][INFO] - + Loading no weights AutoModel
|
| 16 |
+
[PROC-0][2024-12-04 16:26:02,775][pytorch][INFO] - + Loading model directly on device: cuda
|
| 17 |
+
[PROC-0][2024-12-04 16:26:02,932][pytorch][INFO] - + Turning on model's eval mode
|
| 18 |
+
[PROC-0][2024-12-04 16:26:02,938][benchmark][INFO] - Allocating energy_star benchmark
|
| 19 |
+
[PROC-0][2024-12-04 16:26:02,938][energy_star][INFO] - + Loading raw dataset
|
| 20 |
+
[PROC-0][2024-12-04 16:26:04,437][energy_star][INFO] - + Updating Text Generation kwargs with default values
|
| 21 |
+
[PROC-0][2024-12-04 16:26:04,437][energy_star][INFO] - + Initializing Text Generation report
|
| 22 |
+
[PROC-0][2024-12-04 16:26:04,437][energy][INFO] - + Tracking GPU energy on devices [0]
|
| 23 |
+
[PROC-0][2024-12-04 16:26:08,638][energy_star][INFO] - + Preprocessing dataset
|
| 24 |
+
[PROC-0][2024-12-04 16:26:09,519][energy][INFO] - + Saving codecarbon emission data to preprocess_codecarbon.json
|
| 25 |
+
[PROC-0][2024-12-04 16:26:09,520][energy_star][INFO] - + Preparing backend for Inference
|
| 26 |
+
[PROC-0][2024-12-04 16:26:09,520][energy_star][INFO] - + Initialising dataloader
|
| 27 |
+
[PROC-0][2024-12-04 16:26:09,520][energy_star][INFO] - + Warming up backend for Inference
|
| 28 |
+
[PROC-0][2024-12-04 16:26:10,457][energy_star][INFO] - + Additional warmup for Text Generation
|
| 29 |
+
[PROC-0][2024-12-04 16:26:10,522][energy_star][INFO] - + Running Text Generation energy tracking for 10 iterations
|
| 30 |
+
[PROC-0][2024-12-04 16:26:10,522][energy_star][INFO] - + Prefill iteration 1/10
|
| 31 |
+
[PROC-0][2024-12-04 16:26:20,596][energy][INFO] - + Saving codecarbon emission data to prefill_codecarbon.json
|
| 32 |
+
[PROC-0][2024-12-04 16:26:20,597][energy_star][INFO] - + Prefill iteration 2/10
|
| 33 |
+
[PROC-0][2024-12-04 16:26:30,622][energy][INFO] - + Saving codecarbon emission data to prefill_codecarbon.json
|
| 34 |
+
[PROC-0][2024-12-04 16:26:30,622][energy_star][INFO] - + Prefill iteration 3/10
|
| 35 |
+
[PROC-0][2024-12-04 16:26:40,676][energy][INFO] - + Saving codecarbon emission data to prefill_codecarbon.json
|
| 36 |
+
[PROC-0][2024-12-04 16:26:40,676][energy_star][INFO] - + Prefill iteration 4/10
|
| 37 |
+
[PROC-0][2024-12-04 16:26:50,714][energy][INFO] - + Saving codecarbon emission data to prefill_codecarbon.json
|
| 38 |
+
[PROC-0][2024-12-04 16:26:50,714][energy_star][INFO] - + Prefill iteration 5/10
|
| 39 |
+
[PROC-0][2024-12-04 16:27:00,772][energy][INFO] - + Saving codecarbon emission data to prefill_codecarbon.json
|
| 40 |
+
[PROC-0][2024-12-04 16:27:00,773][energy_star][INFO] - + Prefill iteration 6/10
|
| 41 |
+
[PROC-0][2024-12-04 16:27:10,823][energy][INFO] - + Saving codecarbon emission data to prefill_codecarbon.json
|
| 42 |
+
[PROC-0][2024-12-04 16:27:10,824][energy_star][INFO] - + Prefill iteration 7/10
|
| 43 |
+
[PROC-0][2024-12-04 16:27:20,845][energy][INFO] - + Saving codecarbon emission data to prefill_codecarbon.json
|
| 44 |
+
[PROC-0][2024-12-04 16:27:20,845][energy_star][INFO] - + Prefill iteration 8/10
|
| 45 |
+
[PROC-0][2024-12-04 16:27:30,893][energy][INFO] - + Saving codecarbon emission data to prefill_codecarbon.json
|
| 46 |
+
[PROC-0][2024-12-04 16:27:30,894][energy_star][INFO] - + Prefill iteration 9/10
|
| 47 |
+
[PROC-0][2024-12-04 16:27:40,952][energy][INFO] - + Saving codecarbon emission data to prefill_codecarbon.json
|
| 48 |
+
[PROC-0][2024-12-04 16:27:40,952][energy_star][INFO] - + Prefill iteration 10/10
|
| 49 |
+
[PROC-0][2024-12-04 16:27:51,027][energy][INFO] - + Saving codecarbon emission data to prefill_codecarbon.json
|
| 50 |
+
[PROC-0][2024-12-04 16:27:51,028][energy_star][INFO] - + Decoding iteration 1/10
|
| 51 |
+
[PROC-0][2024-12-04 16:28:55,044][energy][INFO] - + Saving codecarbon emission data to generate_codecarbon.json
|
| 52 |
+
[PROC-0][2024-12-04 16:28:55,045][energy_star][INFO] - + Decoding iteration 2/10
|
| 53 |
+
[PROC-0][2024-12-04 16:29:58,011][energy][INFO] - + Saving codecarbon emission data to generate_codecarbon.json
|
| 54 |
+
[PROC-0][2024-12-04 16:29:58,011][energy_star][INFO] - + Decoding iteration 3/10
|
| 55 |
+
[PROC-0][2024-12-04 16:31:01,288][energy][INFO] - + Saving codecarbon emission data to generate_codecarbon.json
|
| 56 |
+
[PROC-0][2024-12-04 16:31:01,289][energy_star][INFO] - + Decoding iteration 4/10
|
| 57 |
+
[PROC-0][2024-12-04 16:32:04,524][energy][INFO] - + Saving codecarbon emission data to generate_codecarbon.json
|
| 58 |
+
[PROC-0][2024-12-04 16:32:04,525][energy_star][INFO] - + Decoding iteration 5/10
|
| 59 |
+
[PROC-0][2024-12-04 16:33:07,035][energy][INFO] - + Saving codecarbon emission data to generate_codecarbon.json
|
| 60 |
+
[PROC-0][2024-12-04 16:33:07,035][energy_star][INFO] - + Decoding iteration 6/10
|
| 61 |
+
[PROC-0][2024-12-04 16:34:09,979][energy][INFO] - + Saving codecarbon emission data to generate_codecarbon.json
|
| 62 |
+
[PROC-0][2024-12-04 16:34:09,980][energy_star][INFO] - + Decoding iteration 7/10
|
| 63 |
+
[PROC-0][2024-12-04 16:35:13,164][energy][INFO] - + Saving codecarbon emission data to generate_codecarbon.json
|
| 64 |
+
[PROC-0][2024-12-04 16:35:13,165][energy_star][INFO] - + Decoding iteration 8/10
|
| 65 |
+
[PROC-0][2024-12-04 16:36:16,216][energy][INFO] - + Saving codecarbon emission data to generate_codecarbon.json
|
| 66 |
+
[PROC-0][2024-12-04 16:36:16,216][energy_star][INFO] - + Decoding iteration 9/10
|
| 67 |
+
[PROC-0][2024-12-04 16:37:19,713][energy][INFO] - + Saving codecarbon emission data to generate_codecarbon.json
|
| 68 |
+
[PROC-0][2024-12-04 16:37:19,714][energy_star][INFO] - + Decoding iteration 10/10
|
| 69 |
+
[PROC-0][2024-12-04 16:38:22,658][energy][INFO] - + Saving codecarbon emission data to generate_codecarbon.json
|
| 70 |
+
[PROC-0][2024-12-04 16:38:22,659][energy][INFO] - + prefill energy consumption:
|
| 71 |
+
[PROC-0][2024-12-04 16:38:22,659][energy][INFO] - + CPU: 0.000107 (kWh)
|
| 72 |
+
[PROC-0][2024-12-04 16:38:22,659][energy][INFO] - + GPU: 0.000560 (kWh)
|
| 73 |
+
[PROC-0][2024-12-04 16:38:22,659][energy][INFO] - + RAM: 0.000001 (kWh)
|
| 74 |
+
[PROC-0][2024-12-04 16:38:22,659][energy][INFO] - + total: 0.000668 (kWh)
|
| 75 |
+
[PROC-0][2024-12-04 16:38:22,659][energy][INFO] - + prefill_iteration_1 energy consumption:
|
| 76 |
+
[PROC-0][2024-12-04 16:38:22,659][energy][INFO] - + CPU: 0.000119 (kWh)
|
| 77 |
+
[PROC-0][2024-12-04 16:38:22,659][energy][INFO] - + GPU: 0.000612 (kWh)
|
| 78 |
+
[PROC-0][2024-12-04 16:38:22,659][energy][INFO] - + RAM: 0.000001 (kWh)
|
| 79 |
+
[PROC-0][2024-12-04 16:38:22,660][energy][INFO] - + total: 0.000732 (kWh)
|
| 80 |
+
[PROC-0][2024-12-04 16:38:22,660][energy][INFO] - + prefill_iteration_2 energy consumption:
|
| 81 |
+
[PROC-0][2024-12-04 16:38:22,660][energy][INFO] - + CPU: 0.000118 (kWh)
|
| 82 |
+
[PROC-0][2024-12-04 16:38:22,660][energy][INFO] - + GPU: 0.000614 (kWh)
|
| 83 |
+
[PROC-0][2024-12-04 16:38:22,660][energy][INFO] - + RAM: 0.000001 (kWh)
|
| 84 |
+
[PROC-0][2024-12-04 16:38:22,660][energy][INFO] - + total: 0.000733 (kWh)
|
| 85 |
+
[PROC-0][2024-12-04 16:38:22,660][energy][INFO] - + prefill_iteration_3 energy consumption:
|
| 86 |
+
[PROC-0][2024-12-04 16:38:22,660][energy][INFO] - + CPU: 0.000119 (kWh)
|
| 87 |
+
[PROC-0][2024-12-04 16:38:22,660][energy][INFO] - + GPU: 0.000624 (kWh)
|
| 88 |
+
[PROC-0][2024-12-04 16:38:22,660][energy][INFO] - + RAM: 0.000001 (kWh)
|
| 89 |
+
[PROC-0][2024-12-04 16:38:22,660][energy][INFO] - + total: 0.000744 (kWh)
|
| 90 |
+
[PROC-0][2024-12-04 16:38:22,660][energy][INFO] - + prefill_iteration_4 energy consumption:
|
| 91 |
+
[PROC-0][2024-12-04 16:38:22,661][energy][INFO] - + CPU: 0.000118 (kWh)
|
| 92 |
+
[PROC-0][2024-12-04 16:38:22,661][energy][INFO] - + GPU: 0.000621 (kWh)
|
| 93 |
+
[PROC-0][2024-12-04 16:38:22,661][energy][INFO] - + RAM: 0.000001 (kWh)
|
| 94 |
+
[PROC-0][2024-12-04 16:38:22,661][energy][INFO] - + total: 0.000741 (kWh)
|
| 95 |
+
[PROC-0][2024-12-04 16:38:22,661][energy][INFO] - + prefill_iteration_5 energy consumption:
|
| 96 |
+
[PROC-0][2024-12-04 16:38:22,661][energy][INFO] - + CPU: 0.000119 (kWh)
|
| 97 |
+
[PROC-0][2024-12-04 16:38:22,661][energy][INFO] - + GPU: 0.000630 (kWh)
|
| 98 |
+
[PROC-0][2024-12-04 16:38:22,661][energy][INFO] - + RAM: 0.000001 (kWh)
|
| 99 |
+
[PROC-0][2024-12-04 16:38:22,661][energy][INFO] - + total: 0.000750 (kWh)
|
| 100 |
+
[PROC-0][2024-12-04 16:38:22,661][energy][INFO] - + prefill_iteration_6 energy consumption:
|
| 101 |
+
[PROC-0][2024-12-04 16:38:22,661][energy][INFO] - + CPU: 0.000119 (kWh)
|
| 102 |
+
[PROC-0][2024-12-04 16:38:22,661][energy][INFO] - + GPU: 0.000620 (kWh)
|
| 103 |
+
[PROC-0][2024-12-04 16:38:22,662][energy][INFO] - + RAM: 0.000001 (kWh)
|
| 104 |
+
[PROC-0][2024-12-04 16:38:22,662][energy][INFO] - + total: 0.000740 (kWh)
|
| 105 |
+
[PROC-0][2024-12-04 16:38:22,662][energy][INFO] - + prefill_iteration_7 energy consumption:
|
| 106 |
+
[PROC-0][2024-12-04 16:38:22,662][energy][INFO] - + CPU: 0.000000 (kWh)
|
| 107 |
+
[PROC-0][2024-12-04 16:38:22,662][energy][INFO] - + GPU: 0.000000 (kWh)
|
| 108 |
+
[PROC-0][2024-12-04 16:38:22,662][energy][INFO] - + RAM: 0.000000 (kWh)
|
| 109 |
+
[PROC-0][2024-12-04 16:38:22,662][energy][INFO] - + total: 0.000000 (kWh)
|
| 110 |
+
[PROC-0][2024-12-04 16:38:22,662][energy][INFO] - + prefill_iteration_8 energy consumption:
|
| 111 |
+
[PROC-0][2024-12-04 16:38:22,662][energy][INFO] - + CPU: 0.000119 (kWh)
|
| 112 |
+
[PROC-0][2024-12-04 16:38:22,662][energy][INFO] - + GPU: 0.000626 (kWh)
|
| 113 |
+
[PROC-0][2024-12-04 16:38:22,662][energy][INFO] - + RAM: 0.000001 (kWh)
|
| 114 |
+
[PROC-0][2024-12-04 16:38:22,662][energy][INFO] - + total: 0.000745 (kWh)
|
| 115 |
+
[PROC-0][2024-12-04 16:38:22,662][energy][INFO] - + prefill_iteration_9 energy consumption:
|
| 116 |
+
[PROC-0][2024-12-04 16:38:22,663][energy][INFO] - + CPU: 0.000119 (kWh)
|
| 117 |
+
[PROC-0][2024-12-04 16:38:22,663][energy][INFO] - + GPU: 0.000632 (kWh)
|
| 118 |
+
[PROC-0][2024-12-04 16:38:22,663][energy][INFO] - + RAM: 0.000001 (kWh)
|
| 119 |
+
[PROC-0][2024-12-04 16:38:22,663][energy][INFO] - + total: 0.000752 (kWh)
|
| 120 |
+
[PROC-0][2024-12-04 16:38:22,663][energy][INFO] - + prefill_iteration_10 energy consumption:
|
| 121 |
+
[PROC-0][2024-12-04 16:38:22,663][energy][INFO] - + CPU: 0.000119 (kWh)
|
| 122 |
+
[PROC-0][2024-12-04 16:38:22,663][energy][INFO] - + GPU: 0.000624 (kWh)
|
| 123 |
+
[PROC-0][2024-12-04 16:38:22,663][energy][INFO] - + RAM: 0.000001 (kWh)
|
| 124 |
+
[PROC-0][2024-12-04 16:38:22,663][energy][INFO] - + total: 0.000744 (kWh)
|
| 125 |
+
[PROC-0][2024-12-04 16:38:22,663][energy][INFO] - + decode energy consumption:
|
| 126 |
+
[PROC-0][2024-12-04 16:38:22,663][energy][INFO] - + CPU: 0.000565 (kWh)
|
| 127 |
+
[PROC-0][2024-12-04 16:38:22,663][energy][INFO] - + GPU: 0.001215 (kWh)
|
| 128 |
+
[PROC-0][2024-12-04 16:38:22,664][energy][INFO] - + RAM: 0.000005 (kWh)
|
| 129 |
+
[PROC-0][2024-12-04 16:38:22,664][energy][INFO] - + total: 0.001786 (kWh)
|
| 130 |
+
[PROC-0][2024-12-04 16:38:22,664][energy][INFO] - + decode_iteration_1 energy consumption:
|
| 131 |
+
[PROC-0][2024-12-04 16:38:22,664][energy][INFO] - + CPU: 0.000637 (kWh)
|
| 132 |
+
[PROC-0][2024-12-04 16:38:22,664][energy][INFO] - + GPU: 0.001425 (kWh)
|
| 133 |
+
[PROC-0][2024-12-04 16:38:22,664][energy][INFO] - + RAM: 0.000006 (kWh)
|
| 134 |
+
[PROC-0][2024-12-04 16:38:22,664][energy][INFO] - + total: 0.002068 (kWh)
|
| 135 |
+
[PROC-0][2024-12-04 16:38:22,664][energy][INFO] - + decode_iteration_2 energy consumption:
|
| 136 |
+
[PROC-0][2024-12-04 16:38:22,664][energy][INFO] - + CPU: 0.000625 (kWh)
|
| 137 |
+
[PROC-0][2024-12-04 16:38:22,664][energy][INFO] - + GPU: 0.001376 (kWh)
|
| 138 |
+
[PROC-0][2024-12-04 16:38:22,664][energy][INFO] - + RAM: 0.000006 (kWh)
|
| 139 |
+
[PROC-0][2024-12-04 16:38:22,664][energy][INFO] - + total: 0.002007 (kWh)
|
| 140 |
+
[PROC-0][2024-12-04 16:38:22,665][energy][INFO] - + decode_iteration_3 energy consumption:
|
| 141 |
+
[PROC-0][2024-12-04 16:38:22,665][energy][INFO] - + CPU: 0.000628 (kWh)
|
| 142 |
+
[PROC-0][2024-12-04 16:38:22,665][energy][INFO] - + GPU: 0.001366 (kWh)
|
| 143 |
+
[PROC-0][2024-12-04 16:38:22,665][energy][INFO] - + RAM: 0.000006 (kWh)
|
| 144 |
+
[PROC-0][2024-12-04 16:38:22,665][energy][INFO] - + total: 0.002001 (kWh)
|
| 145 |
+
[PROC-0][2024-12-04 16:38:22,665][energy][INFO] - + decode_iteration_4 energy consumption:
|
| 146 |
+
[PROC-0][2024-12-04 16:38:22,665][energy][INFO] - + CPU: 0.000628 (kWh)
|
| 147 |
+
[PROC-0][2024-12-04 16:38:22,665][energy][INFO] - + GPU: 0.001349 (kWh)
|
| 148 |
+
[PROC-0][2024-12-04 16:38:22,665][energy][INFO] - + RAM: 0.000006 (kWh)
|
| 149 |
+
[PROC-0][2024-12-04 16:38:22,665][energy][INFO] - + total: 0.001983 (kWh)
|
| 150 |
+
[PROC-0][2024-12-04 16:38:22,665][energy][INFO] - + decode_iteration_5 energy consumption:
|
| 151 |
+
[PROC-0][2024-12-04 16:38:22,665][energy][INFO] - + CPU: -0.000119 (kWh)
|
| 152 |
+
[PROC-0][2024-12-04 16:38:22,665][energy][INFO] - + GPU: -0.000630 (kWh)
|
| 153 |
+
[PROC-0][2024-12-04 16:38:22,666][energy][INFO] - + RAM: -0.000001 (kWh)
|
| 154 |
+
[PROC-0][2024-12-04 16:38:22,666][energy][INFO] - + total: -0.000750 (kWh)
|
| 155 |
+
[PROC-0][2024-12-04 16:38:22,666][energy][INFO] - + decode_iteration_6 energy consumption:
|
| 156 |
+
[PROC-0][2024-12-04 16:38:22,666][energy][INFO] - + CPU: 0.000624 (kWh)
|
| 157 |
+
[PROC-0][2024-12-04 16:38:22,666][energy][INFO] - + GPU: 0.001314 (kWh)
|
| 158 |
+
[PROC-0][2024-12-04 16:38:22,666][energy][INFO] - + RAM: 0.000006 (kWh)
|
| 159 |
+
[PROC-0][2024-12-04 16:38:22,666][energy][INFO] - + total: 0.001944 (kWh)
|
| 160 |
+
[PROC-0][2024-12-04 16:38:22,666][energy][INFO] - + decode_iteration_7 energy consumption:
|
| 161 |
+
[PROC-0][2024-12-04 16:38:22,666][energy][INFO] - + CPU: 0.000746 (kWh)
|
| 162 |
+
[PROC-0][2024-12-04 16:38:22,666][energy][INFO] - + GPU: 0.001952 (kWh)
|
| 163 |
+
[PROC-0][2024-12-04 16:38:22,666][energy][INFO] - + RAM: 0.000007 (kWh)
|
| 164 |
+
[PROC-0][2024-12-04 16:38:22,666][energy][INFO] - + total: 0.002705 (kWh)
|
| 165 |
+
[PROC-0][2024-12-04 16:38:22,667][energy][INFO] - + decode_iteration_8 energy consumption:
|
| 166 |
+
[PROC-0][2024-12-04 16:38:22,667][energy][INFO] - + CPU: 0.000626 (kWh)
|
| 167 |
+
[PROC-0][2024-12-04 16:38:22,667][energy][INFO] - + GPU: 0.001329 (kWh)
|
| 168 |
+
[PROC-0][2024-12-04 16:38:22,667][energy][INFO] - + RAM: 0.000006 (kWh)
|
| 169 |
+
[PROC-0][2024-12-04 16:38:22,667][energy][INFO] - + total: 0.001961 (kWh)
|
| 170 |
+
[PROC-0][2024-12-04 16:38:22,667][energy][INFO] - + decode_iteration_9 energy consumption:
|
| 171 |
+
[PROC-0][2024-12-04 16:38:22,667][energy][INFO] - + CPU: 0.000631 (kWh)
|
| 172 |
+
[PROC-0][2024-12-04 16:38:22,667][energy][INFO] - + GPU: 0.001339 (kWh)
|
| 173 |
+
[PROC-0][2024-12-04 16:38:22,667][energy][INFO] - + RAM: 0.000006 (kWh)
|
| 174 |
+
[PROC-0][2024-12-04 16:38:22,667][energy][INFO] - + total: 0.001976 (kWh)
|
| 175 |
+
[PROC-0][2024-12-04 16:38:22,667][energy][INFO] - + decode_iteration_10 energy consumption:
|
| 176 |
+
[PROC-0][2024-12-04 16:38:22,667][energy][INFO] - + CPU: 0.000624 (kWh)
|
| 177 |
+
[PROC-0][2024-12-04 16:38:22,667][energy][INFO] - + GPU: 0.001333 (kWh)
|
| 178 |
+
[PROC-0][2024-12-04 16:38:22,668][energy][INFO] - + RAM: 0.000006 (kWh)
|
| 179 |
+
[PROC-0][2024-12-04 16:38:22,668][energy][INFO] - + total: 0.001963 (kWh)
|
| 180 |
+
[PROC-0][2024-12-04 16:38:22,668][energy][INFO] - + preprocess energy consumption:
|
| 181 |
+
[PROC-0][2024-12-04 16:38:22,668][energy][INFO] - + CPU: 0.000010 (kWh)
|
| 182 |
+
[PROC-0][2024-12-04 16:38:22,668][energy][INFO] - + GPU: 0.000018 (kWh)
|
| 183 |
+
[PROC-0][2024-12-04 16:38:22,668][energy][INFO] - + RAM: 0.000000 (kWh)
|
| 184 |
+
[PROC-0][2024-12-04 16:38:22,668][energy][INFO] - + total: 0.000028 (kWh)
|
| 185 |
+
[PROC-0][2024-12-04 16:38:22,668][energy][INFO] - + prefill energy efficiency: 376223882.785107 (tokens/kWh)
|
| 186 |
+
[PROC-0][2024-12-04 16:38:22,669][energy][INFO] - + decode energy efficiency: 5040238.800641 (tokens/kWh)
|
| 187 |
+
[PROC-0][2024-12-04 16:38:22,669][energy][INFO] - + preprocess energy efficiency: 35711297.070853 (samples/kWh)
|
| 188 |
+
[2024-12-04 16:38:23,359][datasets][INFO] - PyTorch version 2.4.0 available.
|
text_generation/openai-community/gpt2/2024-12-04-16-25-55/error.log
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
text_generation/openai-community/gpt2/2024-12-04-16-25-55/experiment_config.json
ADDED
|
@@ -0,0 +1,110 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"experiment_name": "text_generation",
|
| 3 |
+
"backend": {
|
| 4 |
+
"name": "pytorch",
|
| 5 |
+
"version": "2.4.0",
|
| 6 |
+
"_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
|
| 7 |
+
"task": "text-generation",
|
| 8 |
+
"model": "openai-community/gpt2",
|
| 9 |
+
"processor": "openai-community/gpt2",
|
| 10 |
+
"library": "transformers",
|
| 11 |
+
"device": "cuda",
|
| 12 |
+
"device_ids": "0",
|
| 13 |
+
"seed": 42,
|
| 14 |
+
"inter_op_num_threads": null,
|
| 15 |
+
"intra_op_num_threads": null,
|
| 16 |
+
"hub_kwargs": {
|
| 17 |
+
"revision": "main",
|
| 18 |
+
"force_download": false,
|
| 19 |
+
"local_files_only": false,
|
| 20 |
+
"trust_remote_code": true
|
| 21 |
+
},
|
| 22 |
+
"no_weights": true,
|
| 23 |
+
"device_map": null,
|
| 24 |
+
"torch_dtype": null,
|
| 25 |
+
"amp_autocast": false,
|
| 26 |
+
"amp_dtype": null,
|
| 27 |
+
"eval_mode": true,
|
| 28 |
+
"to_bettertransformer": false,
|
| 29 |
+
"low_cpu_mem_usage": null,
|
| 30 |
+
"attn_implementation": null,
|
| 31 |
+
"cache_implementation": null,
|
| 32 |
+
"torch_compile": false,
|
| 33 |
+
"torch_compile_config": {},
|
| 34 |
+
"quantization_scheme": null,
|
| 35 |
+
"quantization_config": {},
|
| 36 |
+
"deepspeed_inference": false,
|
| 37 |
+
"deepspeed_inference_config": {},
|
| 38 |
+
"peft_type": null,
|
| 39 |
+
"peft_config": {}
|
| 40 |
+
},
|
| 41 |
+
"launcher": {
|
| 42 |
+
"name": "process",
|
| 43 |
+
"_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
|
| 44 |
+
"device_isolation": false,
|
| 45 |
+
"device_isolation_action": "warn",
|
| 46 |
+
"start_method": "spawn"
|
| 47 |
+
},
|
| 48 |
+
"benchmark": {
|
| 49 |
+
"name": "energy_star",
|
| 50 |
+
"_target_": "optimum_benchmark.benchmarks.energy_star.benchmark.EnergyStarBenchmark",
|
| 51 |
+
"dataset_name": "EnergyStarAI/text_generation",
|
| 52 |
+
"dataset_config": "",
|
| 53 |
+
"dataset_split": "train",
|
| 54 |
+
"num_samples": 1000,
|
| 55 |
+
"input_shapes": {
|
| 56 |
+
"batch_size": 1
|
| 57 |
+
},
|
| 58 |
+
"text_column_name": "text",
|
| 59 |
+
"truncation": true,
|
| 60 |
+
"max_length": -1,
|
| 61 |
+
"dataset_prefix1": "",
|
| 62 |
+
"dataset_prefix2": "",
|
| 63 |
+
"t5_task": "",
|
| 64 |
+
"image_column_name": "image",
|
| 65 |
+
"resize": false,
|
| 66 |
+
"question_column_name": "question",
|
| 67 |
+
"context_column_name": "context",
|
| 68 |
+
"sentence1_column_name": "sentence1",
|
| 69 |
+
"sentence2_column_name": "sentence2",
|
| 70 |
+
"audio_column_name": "audio",
|
| 71 |
+
"iterations": 10,
|
| 72 |
+
"warmup_runs": 10,
|
| 73 |
+
"energy": true,
|
| 74 |
+
"forward_kwargs": {},
|
| 75 |
+
"generate_kwargs": {
|
| 76 |
+
"max_new_tokens": 10,
|
| 77 |
+
"min_new_tokens": 10
|
| 78 |
+
},
|
| 79 |
+
"call_kwargs": {}
|
| 80 |
+
},
|
| 81 |
+
"environment": {
|
| 82 |
+
"cpu": " AMD EPYC 7R32",
|
| 83 |
+
"cpu_count": 48,
|
| 84 |
+
"cpu_ram_mb": 200472.73984,
|
| 85 |
+
"system": "Linux",
|
| 86 |
+
"machine": "x86_64",
|
| 87 |
+
"platform": "Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35",
|
| 88 |
+
"processor": "x86_64",
|
| 89 |
+
"python_version": "3.9.20",
|
| 90 |
+
"gpu": [
|
| 91 |
+
"NVIDIA A10G"
|
| 92 |
+
],
|
| 93 |
+
"gpu_count": 1,
|
| 94 |
+
"gpu_vram_mb": 24146608128,
|
| 95 |
+
"optimum_benchmark_version": "0.2.0",
|
| 96 |
+
"optimum_benchmark_commit": null,
|
| 97 |
+
"transformers_version": "4.44.0",
|
| 98 |
+
"transformers_commit": null,
|
| 99 |
+
"accelerate_version": "0.33.0",
|
| 100 |
+
"accelerate_commit": null,
|
| 101 |
+
"diffusers_version": "0.30.0",
|
| 102 |
+
"diffusers_commit": null,
|
| 103 |
+
"optimum_version": null,
|
| 104 |
+
"optimum_commit": null,
|
| 105 |
+
"timm_version": null,
|
| 106 |
+
"timm_commit": null,
|
| 107 |
+
"peft_version": null,
|
| 108 |
+
"peft_commit": null
|
| 109 |
+
}
|
| 110 |
+
}
|
text_generation/openai-community/gpt2/2024-12-04-16-25-55/generate_codecarbon.json
ADDED
|
@@ -0,0 +1,33 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"timestamp": "2024-12-04T16:38:22",
|
| 3 |
+
"project_name": "codecarbon",
|
| 4 |
+
"run_id": "afde1576-ebf9-460b-b8e2-a655a80c078d",
|
| 5 |
+
"duration": -1733284530.346387,
|
| 6 |
+
"emissions": 0.0009991043657714086,
|
| 7 |
+
"emissions_rate": 1.5873427648203317e-05,
|
| 8 |
+
"cpu_power": 42.5,
|
| 9 |
+
"gpu_power": 111.89574880758785,
|
| 10 |
+
"ram_power": 0.40994882583618164,
|
| 11 |
+
"cpu_energy": 0.0007430872459569301,
|
| 12 |
+
"gpu_energy": 0.001956359620642001,
|
| 13 |
+
"ram_energy": 7.16737908677588e-06,
|
| 14 |
+
"energy_consumed": 0.002706614245685711,
|
| 15 |
+
"country_name": "United States",
|
| 16 |
+
"country_iso_code": "USA",
|
| 17 |
+
"region": "virginia",
|
| 18 |
+
"cloud_provider": "",
|
| 19 |
+
"cloud_region": "",
|
| 20 |
+
"os": "Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35",
|
| 21 |
+
"python_version": "3.9.20",
|
| 22 |
+
"codecarbon_version": "2.5.1",
|
| 23 |
+
"cpu_count": 48,
|
| 24 |
+
"cpu_model": "AMD EPYC 7R32",
|
| 25 |
+
"gpu_count": 1,
|
| 26 |
+
"gpu_model": "1 x NVIDIA A10G",
|
| 27 |
+
"longitude": -77.4903,
|
| 28 |
+
"latitude": 39.0469,
|
| 29 |
+
"ram_total_size": 186.7047882080078,
|
| 30 |
+
"tracking_mode": "process",
|
| 31 |
+
"on_cloud": "N",
|
| 32 |
+
"pue": 1.0
|
| 33 |
+
}
|
text_generation/openai-community/gpt2/2024-12-04-16-25-55/prefill_codecarbon.json
ADDED
|
@@ -0,0 +1,33 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"timestamp": "2024-12-04T16:27:51",
|
| 3 |
+
"project_name": "codecarbon",
|
| 4 |
+
"run_id": "afde1576-ebf9-460b-b8e2-a655a80c078d",
|
| 5 |
+
"duration": -1733284583.215868,
|
| 6 |
+
"emissions": 0.0002745477385130662,
|
| 7 |
+
"emissions_rate": 2.7257328036196253e-05,
|
| 8 |
+
"cpu_power": 42.5,
|
| 9 |
+
"gpu_power": 222.9195650721471,
|
| 10 |
+
"ram_power": 0.4099001884460449,
|
| 11 |
+
"cpu_energy": 0.0001189331710993479,
|
| 12 |
+
"gpu_energy": 0.0006236810545000582,
|
| 13 |
+
"ram_energy": 1.1467323526854064e-06,
|
| 14 |
+
"energy_consumed": 0.0007437609579520913,
|
| 15 |
+
"country_name": "United States",
|
| 16 |
+
"country_iso_code": "USA",
|
| 17 |
+
"region": "virginia",
|
| 18 |
+
"cloud_provider": "",
|
| 19 |
+
"cloud_region": "",
|
| 20 |
+
"os": "Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35",
|
| 21 |
+
"python_version": "3.9.20",
|
| 22 |
+
"codecarbon_version": "2.5.1",
|
| 23 |
+
"cpu_count": 48,
|
| 24 |
+
"cpu_model": "AMD EPYC 7R32",
|
| 25 |
+
"gpu_count": 1,
|
| 26 |
+
"gpu_model": "1 x NVIDIA A10G",
|
| 27 |
+
"longitude": -77.4903,
|
| 28 |
+
"latitude": 39.0469,
|
| 29 |
+
"ram_total_size": 186.7047882080078,
|
| 30 |
+
"tracking_mode": "process",
|
| 31 |
+
"on_cloud": "N",
|
| 32 |
+
"pue": 1.0
|
| 33 |
+
}
|
text_generation/openai-community/gpt2/2024-12-04-16-25-55/preprocess_codecarbon.json
ADDED
|
@@ -0,0 +1,33 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"timestamp": "2024-12-04T16:26:09",
|
| 3 |
+
"project_name": "codecarbon",
|
| 4 |
+
"run_id": "afde1576-ebf9-460b-b8e2-a655a80c078d",
|
| 5 |
+
"duration": -1733284592.4103684,
|
| 6 |
+
"emissions": 1.0336627570775682e-05,
|
| 7 |
+
"emissions_rate": 1.1772892293402276e-05,
|
| 8 |
+
"cpu_power": 42.5,
|
| 9 |
+
"gpu_power": 71.93685272733893,
|
| 10 |
+
"ram_power": 0.2868003845214844,
|
| 11 |
+
"cpu_energy": 1.0390234946579667e-05,
|
| 12 |
+
"gpu_energy": 1.754223625605178e-05,
|
| 13 |
+
"ram_energy": 6.987208991124449e-08,
|
| 14 |
+
"energy_consumed": 2.800234329254269e-05,
|
| 15 |
+
"country_name": "United States",
|
| 16 |
+
"country_iso_code": "USA",
|
| 17 |
+
"region": "virginia",
|
| 18 |
+
"cloud_provider": "",
|
| 19 |
+
"cloud_region": "",
|
| 20 |
+
"os": "Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35",
|
| 21 |
+
"python_version": "3.9.20",
|
| 22 |
+
"codecarbon_version": "2.5.1",
|
| 23 |
+
"cpu_count": 48,
|
| 24 |
+
"cpu_model": "AMD EPYC 7R32",
|
| 25 |
+
"gpu_count": 1,
|
| 26 |
+
"gpu_model": "1 x NVIDIA A10G",
|
| 27 |
+
"longitude": -77.4903,
|
| 28 |
+
"latitude": 39.0469,
|
| 29 |
+
"ram_total_size": 186.7047882080078,
|
| 30 |
+
"tracking_mode": "process",
|
| 31 |
+
"on_cloud": "N",
|
| 32 |
+
"pue": 1.0
|
| 33 |
+
}
|