sasha HF Staff commited on
Commit
1f997fd
·
verified ·
1 Parent(s): 7fbfe46

Upload folder using huggingface_hub

Browse files
text_generation/meta-llama/Llama-3.2-3B/2024-10-24-16-25-47/.hydra/config.yaml ADDED
@@ -0,0 +1,96 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ backend:
2
+ name: pytorch
3
+ version: 2.4.0
4
+ _target_: optimum_benchmark.backends.pytorch.backend.PyTorchBackend
5
+ task: text-generation
6
+ model: meta-llama/Llama-3.2-3B
7
+ processor: meta-llama/Llama-3.2-3B
8
+ library: null
9
+ device: cuda
10
+ device_ids: '0'
11
+ seed: 42
12
+ inter_op_num_threads: null
13
+ intra_op_num_threads: null
14
+ hub_kwargs: {}
15
+ no_weights: true
16
+ device_map: null
17
+ torch_dtype: null
18
+ amp_autocast: false
19
+ amp_dtype: null
20
+ eval_mode: true
21
+ to_bettertransformer: false
22
+ low_cpu_mem_usage: null
23
+ attn_implementation: null
24
+ cache_implementation: null
25
+ torch_compile: false
26
+ torch_compile_config: {}
27
+ quantization_scheme: null
28
+ quantization_config: {}
29
+ deepspeed_inference: false
30
+ deepspeed_inference_config: {}
31
+ peft_type: null
32
+ peft_config: {}
33
+ launcher:
34
+ name: process
35
+ _target_: optimum_benchmark.launchers.process.launcher.ProcessLauncher
36
+ device_isolation: false
37
+ device_isolation_action: warn
38
+ start_method: spawn
39
+ benchmark:
40
+ name: energy_star
41
+ _target_: optimum_benchmark.benchmarks.energy_star.benchmark.EnergyStarBenchmark
42
+ dataset_name: EnergyStarAI/text_generation
43
+ dataset_config: ''
44
+ dataset_split: train
45
+ num_samples: 1000
46
+ input_shapes:
47
+ batch_size: 1
48
+ text_column_name: text
49
+ truncation: true
50
+ max_length: -1
51
+ dataset_prefix1: ''
52
+ dataset_prefix2: ''
53
+ t5_task: ''
54
+ image_column_name: image
55
+ resize: false
56
+ question_column_name: question
57
+ context_column_name: context
58
+ sentence1_column_name: sentence1
59
+ sentence2_column_name: sentence2
60
+ audio_column_name: audio
61
+ iterations: 10
62
+ warmup_runs: 10
63
+ energy: true
64
+ forward_kwargs: {}
65
+ generate_kwargs:
66
+ max_new_tokens: 10
67
+ min_new_tokens: 10
68
+ call_kwargs: {}
69
+ experiment_name: text_generation
70
+ environment:
71
+ cpu: ' AMD EPYC 7R32'
72
+ cpu_count: 48
73
+ cpu_ram_mb: 200472.73984
74
+ system: Linux
75
+ machine: x86_64
76
+ platform: Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35
77
+ processor: x86_64
78
+ python_version: 3.9.20
79
+ gpu:
80
+ - NVIDIA A10G
81
+ gpu_count: 1
82
+ gpu_vram_mb: 24146608128
83
+ optimum_benchmark_version: 0.2.0
84
+ optimum_benchmark_commit: null
85
+ transformers_version: 4.44.0
86
+ transformers_commit: null
87
+ accelerate_version: 0.33.0
88
+ accelerate_commit: null
89
+ diffusers_version: 0.30.0
90
+ diffusers_commit: null
91
+ optimum_version: null
92
+ optimum_commit: null
93
+ timm_version: null
94
+ timm_commit: null
95
+ peft_version: null
96
+ peft_commit: null
text_generation/meta-llama/Llama-3.2-3B/2024-10-24-16-25-47/.hydra/hydra.yaml ADDED
@@ -0,0 +1,175 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: ./runs/text_generation/meta-llama/Llama-3.2-3B/2024-10-24-16-25-47
4
+ sweep:
5
+ dir: sweeps/${experiment_name}/${backend.model}/${now:%Y-%m-%d-%H-%M-%S}
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ formatters:
71
+ colorlog:
72
+ (): colorlog.ColoredFormatter
73
+ format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s'
74
+ handlers:
75
+ console:
76
+ class: logging.StreamHandler
77
+ formatter: colorlog
78
+ stream: ext://sys.stdout
79
+ root:
80
+ level: INFO
81
+ handlers:
82
+ - console
83
+ disable_existing_loggers: false
84
+ job_logging:
85
+ version: 1
86
+ formatters:
87
+ simple:
88
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
89
+ colorlog:
90
+ (): colorlog.ColoredFormatter
91
+ format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s]
92
+ - %(message)s'
93
+ log_colors:
94
+ DEBUG: purple
95
+ INFO: green
96
+ WARNING: yellow
97
+ ERROR: red
98
+ CRITICAL: red
99
+ handlers:
100
+ console:
101
+ class: logging.StreamHandler
102
+ formatter: colorlog
103
+ stream: ext://sys.stdout
104
+ file:
105
+ class: logging.FileHandler
106
+ formatter: simple
107
+ filename: ${hydra.job.name}.log
108
+ root:
109
+ level: INFO
110
+ handlers:
111
+ - console
112
+ - file
113
+ disable_existing_loggers: false
114
+ env: {}
115
+ mode: RUN
116
+ searchpath: []
117
+ callbacks: {}
118
+ output_subdir: .hydra
119
+ overrides:
120
+ hydra:
121
+ - hydra.run.dir=./runs/text_generation/meta-llama/Llama-3.2-3B/2024-10-24-16-25-47
122
+ - hydra.mode=RUN
123
+ task:
124
+ - backend.model=meta-llama/Llama-3.2-3B
125
+ - backend.processor=meta-llama/Llama-3.2-3B
126
+ job:
127
+ name: cli
128
+ chdir: true
129
+ override_dirname: backend.model=meta-llama/Llama-3.2-3B,backend.processor=meta-llama/Llama-3.2-3B
130
+ id: ???
131
+ num: ???
132
+ config_name: text_generation
133
+ env_set:
134
+ OVERRIDE_BENCHMARKS: '1'
135
+ env_copy: []
136
+ config:
137
+ override_dirname:
138
+ kv_sep: '='
139
+ item_sep: ','
140
+ exclude_keys: []
141
+ runtime:
142
+ version: 1.3.2
143
+ version_base: '1.3'
144
+ cwd: /
145
+ config_sources:
146
+ - path: hydra.conf
147
+ schema: pkg
148
+ provider: hydra
149
+ - path: optimum_benchmark
150
+ schema: pkg
151
+ provider: main
152
+ - path: hydra_plugins.hydra_colorlog.conf
153
+ schema: pkg
154
+ provider: hydra-colorlog
155
+ - path: /optimum-benchmark/examples/energy_star
156
+ schema: file
157
+ provider: command-line
158
+ - path: ''
159
+ schema: structured
160
+ provider: schema
161
+ output_dir: /runs/text_generation/meta-llama/Llama-3.2-3B/2024-10-24-16-25-47
162
+ choices:
163
+ benchmark: energy_star
164
+ launcher: process
165
+ backend: pytorch
166
+ hydra/env: default
167
+ hydra/callbacks: null
168
+ hydra/job_logging: colorlog
169
+ hydra/hydra_logging: colorlog
170
+ hydra/hydra_help: default
171
+ hydra/help: default
172
+ hydra/sweeper: basic
173
+ hydra/launcher: basic
174
+ hydra/output: default
175
+ verbose: false
text_generation/meta-llama/Llama-3.2-3B/2024-10-24-16-25-47/.hydra/overrides.yaml ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ - backend.model=meta-llama/Llama-3.2-3B
2
+ - backend.processor=meta-llama/Llama-3.2-3B
text_generation/meta-llama/Llama-3.2-3B/2024-10-24-16-25-47/cli.log ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [2024-10-24 16:25:50,576][launcher][INFO] - ََAllocating process launcher
2
+ [2024-10-24 16:25:50,577][process][INFO] - + Setting multiprocessing start method to spawn.
3
+ [2024-10-24 16:25:50,589][process][INFO] - + Launched benchmark in isolated process 180.
4
+ [PROC-0][2024-10-24 16:25:53,122][datasets][INFO] - PyTorch version 2.4.0 available.
5
+ [PROC-0][2024-10-24 16:25:54,069][backend][INFO] - َAllocating pytorch backend
6
+ [PROC-0][2024-10-24 16:25:54,069][backend][INFO] - + Setting random seed to 42
7
+ [PROC-0][2024-10-24 16:25:55,366][pytorch][INFO] - + Using AutoModel class AutoModelForCausalLM
8
+ [PROC-0][2024-10-24 16:25:55,367][pytorch][INFO] - + Creating backend temporary directory
9
+ [PROC-0][2024-10-24 16:25:55,367][pytorch][INFO] - + Loading model with random weights
10
+ [PROC-0][2024-10-24 16:25:55,367][pytorch][INFO] - + Creating no weights model
11
+ [PROC-0][2024-10-24 16:25:55,367][pytorch][INFO] - + Creating no weights model directory
12
+ [PROC-0][2024-10-24 16:25:55,367][pytorch][INFO] - + Creating no weights model state dict
13
+ [PROC-0][2024-10-24 16:25:55,389][pytorch][INFO] - + Saving no weights model safetensors
14
+ [PROC-0][2024-10-24 16:25:55,389][pytorch][INFO] - + Saving no weights model pretrained config
15
+ [PROC-0][2024-10-24 16:25:55,390][pytorch][INFO] - + Loading no weights AutoModel
16
+ [PROC-0][2024-10-24 16:25:55,390][pytorch][INFO] - + Loading model directly on device: cuda
17
+ [PROC-0][2024-10-24 16:25:55,801][pytorch][INFO] - + Turning on model's eval mode
18
+ [PROC-0][2024-10-24 16:25:55,808][benchmark][INFO] - Allocating energy_star benchmark
19
+ [PROC-0][2024-10-24 16:25:55,808][energy_star][INFO] - + Loading raw dataset
20
+ [PROC-0][2024-10-24 16:25:56,980][energy_star][INFO] - + Updating Text Generation kwargs with default values
21
+ [PROC-0][2024-10-24 16:25:56,980][energy_star][INFO] - + Initializing Text Generation report
22
+ [PROC-0][2024-10-24 16:25:56,980][energy][INFO] - + Tracking GPU energy on devices [0]
23
+ [PROC-0][2024-10-24 16:26:01,188][energy_star][INFO] - + Preprocessing dataset
24
+ [PROC-0][2024-10-24 16:26:02,092][energy][INFO] - + Saving codecarbon emission data to preprocess_codecarbon.json
25
+ [PROC-0][2024-10-24 16:26:02,092][energy_star][INFO] - + Preparing backend for Inference
26
+ [PROC-0][2024-10-24 16:26:02,092][energy_star][INFO] - + Initialising dataloader
27
+ [PROC-0][2024-10-24 16:26:02,093][energy_star][INFO] - + Warming up backend for Inference
28
+ [PROC-0][2024-10-24 16:26:04,452][energy_star][INFO] - + Additional warmup for Text Generation
29
+ [PROC-0][2024-10-24 16:26:04,877][energy_star][INFO] - + Running Text Generation energy tracking for 10 iterations
30
+ [PROC-0][2024-10-24 16:26:04,878][energy_star][INFO] - + Prefill iteration 1/10
31
+ [2024-10-24 16:27:18,508][experiment][ERROR] - Error during experiment
text_generation/meta-llama/Llama-3.2-3B/2024-10-24-16-25-47/error.log ADDED
@@ -0,0 +1,96 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
0
  0%| | 0/1000 [00:00<?, ?it/s]
1
  0%| | 1/1000 [00:00<02:15, 7.40it/s]
2
  0%| | 3/1000 [00:00<01:30, 11.00it/s]
3
  0%| | 5/1000 [00:00<02:48, 5.90it/s]
4
  1%| | 7/1000 [00:00<02:10, 7.63it/s]
5
  1%| | 8/1000 [00:02<06:03, 2.73it/s]
6
  1%| | 10/1000 [00:02<04:08, 3.99it/s]
7
  1%| | 11/1000 [00:02<04:52, 3.38it/s]
8
  1%|▏ | 13/1000 [00:02<03:27, 4.76it/s]
9
  1%|▏ | 14/1000 [00:03<03:48, 4.31it/s]
10
  2%|▏ | 15/1000 [00:03<03:32, 4.63it/s]
11
  2%|▏ | 18/1000 [00:03<02:10, 7.53it/s]
12
  2%|▏ | 20/1000 [00:03<01:49, 8.91it/s]
13
  2%|▏ | 23/1000 [00:03<01:23, 11.74it/s]
14
  2%|▎ | 25/1000 [00:03<01:13, 13.21it/s]
15
  3%|▎ | 27/1000 [00:04<01:42, 9.54it/s]
16
  3%|▎ | 30/1000 [00:04<01:17, 12.49it/s]
17
  3%|▎ | 32/1000 [00:04<01:16, 12.65it/s]
18
  4%|▎ | 35/1000 [00:04<01:12, 13.24it/s]
19
  4%|▎ | 37/1000 [00:04<01:11, 13.45it/s]
20
  4%|▍ | 39/1000 [00:05<01:33, 10.24it/s]
21
  4%|▍ | 41/1000 [00:05<01:57, 8.19it/s]
22
  4%|▍ | 43/1000 [00:05<01:37, 9.78it/s]
23
  5%|▍ | 46/1000 [00:05<01:17, 12.38it/s]
24
  5%|▍ | 48/1000 [00:05<01:18, 12.07it/s]
25
  5%|▌ | 50/1000 [00:06<01:32, 10.25it/s]
26
  5%|▌ | 52/1000 [00:06<01:22, 11.42it/s]
27
  5%|▌ | 54/1000 [00:06<01:40, 9.45it/s]
28
  6%|▌ | 56/1000 [00:06<01:40, 9.36it/s]
29
  6%|▌ | 58/1000 [00:06<01:26, 10.93it/s]
30
  6%|▌ | 60/1000 [00:07<01:34, 9.97it/s]
31
  6%|▌ | 62/1000 [00:07<01:49, 8.55it/s]
32
  6%|▋ | 64/1000 [00:07<01:32, 10.17it/s]
33
  7%|▋ | 66/1000 [00:07<01:57, 7.98it/s]
34
  7%|▋ | 68/1000 [00:08<01:48, 8.61it/s]
35
  7%|▋ | 70/1000 [00:08<01:40, 9.28it/s]
36
  7%|▋ | 72/1000 [00:08<01:58, 7.85it/s]
37
  7%|▋ | 74/1000 [00:08<01:54, 8.10it/s]
38
  8%|▊ | 76/1000 [00:09<01:37, 9.52it/s]
39
  8%|▊ | 79/1000 [00:09<01:21, 11.31it/s]
40
  8%|▊ | 81/1000 [00:09<01:45, 8.74it/s]
41
  8%|▊ | 83/1000 [00:09<01:36, 9.55it/s]
42
  8%|▊ | 85/1000 [00:09<01:22, 11.10it/s]
43
  9%|▊ | 87/1000 [00:13<08:07, 1.87it/s]
44
  9%|▉ | 88/1000 [00:13<07:05, 2.14it/s]
45
  9%|▉ | 89/1000 [00:13<06:28, 2.35it/s]
46
  9%|▉ | 90/1000 [00:13<05:31, 2.75it/s]
47
  9%|▉ | 92/1000 [00:13<03:53, 3.88it/s]
48
  9%|▉ | 93/1000 [00:14<03:59, 3.78it/s]
49
  10%|▉ | 95/1000 [00:14<03:16, 4.62it/s]
50
  10%|▉ | 98/1000 [00:14<02:17, 6.56it/s]
51
  10%|█ | 100/1000 [00:14<01:52, 7.97it/s]
52
  10%|█ | 102/1000 [00:15<02:12, 6.80it/s]
53
  10%|█ | 104/1000 [00:15<01:51, 8.03it/s]
54
  11%|█ | 106/1000 [00:15<02:04, 7.17it/s]
55
  11%|█ | 107/1000 [00:15<02:08, 6.95it/s]
56
  11%|█ | 109/1000 [00:16<02:03, 7.21it/s]
57
  11%|█ | 110/1000 [00:16<02:07, 6.96it/s]
58
  11%|█ | 112/1000 [00:16<01:40, 8.87it/s]
59
  11%|█▏ | 114/1000 [00:16<01:32, 9.55it/s]
60
  12%|█▏ | 116/1000 [00:16<02:01, 7.29it/s]
61
  12%|█▏ | 119/1000 [00:17<01:30, 9.71it/s]
62
  12%|█▏ | 121/1000 [00:17<01:17, 11.34it/s]
63
  12%|█▏ | 123/1000 [00:17<01:23, 10.48it/s]
64
  12%|█▎ | 125/1000 [00:18<02:21, 6.16it/s]
65
  13%|█▎ | 127/1000 [00:18<01:54, 7.60it/s]
66
  13%|█▎ | 129/1000 [00:18<01:40, 8.66it/s]
67
  13%|█▎ | 131/1000 [00:18<02:03, 7.01it/s]
68
  13%|█▎ | 134/1000 [00:19<03:27, 4.17it/s]
69
  14%|█▎ | 135/1000 [00:20<04:06, 3.50it/s]
70
  14%|█▎ | 136/1000 [00:20<03:44, 3.85it/s]
71
  14%|█▎ | 137/1000 [00:20<03:44, 3.84it/s]
72
  14%|█▍ | 138/1000 [00:21<03:44, 3.84it/s]
73
  14%|█▍ | 139/1000 [00:21<03:11, 4.50it/s]
74
  14%|█▍ | 140/1000 [00:21<02:45, 5.20it/s]
75
  14%|█▍ | 142/1000 [00:21<01:59, 7.21it/s]
76
  14%|█▍ | 144/1000 [00:21<01:41, 8.44it/s]
77
  15%|█▍ | 146/1000 [00:21<01:25, 10.02it/s]
78
  15%|█▍ | 148/1000 [00:21<01:11, 11.85it/s]
79
  15%|█▌ | 150/1000 [00:22<01:12, 11.66it/s]
80
  15%|█▌ | 152/1000 [00:24<06:14, 2.26it/s]
81
  15%|█▌ | 153/1000 [00:24<05:22, 2.63it/s]
82
  16%|█▌ | 156/1000 [00:25<03:48, 3.69it/s]
83
  16%|█▌ | 157/1000 [00:25<03:24, 4.13it/s]
84
  16%|█▌ | 158/1000 [00:25<03:27, 4.06it/s]
85
  16%|█▌ | 159/1000 [00:25<03:01, 4.63it/s]
86
  16%|█▌ | 161/1000 [00:25<02:19, 6.02it/s]
87
  16%|█▋ | 163/1000 [00:25<01:45, 7.94it/s]
88
  16%|█▋ | 165/1000 [00:25<01:29, 9.34it/s]
89
  17%|█▋ | 167/1000 [00:26<02:53, 4.81it/s]
90
  17%|█▋ | 169/1000 [00:26<02:13, 6.25it/s]
91
  17%|█▋ | 171/1000 [00:27<01:56, 7.09it/s]
92
  17%|█▋ | 174/1000 [00:27<01:25, 9.61it/s]
93
  18%|█▊ | 176/1000 [00:27<01:25, 9.65it/s]
94
  18%|█▊ | 178/1000 [00:27<01:42, 8.02it/s]
95
  18%|█▊ | 180/1000 [00:27<01:24, 9.67it/s]
96
  18%|█▊ | 182/1000 [00:27<01:15, 10.86it/s]
97
  18%|█▊ | 184/1000 [00:28<01:18, 10.42it/s]
98
  19%|█▊ | 187/1000 [00:28<01:00, 13.44it/s]
99
  19%|█▉ | 189/1000 [00:28<01:18, 10.35it/s]
100
  19%|█▉ | 191/1000 [00:29<01:46, 7.57it/s]
101
  19%|█▉ | 193/1000 [00:29<01:34, 8.58it/s]
102
  20%|█▉ | 195/1000 [00:29<01:45, 7.63it/s]
103
  20%|█▉ | 196/1000 [00:29<01:49, 7.31it/s]
104
  20%|█▉ | 198/1000 [00:30<01:50, 7.24it/s]
105
  20%|██ | 200/1000 [00:31<04:00, 3.33it/s]
106
  20%|██ | 201/1000 [00:31<03:39, 3.64it/s]
107
  20%|██ | 203/1000 [00:31<02:52, 4.61it/s]
108
  21%|██ | 206/1000 [00:31<02:04, 6.36it/s]
109
  21%|██ | 208/1000 [00:32<01:46, 7.45it/s]
110
  21%|██ | 210/1000 [00:32<02:54, 4.52it/s]
111
  21%|██ | 212/1000 [00:33<02:55, 4.49it/s]
112
  21%|██▏ | 214/1000 [00:33<02:15, 5.78it/s]
113
  22%|██▏ | 216/1000 [00:33<01:50, 7.07it/s]
114
  22%|██▏ | 218/1000 [00:33<01:42, 7.64it/s]
115
  22%|██▏ | 220/1000 [00:33<01:29, 8.76it/s]
116
  22%|██▏ | 223/1000 [00:34<01:13, 10.62it/s]
117
  22%|██▎ | 225/1000 [00:34<01:06, 11.59it/s]
118
  23%|██▎ | 227/1000 [00:35<02:45, 4.67it/s]
119
  23%|██▎ | 229/1000 [00:35<02:13, 5.79it/s]
120
  23%|██▎ | 231/1000 [00:37<05:12, 2.46it/s]
121
  23%|██▎ | 233/1000 [00:37<04:11, 3.05it/s]
122
  24%|██▎ | 235/1000 [00:37<03:21, 3.79it/s]
123
  24%|██▎ | 237/1000 [00:38<02:38, 4.80it/s]
124
  24%|██▍ | 239/1000 [00:38<02:06, 6.02it/s]
125
  24%|██▍ | 241/1000 [00:38<01:44, 7.27it/s]
126
  24%|██▍ | 243/1000 [00:38<01:32, 8.15it/s]
127
  24%|██▍ | 245/1000 [00:38<01:44, 7.20it/s]
128
  25%|██▍ | 247/1000 [00:39<01:31, 8.24it/s]
129
  25%|██▍ | 249/1000 [00:39<01:49, 6.83it/s]
130
  25%|██▌ | 251/1000 [00:39<01:32, 8.07it/s]
131
  25%|██▌ | 253/1000 [00:39<01:19, 9.38it/s]
132
  26%|██▌ | 255/1000 [00:40<01:26, 8.57it/s]
133
  26%|██▌ | 257/1000 [00:40<02:23, 5.19it/s]
134
  26%|██▌ | 258/1000 [00:40<02:15, 5.47it/s]
135
  26%|██▌ | 260/1000 [00:41<01:43, 7.12it/s]
136
  26%|██▌ | 262/1000 [00:41<01:48, 6.78it/s]
137
  26%|██▋ | 263/1000 [00:41<02:08, 5.74it/s]
138
  26%|██▋ | 265/1000 [00:41<01:45, 6.98it/s]
139
  27%|██▋ | 266/1000 [00:42<01:48, 6.77it/s]
140
  27%|██▋ | 267/1000 [00:42<01:41, 7.19it/s]
141
  27%|██▋ | 269/1000 [00:42<01:20, 9.13it/s]
142
  27%|██▋ | 271/1000 [00:42<01:40, 7.22it/s]
143
  27%|██▋ | 273/1000 [00:42<01:46, 6.82it/s]
144
  27%|██▋ | 274/1000 [00:43<01:40, 7.20it/s]
145
  28%|██▊ | 275/1000 [00:43<01:52, 6.43it/s]
146
  28%|██▊ | 277/1000 [00:43<01:30, 7.97it/s]
147
  28%|██▊ | 278/1000 [00:43<01:27, 8.24it/s]
148
  28%|██▊ | 279/1000 [00:43<01:42, 7.01it/s]
149
  28%|██▊ | 282/1000 [00:43<01:07, 10.64it/s]
150
  28%|██▊ | 284/1000 [00:44<01:58, 6.02it/s]
151
  28%|██▊ | 285/1000 [00:44<02:09, 5.50it/s]
152
  29%|██▊ | 287/1000 [00:45<02:09, 5.52it/s]
153
  29%|██▉ | 290/1000 [00:45<01:37, 7.26it/s]
154
  29%|██▉ | 292/1000 [00:45<01:25, 8.33it/s]
155
  29%|██▉ | 294/1000 [00:46<01:53, 6.23it/s]
156
  30%|██▉ | 296/1000 [00:46<01:37, 7.24it/s]
157
  30%|██▉ | 299/1000 [00:47<02:51, 4.08it/s]
158
  30%|███ | 301/1000 [00:47<02:14, 5.18it/s]
159
  30%|███ | 303/1000 [00:47<02:08, 5.42it/s]
160
  30%|███ | 305/1000 [00:48<02:04, 5.60it/s]
161
  31%|███ | 307/1000 [00:48<01:38, 7.03it/s]
162
  31%|███ | 309/1000 [00:48<01:33, 7.43it/s]
163
  31%|███ | 311/1000 [00:48<01:24, 8.12it/s]
164
  31%|███▏ | 314/1000 [00:49<01:17, 8.85it/s]
165
  32%|███▏ | 316/1000 [00:49<01:32, 7.43it/s]
166
  32%|███▏ | 317/1000 [00:50<03:11, 3.57it/s]
167
  32%|███▏ | 319/1000 [00:50<02:27, 4.62it/s]
168
  32%|███▏ | 320/1000 [00:52<05:20, 2.12it/s]
169
  32%|███▏ | 321/1000 [00:52<05:17, 2.14it/s]
170
  32%|███▏ | 324/1000 [00:53<03:37, 3.10it/s]
171
  32%|███▎ | 325/1000 [00:53<03:11, 3.53it/s]
172
  33%|███▎ | 326/1000 [00:53<02:54, 3.87it/s]
173
  33%|███▎ | 327/1000 [00:53<02:46, 4.05it/s]
174
  33%|███▎ | 328/1000 [00:53<02:28, 4.53it/s]
175
  33%|███▎ | 330/1000 [00:53<01:44, 6.40it/s]
176
  33%|███▎ | 331/1000 [00:54<01:45, 6.37it/s]
177
  33%|███▎ | 332/1000 [00:54<01:49, 6.10it/s]
178
  33%|███▎ | 334/1000 [00:54<01:18, 8.47it/s]
179
  34%|███▎ | 336/1000 [00:54<01:10, 9.38it/s]
180
  34%|███▍ | 338/1000 [00:54<01:13, 9.06it/s]
181
  34%|███▍ | 340/1000 [00:54<01:00, 10.94it/s]
182
  34%|███▍ | 342/1000 [00:55<01:11, 9.17it/s]
183
  34%|███▍ | 344/1000 [00:55<01:01, 10.67it/s]
184
  35%|███▍ | 346/1000 [00:55<00:53, 12.19it/s]
185
  35%|███▍ | 348/1000 [00:56<02:08, 5.06it/s]
186
  35%|███▌ | 350/1000 [00:57<02:51, 3.79it/s]
187
  35%|███▌ | 353/1000 [00:57<02:52, 3.76it/s]
188
  36%|███▌ | 356/1000 [00:58<02:09, 4.96it/s]
189
  36%|███▌ | 358/1000 [00:58<01:45, 6.09it/s]
190
  36%|███▌ | 360/1000 [00:58<01:34, 6.76it/s]
191
  36%|███▌ | 362/1000 [00:58<01:23, 7.61it/s]
192
  36%|███▋ | 364/1000 [00:58<01:09, 9.11it/s]
193
  37%|███▋ | 366/1000 [00:59<02:16, 4.65it/s]
194
  37%|███▋ | 368/1000 [01:00<02:05, 5.02it/s]
195
  37%|███▋ | 370/1000 [01:00<01:45, 5.98it/s]
196
  37%|███▋ | 371/1000 [01:00<01:38, 6.37it/s]
197
  37%|███▋ | 373/1000 [01:00<01:25, 7.30it/s]
198
  37%|███▋ | 374/1000 [01:00<01:41, 6.18it/s]
199
  38%|███▊ | 376/1000 [01:12<02:01, 5.16it/s]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+
3
+
4
+
5
+
6
+ [codecarbon INFO @ 16:25:56] [setup] RAM Tracking...
7
+ [codecarbon INFO @ 16:25:56] [setup] GPU Tracking...
8
+ [codecarbon INFO @ 16:25:56] Tracking Nvidia GPU via pynvml
9
+ [codecarbon DEBUG @ 16:25:56] GPU available. Starting setup
10
+ [codecarbon INFO @ 16:25:56] [setup] CPU Tracking...
11
+ [codecarbon DEBUG @ 16:25:56] Not using PowerGadget, an exception occurred while instantiating IntelPowerGadget : Platform not supported by Intel Power Gadget
12
+ [codecarbon DEBUG @ 16:25:56] Not using the RAPL interface, an exception occurred while instantiating IntelRAPL : Intel RAPL files not found at /sys/class/powercap/intel-rapl on linux
13
+ [codecarbon DEBUG @ 16:25:56] Not using PowerMetrics, an exception occurred while instantiating Powermetrics : Platform not supported by Powermetrics
14
+ [codecarbon WARNING @ 16:25:56] No CPU tracking mode found. Falling back on CPU constant mode.
15
+ [codecarbon WARNING @ 16:25:58] We saw that you have a AMD EPYC 7R32 but we don't know it. Please contact us.
16
+ [codecarbon INFO @ 16:25:58] CPU Model on constant consumption mode: AMD EPYC 7R32
17
+ [codecarbon INFO @ 16:25:58] >>> Tracker's metadata:
18
+ [codecarbon INFO @ 16:25:58] Platform system: Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35
19
+ [codecarbon INFO @ 16:25:58] Python version: 3.9.20
20
+ [codecarbon INFO @ 16:25:58] CodeCarbon version: 2.5.1
21
+ [codecarbon INFO @ 16:25:58] Available RAM : 186.705 GB
22
+ [codecarbon INFO @ 16:25:58] CPU count: 48
23
+ [codecarbon INFO @ 16:25:58] CPU model: AMD EPYC 7R32
24
+ [codecarbon INFO @ 16:25:58] GPU count: 1
25
+ [codecarbon INFO @ 16:25:58] GPU model: 1 x NVIDIA A10G
26
+ [codecarbon DEBUG @ 16:25:59] Not running on AWS
27
+ [codecarbon DEBUG @ 16:26:00] Not running on Azure
28
+ [codecarbon DEBUG @ 16:26:01] Not running on GCP
29
+ [codecarbon INFO @ 16:26:01] Saving emissions data to file /runs/text_generation/meta-llama/Llama-3.2-3B/2024-10-24-16-25-47/codecarbon.csv
30
+ [codecarbon DEBUG @ 16:26:01] EmissionsData(timestamp='2024-10-24T16:26:01', project_name='codecarbon', run_id='2e624837-816c-4a54-b414-c29d24dd9f24', duration=0.00212410103995353, emissions=0.0, emissions_rate=0.0, cpu_power=0.0, gpu_power=0.0, ram_power=0.0, cpu_energy=0, gpu_energy=0, ram_energy=0, energy_consumed=0, country_name='United States', country_iso_code='USA', region='virginia', cloud_provider='', cloud_region='', os='Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35', python_version='3.9.20', codecarbon_version='2.5.1', cpu_count=48, cpu_model='AMD EPYC 7R32', gpu_count=1, gpu_model='1 x NVIDIA A10G', longitude=-77.4903, latitude=39.0469, ram_total_size=186.7047882080078, tracking_mode='process', on_cloud='N', pue=1.0)
31
+
32
+
33
+ [codecarbon INFO @ 16:26:02] Energy consumed for RAM : 0.000000 kWh. RAM Power : 0.35121631622314453 W
34
+ [codecarbon DEBUG @ 16:26:02] RAM : 0.35 W during 0.90 s [measurement time: 0.0007]
35
+ [codecarbon INFO @ 16:26:02] Energy consumed for all GPUs : 0.000016 kWh. Total GPU Power : 65.0621770882058 W
36
+ [codecarbon DEBUG @ 16:26:02] GPU : 65.06 W during 0.90 s [measurement time: 0.0023]
37
+ [codecarbon INFO @ 16:26:02] Energy consumed for all CPUs : 0.000011 kWh. Total CPU Power : 42.5 W
38
+ [codecarbon DEBUG @ 16:26:02] CPU : 42.50 W during 0.90 s [measurement time: 0.0001]
39
+ [codecarbon INFO @ 16:26:02] 0.000027 kWh of electricity used since the beginning.
40
+ [codecarbon DEBUG @ 16:26:02] last_duration=0.899167487048544
41
+ ------------------------
42
+ [codecarbon DEBUG @ 16:26:02] EmissionsData(timestamp='2024-10-24T16:26:02', project_name='codecarbon', run_id='2e624837-816c-4a54-b414-c29d24dd9f24', duration=0.9025245610391721, emissions=9.969546314489585e-06, emissions_rate=1.1046288095484727e-05, cpu_power=42.5, gpu_power=65.0621770882058, ram_power=0.35121631622314453, cpu_energy=1.065322435606504e-05, gpu_energy=1.6266957458377362e-05, ram_energy=8.772346779188348e-08, energy_consumed=2.7007905282234288e-05, country_name='United States', country_iso_code='USA', region='virginia', cloud_provider='', cloud_region='', os='Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35', python_version='3.9.20', codecarbon_version='2.5.1', cpu_count=48, cpu_model='AMD EPYC 7R32', gpu_count=1, gpu_model='1 x NVIDIA A10G', longitude=-77.4903, latitude=39.0469, ram_total_size=186.7047882080078, tracking_mode='process', on_cloud='N', pue=1.0)
43
+ [codecarbon DEBUG @ 16:26:04] EmissionsData(timestamp='2024-10-24T16:26:04', project_name='codecarbon', run_id='2e624837-816c-4a54-b414-c29d24dd9f24', duration=0.002263403031975031, emissions=9.969546314489585e-06, emissions_rate=0.004404671273144944, cpu_power=42.5, gpu_power=65.0621770882058, ram_power=0.35121631622314453, cpu_energy=1.065322435606504e-05, gpu_energy=1.6266957458377362e-05, ram_energy=8.772346779188348e-08, energy_consumed=2.7007905282234288e-05, country_name='United States', country_iso_code='USA', region='virginia', cloud_provider='', cloud_region='', os='Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35', python_version='3.9.20', codecarbon_version='2.5.1', cpu_count=48, cpu_model='AMD EPYC 7R32', gpu_count=1, gpu_model='1 x NVIDIA A10G', longitude=-77.4903, latitude=39.0469, ram_total_size=186.7047882080078, tracking_mode='process', on_cloud='N', pue=1.0)
44
+
45
  0%| | 0/1000 [00:00<?, ?it/s]
46
  0%| | 1/1000 [00:00<02:15, 7.40it/s]
47
  0%| | 3/1000 [00:00<01:30, 11.00it/s]
48
  0%| | 5/1000 [00:00<02:48, 5.90it/s]
49
  1%| | 7/1000 [00:00<02:10, 7.63it/s]
50
  1%| | 8/1000 [00:02<06:03, 2.73it/s]
51
  1%| | 10/1000 [00:02<04:08, 3.99it/s]
52
  1%| | 11/1000 [00:02<04:52, 3.38it/s]
53
  1%|▏ | 13/1000 [00:02<03:27, 4.76it/s]
54
  1%|▏ | 14/1000 [00:03<03:48, 4.31it/s]
55
  2%|▏ | 15/1000 [00:03<03:32, 4.63it/s]
56
  2%|▏ | 18/1000 [00:03<02:10, 7.53it/s]
57
  2%|▏ | 20/1000 [00:03<01:49, 8.91it/s]
58
  2%|▏ | 23/1000 [00:03<01:23, 11.74it/s]
59
  2%|▎ | 25/1000 [00:03<01:13, 13.21it/s]
60
  3%|▎ | 27/1000 [00:04<01:42, 9.54it/s]
61
  3%|▎ | 30/1000 [00:04<01:17, 12.49it/s]
62
  3%|▎ | 32/1000 [00:04<01:16, 12.65it/s]
63
  4%|▎ | 35/1000 [00:04<01:12, 13.24it/s]
64
  4%|▎ | 37/1000 [00:04<01:11, 13.45it/s]
65
  4%|▍ | 39/1000 [00:05<01:33, 10.24it/s]
66
  4%|▍ | 41/1000 [00:05<01:57, 8.19it/s]
67
  4%|▍ | 43/1000 [00:05<01:37, 9.78it/s]
68
  5%|▍ | 46/1000 [00:05<01:17, 12.38it/s]
69
  5%|▍ | 48/1000 [00:05<01:18, 12.07it/s]
70
  5%|▌ | 50/1000 [00:06<01:32, 10.25it/s]
71
  5%|▌ | 52/1000 [00:06<01:22, 11.42it/s]
72
  5%|▌ | 54/1000 [00:06<01:40, 9.45it/s]
73
  6%|▌ | 56/1000 [00:06<01:40, 9.36it/s]
74
  6%|▌ | 58/1000 [00:06<01:26, 10.93it/s]
75
  6%|▌ | 60/1000 [00:07<01:34, 9.97it/s]
76
  6%|▌ | 62/1000 [00:07<01:49, 8.55it/s]
77
  6%|▋ | 64/1000 [00:07<01:32, 10.17it/s]
78
  7%|▋ | 66/1000 [00:07<01:57, 7.98it/s]
79
  7%|▋ | 68/1000 [00:08<01:48, 8.61it/s]
80
  7%|▋ | 70/1000 [00:08<01:40, 9.28it/s]
81
  7%|▋ | 72/1000 [00:08<01:58, 7.85it/s]
82
  7%|▋ | 74/1000 [00:08<01:54, 8.10it/s]
83
  8%|▊ | 76/1000 [00:09<01:37, 9.52it/s]
84
  8%|▊ | 79/1000 [00:09<01:21, 11.31it/s]
85
  8%|▊ | 81/1000 [00:09<01:45, 8.74it/s]
86
  8%|▊ | 83/1000 [00:09<01:36, 9.55it/s]
87
  8%|▊ | 85/1000 [00:09<01:22, 11.10it/s]
88
  9%|▊ | 87/1000 [00:13<08:07, 1.87it/s]
89
  9%|▉ | 88/1000 [00:13<07:05, 2.14it/s]
90
  9%|▉ | 89/1000 [00:13<06:28, 2.35it/s]
91
  9%|▉ | 90/1000 [00:13<05:31, 2.75it/s]
92
  9%|▉ | 92/1000 [00:13<03:53, 3.88it/s]
93
  9%|▉ | 93/1000 [00:14<03:59, 3.78it/s]
94
  10%|▉ | 95/1000 [00:14<03:16, 4.62it/s]
95
  10%|▉ | 98/1000 [00:14<02:17, 6.56it/s]
96
  10%|█ | 100/1000 [00:14<01:52, 7.97it/s]
97
  10%|█ | 102/1000 [00:15<02:12, 6.80it/s]
98
  10%|█ | 104/1000 [00:15<01:51, 8.03it/s]
99
  11%|█ | 106/1000 [00:15<02:04, 7.17it/s]
100
  11%|█ | 107/1000 [00:15<02:08, 6.95it/s]
101
  11%|█ | 109/1000 [00:16<02:03, 7.21it/s]
102
  11%|█ | 110/1000 [00:16<02:07, 6.96it/s]
103
  11%|█ | 112/1000 [00:16<01:40, 8.87it/s]
104
  11%|█▏ | 114/1000 [00:16<01:32, 9.55it/s]
105
  12%|█▏ | 116/1000 [00:16<02:01, 7.29it/s]
106
  12%|█▏ | 119/1000 [00:17<01:30, 9.71it/s]
107
  12%|█▏ | 121/1000 [00:17<01:17, 11.34it/s]
108
  12%|█▏ | 123/1000 [00:17<01:23, 10.48it/s]
109
  12%|█▎ | 125/1000 [00:18<02:21, 6.16it/s]
110
  13%|█▎ | 127/1000 [00:18<01:54, 7.60it/s]
111
  13%|█▎ | 129/1000 [00:18<01:40, 8.66it/s]
112
  13%|█▎ | 131/1000 [00:18<02:03, 7.01it/s]
113
  13%|█▎ | 134/1000 [00:19<03:27, 4.17it/s]
114
  14%|█▎ | 135/1000 [00:20<04:06, 3.50it/s]
115
  14%|█▎ | 136/1000 [00:20<03:44, 3.85it/s]
116
  14%|█▎ | 137/1000 [00:20<03:44, 3.84it/s]
117
  14%|█▍ | 138/1000 [00:21<03:44, 3.84it/s]
118
  14%|█▍ | 139/1000 [00:21<03:11, 4.50it/s]
119
  14%|█▍ | 140/1000 [00:21<02:45, 5.20it/s]
120
  14%|█▍ | 142/1000 [00:21<01:59, 7.21it/s]
121
  14%|█▍ | 144/1000 [00:21<01:41, 8.44it/s]
122
  15%|█▍ | 146/1000 [00:21<01:25, 10.02it/s]
123
  15%|█▍ | 148/1000 [00:21<01:11, 11.85it/s]
124
  15%|█▌ | 150/1000 [00:22<01:12, 11.66it/s]
125
  15%|█▌ | 152/1000 [00:24<06:14, 2.26it/s]
126
  15%|█▌ | 153/1000 [00:24<05:22, 2.63it/s]
127
  16%|█▌ | 156/1000 [00:25<03:48, 3.69it/s]
128
  16%|█▌ | 157/1000 [00:25<03:24, 4.13it/s]
129
  16%|█▌ | 158/1000 [00:25<03:27, 4.06it/s]
130
  16%|█▌ | 159/1000 [00:25<03:01, 4.63it/s]
131
  16%|█▌ | 161/1000 [00:25<02:19, 6.02it/s]
132
  16%|█▋ | 163/1000 [00:25<01:45, 7.94it/s]
133
  16%|█▋ | 165/1000 [00:25<01:29, 9.34it/s]
134
  17%|█▋ | 167/1000 [00:26<02:53, 4.81it/s]
135
  17%|█▋ | 169/1000 [00:26<02:13, 6.25it/s]
136
  17%|█▋ | 171/1000 [00:27<01:56, 7.09it/s]
137
  17%|█▋ | 174/1000 [00:27<01:25, 9.61it/s]
138
  18%|█▊ | 176/1000 [00:27<01:25, 9.65it/s]
139
  18%|█▊ | 178/1000 [00:27<01:42, 8.02it/s]
140
  18%|█▊ | 180/1000 [00:27<01:24, 9.67it/s]
141
  18%|█▊ | 182/1000 [00:27<01:15, 10.86it/s]
142
  18%|█▊ | 184/1000 [00:28<01:18, 10.42it/s]
143
  19%|█▊ | 187/1000 [00:28<01:00, 13.44it/s]
144
  19%|█▉ | 189/1000 [00:28<01:18, 10.35it/s]
145
  19%|█▉ | 191/1000 [00:29<01:46, 7.57it/s]
146
  19%|█▉ | 193/1000 [00:29<01:34, 8.58it/s]
147
  20%|█▉ | 195/1000 [00:29<01:45, 7.63it/s]
148
  20%|█▉ | 196/1000 [00:29<01:49, 7.31it/s]
149
  20%|█▉ | 198/1000 [00:30<01:50, 7.24it/s]
150
  20%|██ | 200/1000 [00:31<04:00, 3.33it/s]
151
  20%|██ | 201/1000 [00:31<03:39, 3.64it/s]
152
  20%|██ | 203/1000 [00:31<02:52, 4.61it/s]
153
  21%|██ | 206/1000 [00:31<02:04, 6.36it/s]
154
  21%|██ | 208/1000 [00:32<01:46, 7.45it/s]
155
  21%|██ | 210/1000 [00:32<02:54, 4.52it/s]
156
  21%|██ | 212/1000 [00:33<02:55, 4.49it/s]
157
  21%|██▏ | 214/1000 [00:33<02:15, 5.78it/s]
158
  22%|██▏ | 216/1000 [00:33<01:50, 7.07it/s]
159
  22%|██▏ | 218/1000 [00:33<01:42, 7.64it/s]
160
  22%|██▏ | 220/1000 [00:33<01:29, 8.76it/s]
161
  22%|██▏ | 223/1000 [00:34<01:13, 10.62it/s]
162
  22%|██▎ | 225/1000 [00:34<01:06, 11.59it/s]
163
  23%|██▎ | 227/1000 [00:35<02:45, 4.67it/s]
164
  23%|██▎ | 229/1000 [00:35<02:13, 5.79it/s]
165
  23%|██▎ | 231/1000 [00:37<05:12, 2.46it/s]
166
  23%|██▎ | 233/1000 [00:37<04:11, 3.05it/s]
167
  24%|██▎ | 235/1000 [00:37<03:21, 3.79it/s]
168
  24%|██▎ | 237/1000 [00:38<02:38, 4.80it/s]
169
  24%|██▍ | 239/1000 [00:38<02:06, 6.02it/s]
170
  24%|██▍ | 241/1000 [00:38<01:44, 7.27it/s]
171
  24%|██▍ | 243/1000 [00:38<01:32, 8.15it/s]
172
  24%|██▍ | 245/1000 [00:38<01:44, 7.20it/s]
173
  25%|██▍ | 247/1000 [00:39<01:31, 8.24it/s]
174
  25%|██▍ | 249/1000 [00:39<01:49, 6.83it/s]
175
  25%|██▌ | 251/1000 [00:39<01:32, 8.07it/s]
176
  25%|██▌ | 253/1000 [00:39<01:19, 9.38it/s]
177
  26%|██▌ | 255/1000 [00:40<01:26, 8.57it/s]
178
  26%|██▌ | 257/1000 [00:40<02:23, 5.19it/s]
179
  26%|██▌ | 258/1000 [00:40<02:15, 5.47it/s]
180
  26%|██▌ | 260/1000 [00:41<01:43, 7.12it/s]
181
  26%|██▌ | 262/1000 [00:41<01:48, 6.78it/s]
182
  26%|██▋ | 263/1000 [00:41<02:08, 5.74it/s]
183
  26%|██▋ | 265/1000 [00:41<01:45, 6.98it/s]
184
  27%|██▋ | 266/1000 [00:42<01:48, 6.77it/s]
185
  27%|██▋ | 267/1000 [00:42<01:41, 7.19it/s]
186
  27%|██▋ | 269/1000 [00:42<01:20, 9.13it/s]
187
  27%|██▋ | 271/1000 [00:42<01:40, 7.22it/s]
188
  27%|██▋ | 273/1000 [00:42<01:46, 6.82it/s]
189
  27%|██▋ | 274/1000 [00:43<01:40, 7.20it/s]
190
  28%|██▊ | 275/1000 [00:43<01:52, 6.43it/s]
191
  28%|██▊ | 277/1000 [00:43<01:30, 7.97it/s]
192
  28%|██▊ | 278/1000 [00:43<01:27, 8.24it/s]
193
  28%|██▊ | 279/1000 [00:43<01:42, 7.01it/s]
194
  28%|██▊ | 282/1000 [00:43<01:07, 10.64it/s]
195
  28%|██▊ | 284/1000 [00:44<01:58, 6.02it/s]
196
  28%|██▊ | 285/1000 [00:44<02:09, 5.50it/s]
197
  29%|██▊ | 287/1000 [00:45<02:09, 5.52it/s]
198
  29%|██▉ | 290/1000 [00:45<01:37, 7.26it/s]
199
  29%|██▉ | 292/1000 [00:45<01:25, 8.33it/s]
200
  29%|██▉ | 294/1000 [00:46<01:53, 6.23it/s]
201
  30%|██▉ | 296/1000 [00:46<01:37, 7.24it/s]
202
  30%|██▉ | 299/1000 [00:47<02:51, 4.08it/s]
203
  30%|███ | 301/1000 [00:47<02:14, 5.18it/s]
204
  30%|███ | 303/1000 [00:47<02:08, 5.42it/s]
205
  30%|███ | 305/1000 [00:48<02:04, 5.60it/s]
206
  31%|███ | 307/1000 [00:48<01:38, 7.03it/s]
207
  31%|███ | 309/1000 [00:48<01:33, 7.43it/s]
208
  31%|███ | 311/1000 [00:48<01:24, 8.12it/s]
209
  31%|███▏ | 314/1000 [00:49<01:17, 8.85it/s]
210
  32%|███▏ | 316/1000 [00:49<01:32, 7.43it/s]
211
  32%|███▏ | 317/1000 [00:50<03:11, 3.57it/s]
212
  32%|███▏ | 319/1000 [00:50<02:27, 4.62it/s]
213
  32%|███▏ | 320/1000 [00:52<05:20, 2.12it/s]
214
  32%|███▏ | 321/1000 [00:52<05:17, 2.14it/s]
215
  32%|███▏ | 324/1000 [00:53<03:37, 3.10it/s]
216
  32%|███▎ | 325/1000 [00:53<03:11, 3.53it/s]
217
  33%|███▎ | 326/1000 [00:53<02:54, 3.87it/s]
218
  33%|███▎ | 327/1000 [00:53<02:46, 4.05it/s]
219
  33%|███▎ | 328/1000 [00:53<02:28, 4.53it/s]
220
  33%|███▎ | 330/1000 [00:53<01:44, 6.40it/s]
221
  33%|███▎ | 331/1000 [00:54<01:45, 6.37it/s]
222
  33%|███▎ | 332/1000 [00:54<01:49, 6.10it/s]
223
  33%|███▎ | 334/1000 [00:54<01:18, 8.47it/s]
224
  34%|███▎ | 336/1000 [00:54<01:10, 9.38it/s]
225
  34%|███▍ | 338/1000 [00:54<01:13, 9.06it/s]
226
  34%|███▍ | 340/1000 [00:54<01:00, 10.94it/s]
227
  34%|███▍ | 342/1000 [00:55<01:11, 9.17it/s]
228
  34%|███▍ | 344/1000 [00:55<01:01, 10.67it/s]
229
  35%|███▍ | 346/1000 [00:55<00:53, 12.19it/s]
230
  35%|███▍ | 348/1000 [00:56<02:08, 5.06it/s]
231
  35%|███▌ | 350/1000 [00:57<02:51, 3.79it/s]
232
  35%|███▌ | 353/1000 [00:57<02:52, 3.76it/s]
233
  36%|███▌ | 356/1000 [00:58<02:09, 4.96it/s]
234
  36%|███▌ | 358/1000 [00:58<01:45, 6.09it/s]
235
  36%|███▌ | 360/1000 [00:58<01:34, 6.76it/s]
236
  36%|███▌ | 362/1000 [00:58<01:23, 7.61it/s]
237
  36%|███▋ | 364/1000 [00:58<01:09, 9.11it/s]
238
  37%|███▋ | 366/1000 [00:59<02:16, 4.65it/s]
239
  37%|███▋ | 368/1000 [01:00<02:05, 5.02it/s]
240
  37%|███▋ | 370/1000 [01:00<01:45, 5.98it/s]
241
  37%|███▋ | 371/1000 [01:00<01:38, 6.37it/s]
242
  37%|███▋ | 373/1000 [01:00<01:25, 7.30it/s]
243
  37%|███▋ | 374/1000 [01:00<01:41, 6.18it/s]
244
  38%|███▊ | 376/1000 [01:12<02:01, 5.16it/s]
245
+ Error executing job with overrides: ['backend.model=meta-llama/Llama-3.2-3B', 'backend.processor=meta-llama/Llama-3.2-3B']
246
+ Traceback (most recent call last):
247
+ File "/optimum-benchmark/optimum_benchmark/cli.py", line 65, in benchmark_cli
248
+ benchmark_report: BenchmarkReport = launch(experiment_config=experiment_config)
249
+ File "/optimum-benchmark/optimum_benchmark/experiment.py", line 102, in launch
250
+ raise error
251
+ File "/optimum-benchmark/optimum_benchmark/experiment.py", line 90, in launch
252
+ report = launcher.launch(run, experiment_config.benchmark, experiment_config.backend)
253
+ File "/optimum-benchmark/optimum_benchmark/launchers/process/launcher.py", line 47, in launch
254
+ while not process_context.join():
255
+ File "/opt/conda/lib/python3.9/site-packages/torch/multiprocessing/spawn.py", line 189, in join
256
+ raise ProcessRaisedException(msg, error_index, failed_process.pid)
257
+ torch.multiprocessing.spawn.ProcessRaisedException:
258
+
259
+ -- Process 0 terminated with the following error:
260
+ Traceback (most recent call last):
261
+ File "/opt/conda/lib/python3.9/site-packages/torch/multiprocessing/spawn.py", line 76, in _wrap
262
+ fn(i, *args)
263
+ File "/optimum-benchmark/optimum_benchmark/launchers/process/launcher.py", line 63, in entrypoint
264
+ worker_output = worker(*worker_args)
265
+ File "/optimum-benchmark/optimum_benchmark/experiment.py", line 62, in run
266
+ benchmark.run(backend)
267
+ File "/optimum-benchmark/optimum_benchmark/benchmarks/energy_star/benchmark.py", line 174, in run
268
+ self.run_text_generation_energy_tracking(backend)
269
+ File "/optimum-benchmark/optimum_benchmark/benchmarks/energy_star/benchmark.py", line 198, in run_text_generation_energy_tracking
270
+ _ = backend.prefill(inputs, prefill_kwargs)
271
+ File "/opt/conda/lib/python3.9/site-packages/torch/utils/_contextlib.py", line 116, in decorate_context
272
+ return func(*args, **kwargs)
273
+ File "/optimum-benchmark/optimum_benchmark/backends/pytorch/backend.py", line 350, in prefill
274
+ return self.pretrained_model.generate(**inputs, **kwargs)
275
+ File "/opt/conda/lib/python3.9/site-packages/torch/utils/_contextlib.py", line 116, in decorate_context
276
+ return func(*args, **kwargs)
277
+ File "/opt/conda/lib/python3.9/site-packages/transformers/generation/utils.py", line 2024, in generate
278
+ result = self._sample(
279
+ File "/opt/conda/lib/python3.9/site-packages/transformers/generation/utils.py", line 2982, in _sample
280
+ outputs = self(**model_inputs, return_dict=True)
281
+ File "/opt/conda/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1553, in _wrapped_call_impl
282
+ return self._call_impl(*args, **kwargs)
283
+ File "/opt/conda/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1562, in _call_impl
284
+ return forward_call(*args, **kwargs)
285
+ File "/opt/conda/lib/python3.9/site-packages/transformers/models/llama/modeling_llama.py", line 1208, in forward
286
+ logits = self.lm_head(hidden_states)
287
+ File "/opt/conda/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1553, in _wrapped_call_impl
288
+ return self._call_impl(*args, **kwargs)
289
+ File "/opt/conda/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1562, in _call_impl
290
+ return forward_call(*args, **kwargs)
291
+ File "/opt/conda/lib/python3.9/site-packages/torch/nn/modules/linear.py", line 117, in forward
292
+ return F.linear(input, self.weight, self.bias)
293
+ torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 8.57 GiB. GPU 0 has a total capacity of 22.19 GiB of which 3.10 GiB is free. Process 899737 has 19.09 GiB memory in use. Of the allocated memory 16.01 GiB is allocated by PyTorch, and 2.78 GiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)
294
+
295
+
296
+ Set the environment variable HYDRA_FULL_ERROR=1 for a complete stack trace.
text_generation/meta-llama/Llama-3.2-3B/2024-10-24-16-25-47/experiment_config.json ADDED
@@ -0,0 +1,110 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "experiment_name": "text_generation",
3
+ "backend": {
4
+ "name": "pytorch",
5
+ "version": "2.4.0",
6
+ "_target_": "optimum_benchmark.backends.pytorch.backend.PyTorchBackend",
7
+ "task": "text-generation",
8
+ "model": "meta-llama/Llama-3.2-3B",
9
+ "processor": "meta-llama/Llama-3.2-3B",
10
+ "library": "transformers",
11
+ "device": "cuda",
12
+ "device_ids": "0",
13
+ "seed": 42,
14
+ "inter_op_num_threads": null,
15
+ "intra_op_num_threads": null,
16
+ "hub_kwargs": {
17
+ "revision": "main",
18
+ "force_download": false,
19
+ "local_files_only": false,
20
+ "trust_remote_code": true
21
+ },
22
+ "no_weights": true,
23
+ "device_map": null,
24
+ "torch_dtype": null,
25
+ "amp_autocast": false,
26
+ "amp_dtype": null,
27
+ "eval_mode": true,
28
+ "to_bettertransformer": false,
29
+ "low_cpu_mem_usage": null,
30
+ "attn_implementation": null,
31
+ "cache_implementation": null,
32
+ "torch_compile": false,
33
+ "torch_compile_config": {},
34
+ "quantization_scheme": null,
35
+ "quantization_config": {},
36
+ "deepspeed_inference": false,
37
+ "deepspeed_inference_config": {},
38
+ "peft_type": null,
39
+ "peft_config": {}
40
+ },
41
+ "launcher": {
42
+ "name": "process",
43
+ "_target_": "optimum_benchmark.launchers.process.launcher.ProcessLauncher",
44
+ "device_isolation": false,
45
+ "device_isolation_action": "warn",
46
+ "start_method": "spawn"
47
+ },
48
+ "benchmark": {
49
+ "name": "energy_star",
50
+ "_target_": "optimum_benchmark.benchmarks.energy_star.benchmark.EnergyStarBenchmark",
51
+ "dataset_name": "EnergyStarAI/text_generation",
52
+ "dataset_config": "",
53
+ "dataset_split": "train",
54
+ "num_samples": 1000,
55
+ "input_shapes": {
56
+ "batch_size": 1
57
+ },
58
+ "text_column_name": "text",
59
+ "truncation": true,
60
+ "max_length": -1,
61
+ "dataset_prefix1": "",
62
+ "dataset_prefix2": "",
63
+ "t5_task": "",
64
+ "image_column_name": "image",
65
+ "resize": false,
66
+ "question_column_name": "question",
67
+ "context_column_name": "context",
68
+ "sentence1_column_name": "sentence1",
69
+ "sentence2_column_name": "sentence2",
70
+ "audio_column_name": "audio",
71
+ "iterations": 10,
72
+ "warmup_runs": 10,
73
+ "energy": true,
74
+ "forward_kwargs": {},
75
+ "generate_kwargs": {
76
+ "max_new_tokens": 10,
77
+ "min_new_tokens": 10
78
+ },
79
+ "call_kwargs": {}
80
+ },
81
+ "environment": {
82
+ "cpu": " AMD EPYC 7R32",
83
+ "cpu_count": 48,
84
+ "cpu_ram_mb": 200472.73984,
85
+ "system": "Linux",
86
+ "machine": "x86_64",
87
+ "platform": "Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35",
88
+ "processor": "x86_64",
89
+ "python_version": "3.9.20",
90
+ "gpu": [
91
+ "NVIDIA A10G"
92
+ ],
93
+ "gpu_count": 1,
94
+ "gpu_vram_mb": 24146608128,
95
+ "optimum_benchmark_version": "0.2.0",
96
+ "optimum_benchmark_commit": null,
97
+ "transformers_version": "4.44.0",
98
+ "transformers_commit": null,
99
+ "accelerate_version": "0.33.0",
100
+ "accelerate_commit": null,
101
+ "diffusers_version": "0.30.0",
102
+ "diffusers_commit": null,
103
+ "optimum_version": null,
104
+ "optimum_commit": null,
105
+ "timm_version": null,
106
+ "timm_commit": null,
107
+ "peft_version": null,
108
+ "peft_commit": null
109
+ }
110
+ }
text_generation/meta-llama/Llama-3.2-3B/2024-10-24-16-25-47/preprocess_codecarbon.json ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "timestamp": "2024-10-24T16:26:02",
3
+ "project_name": "codecarbon",
4
+ "run_id": "2e624837-816c-4a54-b414-c29d24dd9f24",
5
+ "duration": -1729257494.3352268,
6
+ "emissions": 9.969546314489585e-06,
7
+ "emissions_rate": 1.1072346980473818e-05,
8
+ "cpu_power": 42.5,
9
+ "gpu_power": 65.0621770882058,
10
+ "ram_power": 0.35121631622314453,
11
+ "cpu_energy": 1.065322435606504e-05,
12
+ "gpu_energy": 1.6266957458377362e-05,
13
+ "ram_energy": 8.772346779188348e-08,
14
+ "energy_consumed": 2.7007905282234288e-05,
15
+ "country_name": "United States",
16
+ "country_iso_code": "USA",
17
+ "region": "virginia",
18
+ "cloud_provider": "",
19
+ "cloud_region": "",
20
+ "os": "Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35",
21
+ "python_version": "3.9.20",
22
+ "codecarbon_version": "2.5.1",
23
+ "cpu_count": 48,
24
+ "cpu_model": "AMD EPYC 7R32",
25
+ "gpu_count": 1,
26
+ "gpu_model": "1 x NVIDIA A10G",
27
+ "longitude": -77.4903,
28
+ "latitude": 39.0469,
29
+ "ram_total_size": 186.7047882080078,
30
+ "tracking_mode": "process",
31
+ "on_cloud": "N",
32
+ "pue": 1.0
33
+ }