diff --git a/automatic_speech_recognition/jonatasgrosman/wav2vec2-large-xlsr-53-portuguese/2024-12-07-06-11-42/.hydra/config.yaml b/automatic_speech_recognition/jonatasgrosman/wav2vec2-large-xlsr-53-portuguese/2024-12-07-06-11-42/.hydra/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..8e17e4a375896a1f9d549130d21e5d8e5081dd7e --- /dev/null +++ b/automatic_speech_recognition/jonatasgrosman/wav2vec2-large-xlsr-53-portuguese/2024-12-07-06-11-42/.hydra/config.yaml @@ -0,0 +1,94 @@ +backend: + name: pytorch + version: 2.4.0 + _target_: optimum_benchmark.backends.pytorch.backend.PyTorchBackend + task: automatic-speech-recognition + model: jonatasgrosman/wav2vec2-large-xlsr-53-portuguese + processor: jonatasgrosman/wav2vec2-large-xlsr-53-portuguese + library: null + device: cuda + device_ids: '0' + seed: 42 + inter_op_num_threads: null + intra_op_num_threads: null + hub_kwargs: {} + no_weights: true + device_map: null + torch_dtype: null + amp_autocast: false + amp_dtype: null + eval_mode: true + to_bettertransformer: false + low_cpu_mem_usage: null + attn_implementation: null + cache_implementation: null + torch_compile: false + torch_compile_config: {} + quantization_scheme: null + quantization_config: {} + deepspeed_inference: false + deepspeed_inference_config: {} + peft_type: null + peft_config: {} +launcher: + name: process + _target_: optimum_benchmark.launchers.process.launcher.ProcessLauncher + device_isolation: true + device_isolation_action: warn + start_method: spawn +benchmark: + name: energy_star + _target_: optimum_benchmark.benchmarks.energy_star.benchmark.EnergyStarBenchmark + dataset_name: EnergyStarAI/ASR + dataset_config: '' + dataset_split: train + num_samples: 1000 + input_shapes: + batch_size: 1 + text_column_name: text + truncation: true + max_length: -1 + dataset_prefix1: '' + dataset_prefix2: '' + t5_task: '' + image_column_name: image + resize: false + question_column_name: question + context_column_name: context + sentence1_column_name: sentence1 + sentence2_column_name: sentence2 + audio_column_name: audio + iterations: 10 + warmup_runs: 10 + energy: true + forward_kwargs: {} + generate_kwargs: {} + call_kwargs: {} +experiment_name: automatic_speech_recognition +environment: + cpu: ' AMD EPYC 7R32' + cpu_count: 48 + cpu_ram_mb: 200472.73984 + system: Linux + machine: x86_64 + platform: Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35 + processor: x86_64 + python_version: 3.9.20 + gpu: + - NVIDIA A10G + gpu_count: 1 + gpu_vram_mb: 24146608128 + optimum_benchmark_version: 0.2.0 + optimum_benchmark_commit: null + transformers_version: 4.44.0 + transformers_commit: null + accelerate_version: 0.33.0 + accelerate_commit: null + diffusers_version: 0.30.0 + diffusers_commit: null + optimum_version: null + optimum_commit: null + timm_version: null + timm_commit: null + peft_version: null + peft_commit: null diff --git a/automatic_speech_recognition/jonatasgrosman/wav2vec2-large-xlsr-53-portuguese/2024-12-07-06-11-42/.hydra/hydra.yaml b/automatic_speech_recognition/jonatasgrosman/wav2vec2-large-xlsr-53-portuguese/2024-12-07-06-11-42/.hydra/hydra.yaml new file mode 100644 index 0000000000000000000000000000000000000000..4e28591df82030f570fbf60f919837514e215ecf --- /dev/null +++ b/automatic_speech_recognition/jonatasgrosman/wav2vec2-large-xlsr-53-portuguese/2024-12-07-06-11-42/.hydra/hydra.yaml @@ -0,0 +1,175 @@ +hydra: + run: + dir: /runs/automatic_speech_recognition/jonatasgrosman/wav2vec2-large-xlsr-53-portuguese/2024-12-07-06-11-42 + sweep: + dir: runs/${experiment_name}/${backend.model}/${now:%Y-%m-%d-%H-%M-%S} + subdir: ${hydra.job.num} + launcher: + _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher + sweeper: + _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper + max_batch_size: null + params: null + help: + app_name: ${hydra.job.name} + header: '${hydra.help.app_name} is powered by Hydra. + + ' + footer: 'Powered by Hydra (https://hydra.cc) + + Use --hydra-help to view Hydra specific help + + ' + template: '${hydra.help.header} + + == Configuration groups == + + Compose your configuration from those groups (group=option) + + + $APP_CONFIG_GROUPS + + + == Config == + + Override anything in the config (foo.bar=value) + + + $CONFIG + + + ${hydra.help.footer} + + ' + hydra_help: + template: 'Hydra (${hydra.runtime.version}) + + See https://hydra.cc for more info. + + + == Flags == + + $FLAGS_HELP + + + == Configuration groups == + + Compose your configuration from those groups (For example, append hydra/job_logging=disabled + to command line) + + + $HYDRA_CONFIG_GROUPS + + + Use ''--cfg hydra'' to Show the Hydra config. + + ' + hydra_help: ??? + hydra_logging: + version: 1 + formatters: + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(purple)sHYDRA%(reset)s] %(message)s' + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + root: + level: INFO + handlers: + - console + disable_existing_loggers: false + job_logging: + version: 1 + formatters: + simple: + format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s' + colorlog: + (): colorlog.ColoredFormatter + format: '[%(cyan)s%(asctime)s%(reset)s][%(blue)s%(name)s%(reset)s][%(log_color)s%(levelname)s%(reset)s] + - %(message)s' + log_colors: + DEBUG: purple + INFO: green + WARNING: yellow + ERROR: red + CRITICAL: red + handlers: + console: + class: logging.StreamHandler + formatter: colorlog + stream: ext://sys.stdout + file: + class: logging.FileHandler + formatter: simple + filename: ${hydra.job.name}.log + root: + level: INFO + handlers: + - console + - file + disable_existing_loggers: false + env: {} + mode: RUN + searchpath: [] + callbacks: {} + output_subdir: .hydra + overrides: + hydra: + - hydra.run.dir=/runs/automatic_speech_recognition/jonatasgrosman/wav2vec2-large-xlsr-53-portuguese/2024-12-07-06-11-42 + - hydra.mode=RUN + task: + - backend.model=jonatasgrosman/wav2vec2-large-xlsr-53-portuguese + - backend.processor=jonatasgrosman/wav2vec2-large-xlsr-53-portuguese + job: + name: cli + chdir: true + override_dirname: backend.model=jonatasgrosman/wav2vec2-large-xlsr-53-portuguese,backend.processor=jonatasgrosman/wav2vec2-large-xlsr-53-portuguese + id: ??? + num: ??? + config_name: automatic_speech_recognition + env_set: + OVERRIDE_BENCHMARKS: '1' + env_copy: [] + config: + override_dirname: + kv_sep: '=' + item_sep: ',' + exclude_keys: [] + runtime: + version: 1.3.2 + version_base: '1.3' + cwd: / + config_sources: + - path: hydra.conf + schema: pkg + provider: hydra + - path: optimum_benchmark + schema: pkg + provider: main + - path: hydra_plugins.hydra_colorlog.conf + schema: pkg + provider: hydra-colorlog + - path: /optimum-benchmark/examples/energy_star + schema: file + provider: command-line + - path: '' + schema: structured + provider: schema + output_dir: /runs/automatic_speech_recognition/jonatasgrosman/wav2vec2-large-xlsr-53-portuguese/2024-12-07-06-11-42 + choices: + benchmark: energy_star + launcher: process + backend: pytorch + hydra/env: default + hydra/callbacks: null + hydra/job_logging: colorlog + hydra/hydra_logging: colorlog + hydra/hydra_help: default + hydra/help: default + hydra/sweeper: basic + hydra/launcher: basic + hydra/output: default + verbose: false diff --git a/automatic_speech_recognition/jonatasgrosman/wav2vec2-large-xlsr-53-portuguese/2024-12-07-06-11-42/.hydra/overrides.yaml b/automatic_speech_recognition/jonatasgrosman/wav2vec2-large-xlsr-53-portuguese/2024-12-07-06-11-42/.hydra/overrides.yaml new file mode 100644 index 0000000000000000000000000000000000000000..6a6838f6859f0bb76f2138056ac984bd878ec4d4 --- /dev/null +++ b/automatic_speech_recognition/jonatasgrosman/wav2vec2-large-xlsr-53-portuguese/2024-12-07-06-11-42/.hydra/overrides.yaml @@ -0,0 +1,2 @@ +- backend.model=jonatasgrosman/wav2vec2-large-xlsr-53-portuguese +- backend.processor=jonatasgrosman/wav2vec2-large-xlsr-53-portuguese diff --git a/automatic_speech_recognition/jonatasgrosman/wav2vec2-large-xlsr-53-portuguese/2024-12-07-06-11-42/cli.log b/automatic_speech_recognition/jonatasgrosman/wav2vec2-large-xlsr-53-portuguese/2024-12-07-06-11-42/cli.log new file mode 100644 index 0000000000000000000000000000000000000000..84b2da95be6adbd3e74aae013047efd0486159ac --- /dev/null +++ b/automatic_speech_recognition/jonatasgrosman/wav2vec2-large-xlsr-53-portuguese/2024-12-07-06-11-42/cli.log @@ -0,0 +1,25 @@ +[2024-12-07 06:11:45,176][launcher][INFO] - ََAllocating process launcher +[2024-12-07 06:11:45,176][process][INFO] - + Setting multiprocessing start method to spawn. +[2024-12-07 06:11:45,188][device-isolation][INFO] - + Launched device(s) isolation process 1684 +[2024-12-07 06:11:45,189][device-isolation][INFO] - + Isolating device(s) [0] +[2024-12-07 06:11:45,195][process][INFO] - + Launched benchmark in isolated process 1685. +[PROC-0][2024-12-07 06:11:47,759][datasets][INFO] - PyTorch version 2.4.0 available. +[PROC-0][2024-12-07 06:11:48,710][backend][INFO] - َAllocating pytorch backend +[PROC-0][2024-12-07 06:11:48,710][backend][INFO] - + Setting random seed to 42 +[PROC-0][2024-12-07 06:11:49,096][pytorch][INFO] - + Using AutoModel class AutoModelForCTC +[PROC-0][2024-12-07 06:11:49,096][pytorch][INFO] - + Creating backend temporary directory +[PROC-0][2024-12-07 06:11:49,096][pytorch][INFO] - + Loading model with random weights +[PROC-0][2024-12-07 06:11:49,096][pytorch][INFO] - + Creating no weights model +[PROC-0][2024-12-07 06:11:49,096][pytorch][INFO] - + Creating no weights model directory +[PROC-0][2024-12-07 06:11:49,096][pytorch][INFO] - + Creating no weights model state dict +[PROC-0][2024-12-07 06:11:49,098][pytorch][INFO] - + Saving no weights model safetensors +[PROC-0][2024-12-07 06:11:49,099][pytorch][INFO] - + Saving no weights model pretrained config +[PROC-0][2024-12-07 06:11:49,100][pytorch][INFO] - + Loading no weights AutoModel +[PROC-0][2024-12-07 06:11:49,100][pytorch][INFO] - + Loading model directly on device: cuda +[PROC-0][2024-12-07 06:11:49,419][pytorch][INFO] - + Turning on model's eval mode +[PROC-0][2024-12-07 06:11:49,425][benchmark][INFO] - Allocating energy_star benchmark +[PROC-0][2024-12-07 06:11:49,425][energy_star][INFO] - + Loading raw dataset +[PROC-0][2024-12-07 06:11:55,577][energy_star][INFO] - + Initializing Inference report +[PROC-0][2024-12-07 06:11:55,577][energy][INFO] - + Tracking GPU energy on devices [0] +[PROC-0][2024-12-07 06:11:59,770][energy_star][INFO] - + Preprocessing dataset +[2024-12-07 06:12:06,206][experiment][ERROR] - Error during experiment diff --git a/automatic_speech_recognition/jonatasgrosman/wav2vec2-large-xlsr-53-portuguese/2024-12-07-06-11-42/error.log b/automatic_speech_recognition/jonatasgrosman/wav2vec2-large-xlsr-53-portuguese/2024-12-07-06-11-42/error.log new file mode 100644 index 0000000000000000000000000000000000000000..99ab0cda9f908e12445697b7536114b4dab98fb8 --- /dev/null +++ b/automatic_speech_recognition/jonatasgrosman/wav2vec2-large-xlsr-53-portuguese/2024-12-07-06-11-42/error.log @@ -0,0 +1,95 @@ +/opt/conda/lib/python3.9/site-packages/transformers/tokenization_utils_base.py:1601: FutureWarning: `clean_up_tokenization_spaces` was not set. It will be set to `True` by default. This behavior will be depracted in transformers v4.45, and will be then set to `False` by default. For more details check this issue: https://github.com/huggingface/transformers/issues/31884 + warnings.warn( + Downloading readme: 0%| | 0.00/320 [00:00>> Tracker's metadata: +[codecarbon INFO @ 06:11:56] Platform system: Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35 +[codecarbon INFO @ 06:11:56] Python version: 3.9.20 +[codecarbon INFO @ 06:11:56] CodeCarbon version: 2.5.1 +[codecarbon INFO @ 06:11:56] Available RAM : 186.705 GB +[codecarbon INFO @ 06:11:56] CPU count: 48 +[codecarbon INFO @ 06:11:56] CPU model: AMD EPYC 7R32 +[codecarbon INFO @ 06:11:56] GPU count: 1 +[codecarbon INFO @ 06:11:56] GPU model: 1 x NVIDIA A10G +[codecarbon DEBUG @ 06:11:57] Not running on AWS +[codecarbon DEBUG @ 06:11:58] Not running on Azure +[codecarbon DEBUG @ 06:11:59] Not running on GCP +[codecarbon INFO @ 06:11:59] Saving emissions data to file /runs/automatic_speech_recognition/jonatasgrosman/wav2vec2-large-xlsr-53-portuguese/2024-12-07-06-11-42/codecarbon.csv +[codecarbon DEBUG @ 06:11:59] EmissionsData(timestamp='2024-12-07T06:11:59', project_name='codecarbon', run_id='a03b476e-f4b5-493a-b12b-aca3f913c3d2', duration=0.0021435300004668534, emissions=0.0, emissions_rate=0.0, cpu_power=0.0, gpu_power=0.0, ram_power=0.0, cpu_energy=0, gpu_energy=0, ram_energy=0, energy_consumed=0, country_name='United States', country_iso_code='USA', region='virginia', cloud_provider='', cloud_region='', os='Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35', python_version='3.9.20', codecarbon_version='2.5.1', cpu_count=48, cpu_model='AMD EPYC 7R32', gpu_count=1, gpu_model='1 x NVIDIA A10G', longitude=-77.4903, latitude=39.0469, ram_total_size=186.7047882080078, tracking_mode='process', on_cloud='N', pue=1.0) + Filter: 0%| | 0/1000 [00:00>> Tracker's metadata: +[codecarbon INFO @ 06:12:15] Platform system: Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35 +[codecarbon INFO @ 06:12:15] Python version: 3.9.20 +[codecarbon INFO @ 06:12:15] CodeCarbon version: 2.5.1 +[codecarbon INFO @ 06:12:15] Available RAM : 186.705 GB +[codecarbon INFO @ 06:12:15] CPU count: 48 +[codecarbon INFO @ 06:12:15] CPU model: AMD EPYC 7R32 +[codecarbon INFO @ 06:12:15] GPU count: 1 +[codecarbon INFO @ 06:12:15] GPU model: 1 x NVIDIA A10G +[codecarbon DEBUG @ 06:12:16] Not running on AWS +[codecarbon DEBUG @ 06:12:17] Not running on Azure +[codecarbon DEBUG @ 06:12:18] Not running on GCP +[codecarbon INFO @ 06:12:18] Saving emissions data to file /runs/image_classification/Falconsai/nsfw_image_detection/2024-12-07-06-12-06/codecarbon.csv +[codecarbon DEBUG @ 06:12:18] EmissionsData(timestamp='2024-12-07T06:12:18', project_name='codecarbon', run_id='1227b6b4-bf23-48a4-ac0a-e6b6f69567eb', duration=0.0022014300047885627, emissions=0.0, emissions_rate=0.0, cpu_power=0.0, gpu_power=0.0, ram_power=0.0, cpu_energy=0, gpu_energy=0, ram_energy=0, energy_consumed=0, country_name='United States', country_iso_code='USA', region='virginia', cloud_provider='', cloud_region='', os='Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35', python_version='3.9.20', codecarbon_version='2.5.1', cpu_count=48, cpu_model='AMD EPYC 7R32', gpu_count=1, gpu_model='1 x NVIDIA A10G', longitude=-77.4903, latitude=39.0469, ram_total_size=186.7047882080078, tracking_mode='process', on_cloud='N', pue=1.0) + Running processor on dataset: 0%| | 0/1000 [00:00>> Tracker's metadata: +[codecarbon INFO @ 06:05:14] Platform system: Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35 +[codecarbon INFO @ 06:05:14] Python version: 3.9.20 +[codecarbon INFO @ 06:05:14] CodeCarbon version: 2.5.1 +[codecarbon INFO @ 06:05:14] Available RAM : 186.705 GB +[codecarbon INFO @ 06:05:14] CPU count: 48 +[codecarbon INFO @ 06:05:14] CPU model: AMD EPYC 7R32 +[codecarbon INFO @ 06:05:14] GPU count: 1 +[codecarbon INFO @ 06:05:14] GPU model: 1 x NVIDIA A10G +[codecarbon DEBUG @ 06:05:15] Not running on AWS +[codecarbon DEBUG @ 06:05:16] Not running on Azure +[codecarbon DEBUG @ 06:05:17] Not running on GCP +[codecarbon INFO @ 06:05:17] Saving emissions data to file /runs/image_classification/WinKawaks/vit-tiny-patch16-224/2024-12-07-06-05-03/codecarbon.csv +[codecarbon DEBUG @ 06:05:17] EmissionsData(timestamp='2024-12-07T06:05:17', project_name='codecarbon', run_id='b311f965-506c-43d6-9dbb-d13cb8595e0d', duration=0.0021764500124845654, emissions=0.0, emissions_rate=0.0, cpu_power=0.0, gpu_power=0.0, ram_power=0.0, cpu_energy=0, gpu_energy=0, ram_energy=0, energy_consumed=0, country_name='United States', country_iso_code='USA', region='virginia', cloud_provider='', cloud_region='', os='Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35', python_version='3.9.20', codecarbon_version='2.5.1', cpu_count=48, cpu_model='AMD EPYC 7R32', gpu_count=1, gpu_model='1 x NVIDIA A10G', longitude=-77.4903, latitude=39.0469, ram_total_size=186.7047882080078, tracking_mode='process', on_cloud='N', pue=1.0) + Filter: 0%| | 0/1000 [00:00>> Tracker's metadata: +[codecarbon INFO @ 06:16:51] Platform system: Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35 +[codecarbon INFO @ 06:16:51] Python version: 3.9.20 +[codecarbon INFO @ 06:16:51] CodeCarbon version: 2.5.1 +[codecarbon INFO @ 06:16:51] Available RAM : 186.705 GB +[codecarbon INFO @ 06:16:51] CPU count: 48 +[codecarbon INFO @ 06:16:51] CPU model: AMD EPYC 7R32 +[codecarbon INFO @ 06:16:51] GPU count: 1 +[codecarbon INFO @ 06:16:51] GPU model: 1 x NVIDIA A10G +[codecarbon DEBUG @ 06:16:52] Not running on AWS +[codecarbon DEBUG @ 06:16:53] Not running on Azure +[codecarbon DEBUG @ 06:16:54] Not running on GCP +[codecarbon INFO @ 06:16:54] Saving emissions data to file /runs/image_classification/google/vit-base-patch16-224/2024-12-07-06-16-42/codecarbon.csv +[codecarbon DEBUG @ 06:16:54] EmissionsData(timestamp='2024-12-07T06:16:54', project_name='codecarbon', run_id='46d0333d-2e22-48c6-971d-c3bac025b79e', duration=0.0022072010033298284, emissions=0.0, emissions_rate=0.0, cpu_power=0.0, gpu_power=0.0, ram_power=0.0, cpu_energy=0, gpu_energy=0, ram_energy=0, energy_consumed=0, country_name='United States', country_iso_code='USA', region='virginia', cloud_provider='', cloud_region='', os='Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35', python_version='3.9.20', codecarbon_version='2.5.1', cpu_count=48, cpu_model='AMD EPYC 7R32', gpu_count=1, gpu_model='1 x NVIDIA A10G', longitude=-77.4903, latitude=39.0469, ram_total_size=186.7047882080078, tracking_mode='process', on_cloud='N', pue=1.0) +[codecarbon INFO @ 06:16:54] Energy consumed for RAM : 0.000000 kWh. RAM Power : 0.2560844421386719 W +[codecarbon DEBUG @ 06:16:54] RAM : 0.26 W during 0.01 s [measurement time: 0.0004] +[codecarbon INFO @ 06:16:54] Energy consumed for all GPUs : 0.000000 kWh. Total GPU Power : 0.0 W +[codecarbon DEBUG @ 06:16:54] GPU : 0.00 W during 0.01 s [measurement time: 0.0021] +[codecarbon INFO @ 06:16:54] Energy consumed for all CPUs : 0.000000 kWh. Total CPU Power : 42.5 W +[codecarbon DEBUG @ 06:16:54] CPU : 42.50 W during 0.01 s [measurement time: 0.0000] +[codecarbon INFO @ 06:16:54] 0.000000 kWh of electricity used since the beginning. +[codecarbon DEBUG @ 06:16:54] last_duration=0.009131788014201447 +------------------------ +[codecarbon DEBUG @ 06:16:54] EmissionsData(timestamp='2024-12-07T06:16:54', project_name='codecarbon', run_id='46d0333d-2e22-48c6-971d-c3bac025b79e', duration=0.012067432020558044, emissions=5.2276560675210465e-08, emissions_rate=4.332036889551336e-06, cpu_power=42.5, gpu_power=0.0, ram_power=0.2560844421386719, cpu_energy=1.4096954865736837e-07, gpu_energy=0.0, ram_energy=6.497742946753871e-10, energy_consumed=1.4161932295204375e-07, country_name='United States', country_iso_code='USA', region='virginia', cloud_provider='', cloud_region='', os='Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35', python_version='3.9.20', codecarbon_version='2.5.1', cpu_count=48, cpu_model='AMD EPYC 7R32', gpu_count=1, gpu_model='1 x NVIDIA A10G', longitude=-77.4903, latitude=39.0469, ram_total_size=186.7047882080078, tracking_mode='process', on_cloud='N', pue=1.0) +[codecarbon DEBUG @ 06:16:54] EmissionsData(timestamp='2024-12-07T06:16:54', project_name='codecarbon', run_id='46d0333d-2e22-48c6-971d-c3bac025b79e', duration=0.0022103410155978054, emissions=5.2276560675210465e-08, emissions_rate=2.3650902872592186e-05, cpu_power=42.5, gpu_power=0.0, ram_power=0.2560844421386719, cpu_energy=1.4096954865736837e-07, gpu_energy=0.0, ram_energy=6.497742946753871e-10, energy_consumed=1.4161932295204375e-07, country_name='United States', country_iso_code='USA', region='virginia', cloud_provider='', cloud_region='', os='Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35', python_version='3.9.20', codecarbon_version='2.5.1', cpu_count=48, cpu_model='AMD EPYC 7R32', gpu_count=1, gpu_model='1 x NVIDIA A10G', longitude=-77.4903, latitude=39.0469, ram_total_size=186.7047882080078, tracking_mode='process', on_cloud='N', pue=1.0) + 0%| | 0/1000 [00:00>> Tracker's metadata: +[codecarbon INFO @ 06:10:09] Platform system: Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35 +[codecarbon INFO @ 06:10:09] Python version: 3.9.20 +[codecarbon INFO @ 06:10:09] CodeCarbon version: 2.5.1 +[codecarbon INFO @ 06:10:09] Available RAM : 186.705 GB +[codecarbon INFO @ 06:10:09] CPU count: 48 +[codecarbon INFO @ 06:10:09] CPU model: AMD EPYC 7R32 +[codecarbon INFO @ 06:10:09] GPU count: 1 +[codecarbon INFO @ 06:10:09] GPU model: 1 x NVIDIA A10G +[codecarbon DEBUG @ 06:10:10] Not running on AWS +[codecarbon DEBUG @ 06:10:11] Not running on Azure +[codecarbon DEBUG @ 06:10:12] Not running on GCP +[codecarbon INFO @ 06:10:12] Saving emissions data to file /runs/sentence_similarity/sentence-transformers/bert-base-nli-mean-tokens/2024-12-07-06-09-58/codecarbon.csv +[codecarbon DEBUG @ 06:10:12] EmissionsData(timestamp='2024-12-07T06:10:12', project_name='codecarbon', run_id='e865c20a-47a1-42bb-8b80-58e4f2b5419a', duration=0.002482334995875135, emissions=0.0, emissions_rate=0.0, cpu_power=0.0, gpu_power=0.0, ram_power=0.0, cpu_energy=0, gpu_energy=0, ram_energy=0, energy_consumed=0, country_name='United States', country_iso_code='USA', region='virginia', cloud_provider='', cloud_region='', os='Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35', python_version='3.9.20', codecarbon_version='2.5.1', cpu_count=48, cpu_model='AMD EPYC 7R32', gpu_count=1, gpu_model='1 x NVIDIA A10G', longitude=-77.4903, latitude=39.0469, ram_total_size=186.7047882080078, tracking_mode='process', on_cloud='N', pue=1.0) + Filter: 0%| | 0/1000 [00:00>> Tracker's metadata: +[codecarbon INFO @ 05:59:24] Platform system: Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35 +[codecarbon INFO @ 05:59:24] Python version: 3.9.20 +[codecarbon INFO @ 05:59:24] CodeCarbon version: 2.5.1 +[codecarbon INFO @ 05:59:24] Available RAM : 186.705 GB +[codecarbon INFO @ 05:59:24] CPU count: 48 +[codecarbon INFO @ 05:59:24] CPU model: AMD EPYC 7R32 +[codecarbon INFO @ 05:59:24] GPU count: 1 +[codecarbon INFO @ 05:59:24] GPU model: 1 x NVIDIA A10G +[codecarbon DEBUG @ 05:59:25] Not running on AWS +[codecarbon DEBUG @ 05:59:26] Not running on Azure +[codecarbon DEBUG @ 05:59:27] Not running on GCP +[codecarbon INFO @ 05:59:27] Saving emissions data to file /runs/summarization/facebook/bart-large-cnn/2024-12-07-05-59-14/codecarbon.csv +[codecarbon DEBUG @ 05:59:27] EmissionsData(timestamp='2024-12-07T05:59:27', project_name='codecarbon', run_id='d9969f35-5239-46e5-99c5-b3063a70b123', duration=0.002186931000323966, emissions=0.0, emissions_rate=0.0, cpu_power=0.0, gpu_power=0.0, ram_power=0.0, cpu_energy=0, gpu_energy=0, ram_energy=0, energy_consumed=0, country_name='United States', country_iso_code='USA', region='virginia', cloud_provider='', cloud_region='', os='Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35', python_version='3.9.20', codecarbon_version='2.5.1', cpu_count=48, cpu_model='AMD EPYC 7R32', gpu_count=1, gpu_model='1 x NVIDIA A10G', longitude=-77.4903, latitude=39.0469, ram_total_size=186.7047882080078, tracking_mode='process', on_cloud='N', pue=1.0) + Filter: 0%| | 0/1000 [00:00>> Tracker's metadata: +[codecarbon INFO @ 06:04:58] Platform system: Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35 +[codecarbon INFO @ 06:04:58] Python version: 3.9.20 +[codecarbon INFO @ 06:04:58] CodeCarbon version: 2.5.1 +[codecarbon INFO @ 06:04:58] Available RAM : 186.705 GB +[codecarbon INFO @ 06:04:58] CPU count: 48 +[codecarbon INFO @ 06:04:58] CPU model: AMD EPYC 7R32 +[codecarbon INFO @ 06:04:58] GPU count: 1 +[codecarbon INFO @ 06:04:58] GPU model: 1 x NVIDIA A10G +[codecarbon DEBUG @ 06:04:59] Not running on AWS +[codecarbon DEBUG @ 06:05:00] Not running on Azure +[codecarbon DEBUG @ 06:05:01] Not running on GCP +[codecarbon INFO @ 06:05:01] Saving emissions data to file /runs/text_classification/cardiffnlp/twitter-roberta-base-sentiment-latest/2024-12-07-06-04-49/codecarbon.csv +[codecarbon DEBUG @ 06:05:01] EmissionsData(timestamp='2024-12-07T06:05:01', project_name='codecarbon', run_id='6deca56d-ef84-4ddc-841c-e2c87160ce3e', duration=0.002153368986910209, emissions=0.0, emissions_rate=0.0, cpu_power=0.0, gpu_power=0.0, ram_power=0.0, cpu_energy=0, gpu_energy=0, ram_energy=0, energy_consumed=0, country_name='United States', country_iso_code='USA', region='virginia', cloud_provider='', cloud_region='', os='Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35', python_version='3.9.20', codecarbon_version='2.5.1', cpu_count=48, cpu_model='AMD EPYC 7R32', gpu_count=1, gpu_model='1 x NVIDIA A10G', longitude=-77.4903, latitude=39.0469, ram_total_size=186.7047882080078, tracking_mode='process', on_cloud='N', pue=1.0) + Filter: 0%| | 0/1000 [00:00>> Tracker's metadata: +[codecarbon INFO @ 06:11:19] Platform system: Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35 +[codecarbon INFO @ 06:11:19] Python version: 3.9.20 +[codecarbon INFO @ 06:11:19] CodeCarbon version: 2.5.1 +[codecarbon INFO @ 06:11:19] Available RAM : 186.705 GB +[codecarbon INFO @ 06:11:19] CPU count: 48 +[codecarbon INFO @ 06:11:19] CPU model: AMD EPYC 7R32 +[codecarbon INFO @ 06:11:19] GPU count: 1 +[codecarbon INFO @ 06:11:19] GPU model: 1 x NVIDIA A10G +[codecarbon DEBUG @ 06:11:20] Not running on AWS +[codecarbon DEBUG @ 06:11:21] Not running on Azure +[codecarbon DEBUG @ 06:11:22] Not running on GCP +[codecarbon INFO @ 06:11:22] Saving emissions data to file /runs/text_classification/papluca/xlm-roberta-base-language-detection/2024-12-07-06-11-09/codecarbon.csv +[codecarbon DEBUG @ 06:11:22] EmissionsData(timestamp='2024-12-07T06:11:22', project_name='codecarbon', run_id='923157f8-64f5-41ff-85ca-223e77dee9d3', duration=0.002174720022594556, emissions=0.0, emissions_rate=0.0, cpu_power=0.0, gpu_power=0.0, ram_power=0.0, cpu_energy=0, gpu_energy=0, ram_energy=0, energy_consumed=0, country_name='United States', country_iso_code='USA', region='virginia', cloud_provider='', cloud_region='', os='Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35', python_version='3.9.20', codecarbon_version='2.5.1', cpu_count=48, cpu_model='AMD EPYC 7R32', gpu_count=1, gpu_model='1 x NVIDIA A10G', longitude=-77.4903, latitude=39.0469, ram_total_size=186.7047882080078, tracking_mode='process', on_cloud='N', pue=1.0) + Running tokenizer on dataset: 0%| | 0/1000 [00:00>> Tracker's metadata: +[codecarbon INFO @ 06:11:35] Platform system: Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35 +[codecarbon INFO @ 06:11:35] Python version: 3.9.20 +[codecarbon INFO @ 06:11:35] CodeCarbon version: 2.5.1 +[codecarbon INFO @ 06:11:35] Available RAM : 186.705 GB +[codecarbon INFO @ 06:11:35] CPU count: 48 +[codecarbon INFO @ 06:11:35] CPU model: AMD EPYC 7R32 +[codecarbon INFO @ 06:11:35] GPU count: 1 +[codecarbon INFO @ 06:11:35] GPU model: 1 x NVIDIA A10G +[codecarbon DEBUG @ 06:11:36] Not running on AWS +[codecarbon DEBUG @ 06:11:37] Not running on Azure +[codecarbon DEBUG @ 06:11:38] Not running on GCP +[codecarbon INFO @ 06:11:38] Saving emissions data to file /runs/text_generation/distributed/optimized-gpt2-1b/2024-12-07-06-11-25/codecarbon.csv +[codecarbon DEBUG @ 06:11:38] EmissionsData(timestamp='2024-12-07T06:11:38', project_name='codecarbon', run_id='b9b03429-845c-4691-90c3-af17425fd4b3', duration=0.0022394309926312417, emissions=0.0, emissions_rate=0.0, cpu_power=0.0, gpu_power=0.0, ram_power=0.0, cpu_energy=0, gpu_energy=0, ram_energy=0, energy_consumed=0, country_name='United States', country_iso_code='USA', region='virginia', cloud_provider='', cloud_region='', os='Linux-5.10.192-183.736.amzn2.x86_64-x86_64-with-glibc2.35', python_version='3.9.20', codecarbon_version='2.5.1', cpu_count=48, cpu_model='AMD EPYC 7R32', gpu_count=1, gpu_model='1 x NVIDIA A10G', longitude=-77.4903, latitude=39.0469, ram_total_size=186.7047882080078, tracking_mode='process', on_cloud='N', pue=1.0) + Filter: 0%| | 0/1000 [00:00