Xabi Ezpeleta commited on
Commit
e6c370a
·
1 Parent(s): 74d65f7

Ignore wandb logs

Browse files
wandb/debug-internal.log DELETED
@@ -1,7 +0,0 @@
1
- {"time":"2025-02-17T21:46:18.764731866Z","level":"INFO","msg":"stream: starting","core version":"0.19.6","symlink path":"/home/tknika/xezpeleta/whisper/whisper-tiny-eu/wandb/run-20250217_214618-7bygcjmf/logs/debug-core.log"}
2
- {"time":"2025-02-17T21:46:18.87215251Z","level":"INFO","msg":"created new stream","id":"7bygcjmf"}
3
- {"time":"2025-02-17T21:46:18.8722139Z","level":"INFO","msg":"stream: started","id":"7bygcjmf"}
4
- {"time":"2025-02-17T21:46:18.872282469Z","level":"INFO","msg":"writer: Do: started","stream_id":"7bygcjmf"}
5
- {"time":"2025-02-17T21:46:18.872363578Z","level":"INFO","msg":"handler: started","stream_id":"7bygcjmf"}
6
- {"time":"2025-02-17T21:46:18.87291405Z","level":"INFO","msg":"sender: started","stream_id":"7bygcjmf"}
7
- {"time":"2025-02-17T21:46:19.182639444Z","level":"INFO","msg":"Starting system monitor"}
 
 
 
 
 
 
 
 
wandb/debug.log DELETED
@@ -1,25 +0,0 @@
1
- 2025-02-17 21:46:18,542 INFO MainThread:168016 [wandb_setup.py:_flush():68] Current SDK version is 0.19.6
2
- 2025-02-17 21:46:18,542 INFO MainThread:168016 [wandb_setup.py:_flush():68] Configure stats pid to 168016
3
- 2025-02-17 21:46:18,542 INFO MainThread:168016 [wandb_setup.py:_flush():68] Loading settings from /home/tknika/.config/wandb/settings
4
- 2025-02-17 21:46:18,542 INFO MainThread:168016 [wandb_setup.py:_flush():68] Loading settings from /home/tknika/xezpeleta/whisper/whisper-tiny-eu/wandb/settings
5
- 2025-02-17 21:46:18,542 INFO MainThread:168016 [wandb_setup.py:_flush():68] Loading settings from environment variables
6
- 2025-02-17 21:46:18,542 INFO MainThread:168016 [wandb_init.py:setup_run_log_directory():637] Logging user logs to /home/tknika/xezpeleta/whisper/whisper-tiny-eu/wandb/run-20250217_214618-7bygcjmf/logs/debug.log
7
- 2025-02-17 21:46:18,542 INFO MainThread:168016 [wandb_init.py:setup_run_log_directory():638] Logging internal logs to /home/tknika/xezpeleta/whisper/whisper-tiny-eu/wandb/run-20250217_214618-7bygcjmf/logs/debug-internal.log
8
- 2025-02-17 21:46:18,542 INFO MainThread:168016 [wandb_init.py:init():756] calling init triggers
9
- 2025-02-17 21:46:18,542 INFO MainThread:168016 [wandb_init.py:init():761] wandb.init called with sweep_config: {}
10
- config: {'_wandb': {}}
11
- 2025-02-17 21:46:18,542 INFO MainThread:168016 [wandb_init.py:init():789] starting backend
12
- 2025-02-17 21:46:18,752 INFO MainThread:168016 [wandb_init.py:init():793] sending inform_init request
13
- 2025-02-17 21:46:18,758 INFO MainThread:168016 [backend.py:_multiprocessing_setup():97] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
14
- 2025-02-17 21:46:18,759 INFO MainThread:168016 [wandb_init.py:init():808] backend started and connected
15
- 2025-02-17 21:46:18,761 INFO MainThread:168016 [wandb_init.py:init():901] updated telemetry
16
- 2025-02-17 21:46:18,768 INFO MainThread:168016 [wandb_init.py:init():936] communicating run to backend with 90.0 second timeout
17
- 2025-02-17 21:46:19,179 INFO MainThread:168016 [wandb_init.py:init():994] starting run threads in backend
18
- 2025-02-17 21:46:19,298 INFO MainThread:168016 [wandb_run.py:_console_start():2385] atexit reg
19
- 2025-02-17 21:46:19,298 INFO MainThread:168016 [wandb_run.py:_redirect():2235] redirect: wrap_raw
20
- 2025-02-17 21:46:19,298 INFO MainThread:168016 [wandb_run.py:_redirect():2300] Wrapping output streams.
21
- 2025-02-17 21:46:19,298 INFO MainThread:168016 [wandb_run.py:_redirect():2325] Redirects installed.
22
- 2025-02-17 21:46:19,300 INFO MainThread:168016 [wandb_init.py:init():1036] run started, returning control to user process
23
- 2025-02-17 21:46:19,301 INFO MainThread:168016 [wandb_run.py:_config_callback():1253] config_cb None None {'vocab_size': 51865, 'num_mel_bins': 80, 'd_model': 384, 'encoder_layers': 4, 'encoder_attention_heads': 6, 'decoder_layers': 4, 'decoder_attention_heads': 6, 'decoder_ffn_dim': 1536, 'encoder_ffn_dim': 1536, 'dropout': 0.0, 'attention_dropout': 0.0, 'activation_dropout': 0.0, 'activation_function': 'gelu', 'init_std': 0.02, 'encoder_layerdrop': 0.0, 'decoder_layerdrop': 0.0, 'use_cache': False, 'num_hidden_layers': 4, 'scale_embedding': False, 'max_source_positions': 1500, 'max_target_positions': 448, 'classifier_proj_size': 256, 'use_weighted_layer_sum': False, 'apply_spec_augment': False, 'mask_time_prob': 0.05, 'mask_time_length': 10, 'mask_time_min_masks': 2, 'mask_feature_prob': 0.0, 'mask_feature_length': 10, 'mask_feature_min_masks': 0, 'median_filter_width': 7, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'float32', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': True, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 448, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': [220, 50257], 'architectures': ['WhisperForConditionalGeneration'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 50257, 'pad_token_id': 50257, 'eos_token_id': 50257, 'sep_token_id': None, 'decoder_start_token_id': 50258, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'openai/whisper-tiny', '_attn_implementation_autoset': True, 'transformers_version': '4.49.0.dev0', 'forced_decoder_ids': None, 'model_type': 'whisper', 'output_dir': './', 'overwrite_output_dir': True, 'do_train': True, 'do_eval': True, 'do_predict': False, 'eval_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 32, 'per_device_eval_batch_size': 16, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'torch_empty_cache_steps': None, 'learning_rate': 3.75e-05, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 3.0, 'max_steps': 10000, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 1000, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': './runs/Feb17_21-45-10_tknika', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 25, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 1000, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'restore_callback_states_from_checkpoint': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': 1000, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': 'whisper-tiny-eu-2025.02', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': True, 'metric_for_best_model': 'wer', 'greater_is_better': False, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'input_length', 'report_to': ['wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': None, 'hub_always_push': False, 'gradient_checkpointing': True, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'include_for_metrics': [], 'eval_do_concat_batches': True, 'fp16_backend': 'auto', 'evaluation_strategy': 'steps', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None, 'batch_eval_metrics': False, 'eval_on_start': False, 'use_liger_kernel': False, 'eval_use_gather_object': False, 'average_tokens_across_devices': False, 'sortish_sampler': False, 'predict_with_generate': True, 'generation_max_length': 225, 'generation_num_beams': None, 'generation_config': None}
24
- 2025-02-17 21:46:19,303 INFO MainThread:168016 [wandb_config.py:__setitem__():154] config set model/num_parameters = 37760640 - <bound method Run._config_callback of <wandb.sdk.wandb_run.Run object at 0x735fac35e9c0>>
25
- 2025-02-17 21:46:19,303 INFO MainThread:168016 [wandb_run.py:_config_callback():1253] config_cb model/num_parameters 37760640 None
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
wandb/run-20250217_214618-7bygcjmf/files/output.log DELETED
The diff for this file is too large to render. See raw diff
 
wandb/run-20250217_214618-7bygcjmf/files/requirements.txt DELETED
@@ -1,122 +0,0 @@
1
- aiosignal==1.3.2
2
- Markdown==3.7
3
- more-itertools==10.6.0
4
- requests==2.32.3
5
- sentry-sdk==2.21.0
6
- torchaudio==2.6.0
7
- charset-normalizer==3.4.1
8
- docker-pycreds==0.4.0
9
- nvidia-cusolver-cu12==11.6.1.9
10
- PyYAML==6.0.2
11
- flatbuffers==25.2.10
12
- librosa==0.10.2.post1
13
- soxr==0.5.0.post1
14
- multiprocess==0.70.16
15
- setuptools==75.8.0
16
- nvidia-cufft-cu12==11.2.1.3
17
- joblib==1.4.2
18
- pytz==2025.1
19
- pip==24.0
20
- humanfriendly==10.0
21
- scikit-learn==1.6.1
22
- ctranslate2==4.5.0
23
- certifi==2025.1.31
24
- jiwer==3.1.0
25
- regex==2024.11.6
26
- annotated-types==0.7.0
27
- grpcio==1.70.0
28
- msgpack==1.1.0
29
- mpmath==1.3.0
30
- nvidia-cudnn-cu12==9.1.0.70
31
- soundfile==0.13.1
32
- dill==0.3.8
33
- nvidia-nvtx-cu12==12.4.127
34
- six==1.17.0
35
- nvidia-cuda-cupti-cu12==12.4.127
36
- pyarrow==19.0.0
37
- nvidia-nccl-cu12==2.21.5
38
- decorator==5.1.1
39
- llvmlite==0.44.0
40
- frozenlist==1.5.0
41
- pydantic==2.10.6
42
- networkx==3.4.2
43
- idna==3.10
44
- wandb==0.19.6
45
- aiohttp==3.11.12
46
- RapidFuzz==3.12.1
47
- pandas==2.2.3
48
- python-dateutil==2.9.0.post0
49
- numpy==2.1.3
50
- coloredlogs==15.0.1
51
- tokenizers==0.21.0
52
- nvidia-cusparselt-cu12==0.6.2
53
- typing_extensions==4.12.2
54
- urllib3==2.3.0
55
- setproctitle==1.3.4
56
- onnxruntime==1.20.1
57
- tzdata==2025.1
58
- datasets==3.3.1.dev0
59
- sympy==1.13.1
60
- faster-whisper==1.1.1
61
- pooch==1.8.2
62
- click==8.1.8
63
- pydantic_core==2.27.2
64
- MarkupSafe==3.0.2
65
- scipy==1.15.1
66
- accelerate==1.3.0
67
- tensorboard==2.19.0
68
- protobuf==5.29.3
69
- gitdb==4.0.12
70
- smmap==5.0.2
71
- absl-py==2.1.0
72
- tqdm==4.67.1
73
- yarl==1.18.3
74
- pycparser==2.22
75
- nvidia-cusparse-cu12==12.3.1.170
76
- attrs==25.1.0
77
- lazy_loader==0.4
78
- tensorboard-data-server==0.7.2
79
- threadpoolctl==3.5.0
80
- GitPython==3.1.44
81
- safetensors==0.5.2
82
- fsspec==2024.12.0
83
- nvidia-cuda-nvrtc-cu12==12.4.127
84
- filelock==3.17.0
85
- aiohappyeyeballs==2.4.6
86
- packaging==24.2
87
- audioread==3.0.1
88
- propcache==0.2.1
89
- transformers==4.49.0.dev0
90
- nvidia-cuda-runtime-cu12==12.4.127
91
- cffi==1.17.1
92
- evaluate==0.4.3
93
- Werkzeug==3.1.3
94
- huggingface-hub==0.28.1
95
- Jinja2==3.1.5
96
- torch==2.6.0
97
- psutil==7.0.0
98
- nvidia-curand-cu12==10.3.5.147
99
- xxhash==3.5.0
100
- platformdirs==4.3.6
101
- multidict==6.1.0
102
- nvidia-cublas-cu12==12.4.5.8
103
- av==14.1.0
104
- nvidia-nvjitlink-cu12==12.4.127
105
- triton==3.2.0
106
- numba==0.61.0
107
- importlib_metadata==8.0.0
108
- platformdirs==4.2.2
109
- typeguard==4.3.0
110
- more-itertools==10.3.0
111
- tomli==2.0.1
112
- autocommand==2.2.2
113
- zipp==3.19.2
114
- typing_extensions==4.12.2
115
- backports.tarfile==1.2.0
116
- inflect==7.3.1
117
- jaraco.text==3.12.1
118
- wheel==0.43.0
119
- packaging==24.2
120
- jaraco.collections==5.1.0
121
- jaraco.functools==4.0.1
122
- jaraco.context==5.3.0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
wandb/run-20250217_214618-7bygcjmf/files/wandb-metadata.json DELETED
@@ -1,86 +0,0 @@
1
- {
2
- "os": "Linux-6.8.0-48-generic-x86_64-with-glibc2.39",
3
- "python": "CPython 3.12.3",
4
- "startedAt": "2025-02-17T21:46:18.759581Z",
5
- "args": [
6
- "--model_name_or_path=openai/whisper-tiny",
7
- "--dataset_name=asierhv/composite_corpus_eu_v2.1",
8
- "--language=basque",
9
- "--train_split_name=train",
10
- "--eval_split_name=dev_parl+test_parl+test_cv+test_oslr",
11
- "--model_index_name=Whisper Tiny Basque",
12
- "--max_steps=10000",
13
- "--output_dir=./",
14
- "--per_device_train_batch_size=32",
15
- "--per_device_eval_batch_size=16",
16
- "--gradient_accumulation_steps=1",
17
- "--logging_steps=25",
18
- "--learning_rate=3.75e-5",
19
- "--warmup_steps=1000",
20
- "--evaluation_strategy=steps",
21
- "--eval_steps=1000",
22
- "--save_strategy=steps",
23
- "--save_steps=1000",
24
- "--generation_max_length=225",
25
- "--length_column_name=input_length",
26
- "--max_duration_in_seconds=30",
27
- "--audio_column_name=audio",
28
- "--text_column_name=sentence",
29
- "--freeze_feature_encoder=False",
30
- "--report_to=tensorboard",
31
- "--metric_for_best_model=wer",
32
- "--greater_is_better=False",
33
- "--load_best_model_at_end",
34
- "--gradient_checkpointing",
35
- "--fp16",
36
- "--overwrite_output_dir",
37
- "--do_train",
38
- "--do_eval",
39
- "--predict_with_generate",
40
- "--do_normalize_eval",
41
- "--streaming",
42
- "--use_auth_token",
43
- "--push_to_hub",
44
- "--report_to",
45
- "wandb",
46
- "--run_name",
47
- "whisper-tiny-eu-2025.02"
48
- ],
49
- "program": "/home/tknika/xezpeleta/whisper/whisper-tiny-eu/run_speech_recognition_seq2seq_streaming.py",
50
- "codePath": "run_speech_recognition_seq2seq_streaming.py",
51
- "git": {
52
- "remote": "https://huggingface.co/xezpeleta/whisper-tiny-eu",
53
- "commit": "424998f63b9d278b4ce5a8fa14342bd43c6971ad"
54
- },
55
- "email": "xezpeleta@gmail.com",
56
- "root": "/home/tknika/xezpeleta/whisper/whisper-tiny-eu",
57
- "host": "tknika",
58
- "executable": "/home/tknika/xezpeleta/whisper/.venv/bin/python",
59
- "codePathLocal": "run_speech_recognition_seq2seq_streaming.py",
60
- "cpu_count": 8,
61
- "cpu_count_logical": 8,
62
- "gpu": "NVIDIA L40-48Q",
63
- "gpu_count": 1,
64
- "disk": {
65
- "/": {
66
- "total": "525987168256",
67
- "used": "447815217152"
68
- }
69
- },
70
- "memory": {
71
- "total": "33654022144"
72
- },
73
- "cpu": {
74
- "count": 8,
75
- "countLogical": 8
76
- },
77
- "gpu_nvidia": [
78
- {
79
- "name": "NVIDIA L40-48Q",
80
- "memoryTotal": "51539607552",
81
- "cudaCores": 18176,
82
- "architecture": "Ada"
83
- }
84
- ],
85
- "cudaVersion": "12.4"
86
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
wandb/run-20250217_214618-7bygcjmf/logs/debug-core.log DELETED
@@ -1,6 +0,0 @@
1
- {"time":"2025-02-17T21:46:18.667916236Z","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpzsw7ot30/port-168016.txt","pid":168016,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false}
2
- {"time":"2025-02-17T21:46:18.685302193Z","level":"INFO","msg":"Will exit if parent process dies.","ppid":168016}
3
- {"time":"2025-02-17T21:46:18.685729677Z","level":"INFO","msg":"server is running","addr":{"IP":"127.0.0.1","Port":40697,"Zone":""}}
4
- {"time":"2025-02-17T21:46:18.752659702Z","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"127.0.0.1:58468"}
5
- {"time":"2025-02-17T21:46:18.764274372Z","level":"INFO","msg":"handleInformInit: received","streamId":"7bygcjmf","id":"127.0.0.1:58468"}
6
- {"time":"2025-02-17T21:46:18.87222342Z","level":"INFO","msg":"handleInformInit: stream started","streamId":"7bygcjmf","id":"127.0.0.1:58468"}
 
 
 
 
 
 
 
wandb/run-20250217_214618-7bygcjmf/logs/debug-internal.log DELETED
@@ -1,7 +0,0 @@
1
- {"time":"2025-02-17T21:46:18.764731866Z","level":"INFO","msg":"stream: starting","core version":"0.19.6","symlink path":"/home/tknika/xezpeleta/whisper/whisper-tiny-eu/wandb/run-20250217_214618-7bygcjmf/logs/debug-core.log"}
2
- {"time":"2025-02-17T21:46:18.87215251Z","level":"INFO","msg":"created new stream","id":"7bygcjmf"}
3
- {"time":"2025-02-17T21:46:18.8722139Z","level":"INFO","msg":"stream: started","id":"7bygcjmf"}
4
- {"time":"2025-02-17T21:46:18.872282469Z","level":"INFO","msg":"writer: Do: started","stream_id":"7bygcjmf"}
5
- {"time":"2025-02-17T21:46:18.872363578Z","level":"INFO","msg":"handler: started","stream_id":"7bygcjmf"}
6
- {"time":"2025-02-17T21:46:18.87291405Z","level":"INFO","msg":"sender: started","stream_id":"7bygcjmf"}
7
- {"time":"2025-02-17T21:46:19.182639444Z","level":"INFO","msg":"Starting system monitor"}
 
 
 
 
 
 
 
 
wandb/run-20250217_214618-7bygcjmf/logs/debug.log DELETED
@@ -1,25 +0,0 @@
1
- 2025-02-17 21:46:18,542 INFO MainThread:168016 [wandb_setup.py:_flush():68] Current SDK version is 0.19.6
2
- 2025-02-17 21:46:18,542 INFO MainThread:168016 [wandb_setup.py:_flush():68] Configure stats pid to 168016
3
- 2025-02-17 21:46:18,542 INFO MainThread:168016 [wandb_setup.py:_flush():68] Loading settings from /home/tknika/.config/wandb/settings
4
- 2025-02-17 21:46:18,542 INFO MainThread:168016 [wandb_setup.py:_flush():68] Loading settings from /home/tknika/xezpeleta/whisper/whisper-tiny-eu/wandb/settings
5
- 2025-02-17 21:46:18,542 INFO MainThread:168016 [wandb_setup.py:_flush():68] Loading settings from environment variables
6
- 2025-02-17 21:46:18,542 INFO MainThread:168016 [wandb_init.py:setup_run_log_directory():637] Logging user logs to /home/tknika/xezpeleta/whisper/whisper-tiny-eu/wandb/run-20250217_214618-7bygcjmf/logs/debug.log
7
- 2025-02-17 21:46:18,542 INFO MainThread:168016 [wandb_init.py:setup_run_log_directory():638] Logging internal logs to /home/tknika/xezpeleta/whisper/whisper-tiny-eu/wandb/run-20250217_214618-7bygcjmf/logs/debug-internal.log
8
- 2025-02-17 21:46:18,542 INFO MainThread:168016 [wandb_init.py:init():756] calling init triggers
9
- 2025-02-17 21:46:18,542 INFO MainThread:168016 [wandb_init.py:init():761] wandb.init called with sweep_config: {}
10
- config: {'_wandb': {}}
11
- 2025-02-17 21:46:18,542 INFO MainThread:168016 [wandb_init.py:init():789] starting backend
12
- 2025-02-17 21:46:18,752 INFO MainThread:168016 [wandb_init.py:init():793] sending inform_init request
13
- 2025-02-17 21:46:18,758 INFO MainThread:168016 [backend.py:_multiprocessing_setup():97] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
14
- 2025-02-17 21:46:18,759 INFO MainThread:168016 [wandb_init.py:init():808] backend started and connected
15
- 2025-02-17 21:46:18,761 INFO MainThread:168016 [wandb_init.py:init():901] updated telemetry
16
- 2025-02-17 21:46:18,768 INFO MainThread:168016 [wandb_init.py:init():936] communicating run to backend with 90.0 second timeout
17
- 2025-02-17 21:46:19,179 INFO MainThread:168016 [wandb_init.py:init():994] starting run threads in backend
18
- 2025-02-17 21:46:19,298 INFO MainThread:168016 [wandb_run.py:_console_start():2385] atexit reg
19
- 2025-02-17 21:46:19,298 INFO MainThread:168016 [wandb_run.py:_redirect():2235] redirect: wrap_raw
20
- 2025-02-17 21:46:19,298 INFO MainThread:168016 [wandb_run.py:_redirect():2300] Wrapping output streams.
21
- 2025-02-17 21:46:19,298 INFO MainThread:168016 [wandb_run.py:_redirect():2325] Redirects installed.
22
- 2025-02-17 21:46:19,300 INFO MainThread:168016 [wandb_init.py:init():1036] run started, returning control to user process
23
- 2025-02-17 21:46:19,301 INFO MainThread:168016 [wandb_run.py:_config_callback():1253] config_cb None None {'vocab_size': 51865, 'num_mel_bins': 80, 'd_model': 384, 'encoder_layers': 4, 'encoder_attention_heads': 6, 'decoder_layers': 4, 'decoder_attention_heads': 6, 'decoder_ffn_dim': 1536, 'encoder_ffn_dim': 1536, 'dropout': 0.0, 'attention_dropout': 0.0, 'activation_dropout': 0.0, 'activation_function': 'gelu', 'init_std': 0.02, 'encoder_layerdrop': 0.0, 'decoder_layerdrop': 0.0, 'use_cache': False, 'num_hidden_layers': 4, 'scale_embedding': False, 'max_source_positions': 1500, 'max_target_positions': 448, 'classifier_proj_size': 256, 'use_weighted_layer_sum': False, 'apply_spec_augment': False, 'mask_time_prob': 0.05, 'mask_time_length': 10, 'mask_time_min_masks': 2, 'mask_feature_prob': 0.0, 'mask_feature_length': 10, 'mask_feature_min_masks': 0, 'median_filter_width': 7, 'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'float32', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': True, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 448, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': [220, 50257], 'architectures': ['WhisperForConditionalGeneration'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 50257, 'pad_token_id': 50257, 'eos_token_id': 50257, 'sep_token_id': None, 'decoder_start_token_id': 50258, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'openai/whisper-tiny', '_attn_implementation_autoset': True, 'transformers_version': '4.49.0.dev0', 'forced_decoder_ids': None, 'model_type': 'whisper', 'output_dir': './', 'overwrite_output_dir': True, 'do_train': True, 'do_eval': True, 'do_predict': False, 'eval_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 32, 'per_device_eval_batch_size': 16, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'torch_empty_cache_steps': None, 'learning_rate': 3.75e-05, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 3.0, 'max_steps': 10000, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 1000, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': './runs/Feb17_21-45-10_tknika', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 25, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 1000, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'restore_callback_states_from_checkpoint': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': 1000, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': 'whisper-tiny-eu-2025.02', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': True, 'metric_for_best_model': 'wer', 'greater_is_better': False, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'input_length', 'report_to': ['wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': None, 'hub_always_push': False, 'gradient_checkpointing': True, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'include_for_metrics': [], 'eval_do_concat_batches': True, 'fp16_backend': 'auto', 'evaluation_strategy': 'steps', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None, 'batch_eval_metrics': False, 'eval_on_start': False, 'use_liger_kernel': False, 'eval_use_gather_object': False, 'average_tokens_across_devices': False, 'sortish_sampler': False, 'predict_with_generate': True, 'generation_max_length': 225, 'generation_num_beams': None, 'generation_config': None}
24
- 2025-02-17 21:46:19,303 INFO MainThread:168016 [wandb_config.py:__setitem__():154] config set model/num_parameters = 37760640 - <bound method Run._config_callback of <wandb.sdk.wandb_run.Run object at 0x735fac35e9c0>>
25
- 2025-02-17 21:46:19,303 INFO MainThread:168016 [wandb_run.py:_config_callback():1253] config_cb model/num_parameters 37760640 None
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
wandb/run-20250217_214618-7bygcjmf/run-7bygcjmf.wandb DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:7b24350895068d2eeaab4073bdf0d0a76ef0d5ec72269f30f06e8afe5fbb6315
3
- size 4980736