SiMajid commited on
Commit
31e0a53
·
verified ·
1 Parent(s): 8c9aba0

reward-train-roberta

Browse files
README.md CHANGED
@@ -1,20 +1,22 @@
1
  ---
2
- license: other
3
  library_name: peft
4
  tags:
 
 
5
  - generated_from_trainer
6
- base_model: facebook/opt-350m
7
  model-index:
8
- - name: facebook/opt-350m
9
  results: []
10
  ---
11
 
12
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
13
  should probably proofread and complete it, then remove this comment. -->
14
 
15
- # facebook/opt-350m
16
 
17
- This model is a fine-tuned version of [facebook/opt-350m](https://huggingface.co/facebook/opt-350m) on the None dataset.
18
 
19
  ## Model description
20
 
@@ -37,8 +39,8 @@ The following hyperparameters were used during training:
37
  - train_batch_size: 1
38
  - eval_batch_size: 8
39
  - seed: 42
40
- - gradient_accumulation_steps: 8
41
- - total_train_batch_size: 8
42
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
43
  - lr_scheduler_type: linear
44
  - num_epochs: 10
@@ -50,7 +52,7 @@ The following hyperparameters were used during training:
50
  ### Framework versions
51
 
52
  - PEFT 0.11.1
53
- - Transformers 4.36.2
54
- - Pytorch 2.0.0
55
- - Datasets 2.1.0
56
- - Tokenizers 0.15.0
 
1
  ---
2
+ license: mit
3
  library_name: peft
4
  tags:
5
+ - trl
6
+ - reward-trainer
7
  - generated_from_trainer
8
+ base_model: microsoft/deberta-v3-large
9
  model-index:
10
+ - name: microsoft/deberta-v3-large
11
  results: []
12
  ---
13
 
14
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
15
  should probably proofread and complete it, then remove this comment. -->
16
 
17
+ # microsoft/deberta-v3-large
18
 
19
+ This model is a fine-tuned version of [microsoft/deberta-v3-large](https://huggingface.co/microsoft/deberta-v3-large) on the None dataset.
20
 
21
  ## Model description
22
 
 
39
  - train_batch_size: 1
40
  - eval_batch_size: 8
41
  - seed: 42
42
+ - gradient_accumulation_steps: 4
43
+ - total_train_batch_size: 4
44
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
  - lr_scheduler_type: linear
46
  - num_epochs: 10
 
52
  ### Framework versions
53
 
54
  - PEFT 0.11.1
55
+ - Transformers 4.41.2
56
+ - Pytorch 2.1.2
57
+ - Datasets 2.19.2
58
+ - Tokenizers 0.19.1
adapter_config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
- "base_model_name_or_path": "facebook/opt-350m",
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
@@ -24,15 +24,8 @@
24
  "rank_pattern": {},
25
  "revision": null,
26
  "target_modules": [
27
- "up_proj",
28
- "o_proj",
29
- "q_proj",
30
- "gate_proj",
31
- "lm_head",
32
- "down_proj",
33
- "k_proj",
34
- "v_proj",
35
- "embed_tokens"
36
  ],
37
  "task_type": "SEQ_CLS",
38
  "use_dora": false,
 
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
+ "base_model_name_or_path": "microsoft/deberta-v3-large",
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
 
24
  "rank_pattern": {},
25
  "revision": null,
26
  "target_modules": [
27
+ "value_proj",
28
+ "query_proj"
 
 
 
 
 
 
 
29
  ],
30
  "task_type": "SEQ_CLS",
31
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7ee5aa395d6465b2c9c20577890fb3c3b7a23b9db5de9285bce7b9d5452ce229
3
- size 115669872
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:259280966da7f2340dfea6ca3e8e3b3627c2c98d0cd6c87bbe3b4e305cacad33
3
+ size 6314296
added_tokens.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "[MASK]": 128000
3
+ }
runs/Jun13_09-46-29_3d0b606030ff/events.out.tfevents.1718271990.3d0b606030ff.35.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:07c7e4d831740e4f79f28180ac7ad1e5ac53b6004975624b352c67034447099f
3
+ size 7644
special_tokens_map.json CHANGED
@@ -1,27 +1,12 @@
1
  {
2
- "bos_token": {
3
- "content": "</s>",
4
- "lstrip": false,
5
- "normalized": true,
6
- "rstrip": false,
7
- "single_word": false
8
- },
9
- "eos_token": {
10
- "content": "</s>",
11
- "lstrip": false,
12
- "normalized": true,
13
- "rstrip": false,
14
- "single_word": false
15
- },
16
- "pad_token": {
17
- "content": "<pad>",
18
- "lstrip": false,
19
- "normalized": true,
20
- "rstrip": false,
21
- "single_word": false
22
- },
23
  "unk_token": {
24
- "content": "</s>",
25
  "lstrip": false,
26
  "normalized": true,
27
  "rstrip": false,
 
1
  {
2
+ "bos_token": "[CLS]",
3
+ "cls_token": "[CLS]",
4
+ "eos_token": "[SEP]",
5
+ "mask_token": "[MASK]",
6
+ "pad_token": "[PAD]",
7
+ "sep_token": "[SEP]",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
8
  "unk_token": {
9
+ "content": "[UNK]",
10
  "lstrip": false,
11
  "normalized": true,
12
  "rstrip": false,
spm.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c679fbf93643d19aab7ee10c0b99e460bdbc02fedf34b92b05af343b4af586fd
3
+ size 2464616
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -1,30 +1,58 @@
1
  {
2
- "add_bos_token": true,
3
- "add_prefix_space": false,
4
  "added_tokens_decoder": {
 
 
 
 
 
 
 
 
5
  "1": {
6
- "content": "<pad>",
7
  "lstrip": false,
8
- "normalized": true,
9
  "rstrip": false,
10
  "single_word": false,
11
  "special": true
12
  },
13
  "2": {
14
- "content": "</s>",
 
 
 
 
 
 
 
 
15
  "lstrip": false,
16
  "normalized": true,
17
  "rstrip": false,
18
  "single_word": false,
19
  "special": true
 
 
 
 
 
 
 
 
20
  }
21
  },
22
- "bos_token": "</s>",
23
  "clean_up_tokenization_spaces": true,
24
- "eos_token": "</s>",
25
- "errors": "replace",
 
 
26
  "model_max_length": 1000000000000000019884624838656,
27
- "pad_token": "<pad>",
28
- "tokenizer_class": "GPT2Tokenizer",
29
- "unk_token": "</s>"
 
 
 
 
30
  }
 
1
  {
 
 
2
  "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
  "1": {
12
+ "content": "[CLS]",
13
  "lstrip": false,
14
+ "normalized": false,
15
  "rstrip": false,
16
  "single_word": false,
17
  "special": true
18
  },
19
  "2": {
20
+ "content": "[SEP]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "3": {
28
+ "content": "[UNK]",
29
  "lstrip": false,
30
  "normalized": true,
31
  "rstrip": false,
32
  "single_word": false,
33
  "special": true
34
+ },
35
+ "128000": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
  }
43
  },
44
+ "bos_token": "[CLS]",
45
  "clean_up_tokenization_spaces": true,
46
+ "cls_token": "[CLS]",
47
+ "do_lower_case": false,
48
+ "eos_token": "[SEP]",
49
+ "mask_token": "[MASK]",
50
  "model_max_length": 1000000000000000019884624838656,
51
+ "pad_token": "[PAD]",
52
+ "sep_token": "[SEP]",
53
+ "sp_model_kwargs": {},
54
+ "split_by_punct": false,
55
+ "tokenizer_class": "DebertaV2Tokenizer",
56
+ "unk_token": "[UNK]",
57
+ "vocab_type": "spm"
58
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d9bd88d676a238f9bf792a68394423f8524508ef5f6b92f30c7bedc5a183feaf
3
- size 4283
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f3768f27ab8f9c71bbb3d3e828123612dbdb4a7da88a9368ccff5a8fafec7829
3
+ size 5112
wandb/debug-internal.log CHANGED
The diff for this file is too large to render. See raw diff
 
wandb/debug.log CHANGED
@@ -1,52 +1,38 @@
1
- 2024-06-12 15:06:10,605 INFO MainThread:26 [wandb_setup.py:_flush():76] Current SDK version is 0.16.2
2
- 2024-06-12 15:06:10,606 INFO MainThread:26 [wandb_setup.py:_flush():76] Configure stats pid to 26
3
- 2024-06-12 15:06:10,606 INFO MainThread:26 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
4
- 2024-06-12 15:06:10,606 INFO MainThread:26 [wandb_setup.py:_flush():76] Loading settings from /kaggle/working/wandb/settings
5
- 2024-06-12 15:06:10,606 INFO MainThread:26 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
6
- 2024-06-12 15:06:10,606 INFO MainThread:26 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
7
- 2024-06-12 15:06:10,606 INFO MainThread:26 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program': '<python with no main file>'}
8
- 2024-06-12 15:06:10,606 INFO MainThread:26 [wandb_setup.py:_flush():76] Applying login settings: {'api_key': '***REDACTED***'}
9
- 2024-06-12 15:06:10,606 INFO MainThread:26 [wandb_init.py:_log_setup():526] Logging user logs to /kaggle/working/wandb/run-20240612_150610-7vg9fga2/logs/debug.log
10
- 2024-06-12 15:06:10,606 INFO MainThread:26 [wandb_init.py:_log_setup():527] Logging internal logs to /kaggle/working/wandb/run-20240612_150610-7vg9fga2/logs/debug-internal.log
11
- 2024-06-12 15:06:10,606 INFO MainThread:26 [wandb_init.py:_jupyter_setup():472] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x7a17c1cfc820>
12
- 2024-06-12 15:06:10,607 INFO MainThread:26 [wandb_init.py:init():566] calling init triggers
13
- 2024-06-12 15:06:10,607 INFO MainThread:26 [wandb_init.py:init():573] wandb.init called with sweep_config: {}
 
14
  config: {}
15
- 2024-06-12 15:06:10,607 INFO MainThread:26 [wandb_init.py:init():616] starting backend
16
- 2024-06-12 15:06:10,607 INFO MainThread:26 [wandb_init.py:init():620] setting up manager
17
- 2024-06-12 15:06:10,609 INFO MainThread:26 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
18
- 2024-06-12 15:06:10,611 INFO MainThread:26 [wandb_init.py:init():628] backend started and connected
19
- 2024-06-12 15:06:10,623 INFO MainThread:26 [wandb_run.py:_label_probe_notebook():1294] probe notebook
20
- 2024-06-12 15:06:11,290 INFO MainThread:26 [wandb_init.py:init():720] updated telemetry
21
- 2024-06-12 15:06:11,293 INFO MainThread:26 [wandb_init.py:init():753] communicating run to backend with 90.0 second timeout
22
- 2024-06-12 15:06:11,523 INFO MainThread:26 [wandb_run.py:_on_init():2254] communicating current version
23
- 2024-06-12 15:06:11,583 INFO MainThread:26 [wandb_run.py:_on_init():2263] got version response upgrade_message: "wandb version 0.17.1 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
24
 
25
- 2024-06-12 15:06:11,583 INFO MainThread:26 [wandb_init.py:init():804] starting run threads in backend
26
- 2024-06-12 15:06:41,620 INFO MainThread:26 [wandb_run.py:_console_start():2233] atexit reg
27
- 2024-06-12 15:06:41,620 INFO MainThread:26 [wandb_run.py:_redirect():2088] redirect: wrap_raw
28
- 2024-06-12 15:06:41,622 INFO MainThread:26 [wandb_run.py:_redirect():2153] Wrapping output streams.
29
- 2024-06-12 15:06:41,622 INFO MainThread:26 [wandb_run.py:_redirect():2178] Redirects installed.
30
- 2024-06-12 15:06:41,623 INFO MainThread:26 [wandb_init.py:init():847] run started, returning control to user process
31
- 2024-06-12 15:06:41,629 INFO MainThread:26 [wandb_run.py:_config_callback():1342] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': 'float16', 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 1024, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'chunk_size_feed_forward': 0, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['OPTForCausalLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': '</s>', 'bos_token_id': 2, 'pad_token_id': 1, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'facebook/opt-350m', 'transformers_version': '4.36.2', 'activation_dropout': 0.0, 'model_type': 'opt', 'vocab_size': 50272, 'max_position_embeddings': 2048, 'num_attention_heads': 16, 'word_embed_proj_dim': 512, 'ffn_dim': 4096, 'hidden_size': 1024, 'num_hidden_layers': 24, 'dropout': 0.1, 'attention_dropout': 0.0, 'activation_function': 'relu', 'init_std': 0.02, 'layerdrop': 0.0, 'use_cache': True, 'do_layer_norm_before': False, 'enable_bias': True, 'layer_norm_elementwise_affine': True, '_remove_final_layer_norm': False, 'output_dir': './', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': False, 'do_predict': False, 'evaluation_strategy': 'no', 'prediction_loss_only': False, 'per_device_train_batch_size': 1, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 8, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 1.41e-05, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 10, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': './runs/Jun12_15-05-56_5f4811c18302', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'past_index': -1, 'run_name': 'reward-train-facebook-opt350m', 'disable_tqdm': False, 'remove_unused_columns': False, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': True, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': False, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None}
32
- 2024-06-12 19:18:48,236 INFO MainThread:26 [jupyter.py:save_ipynb():373] not saving jupyter notebook
33
- 2024-06-12 19:18:48,237 INFO MainThread:26 [wandb_init.py:_pause_backend():437] pausing backend
34
- 2024-06-12 19:18:48,246 INFO MainThread:26 [wandb_init.py:_resume_backend():442] resuming backend
35
- 2024-06-12 19:18:55,407 INFO MainThread:26 [jupyter.py:save_ipynb():373] not saving jupyter notebook
36
- 2024-06-12 19:18:55,407 INFO MainThread:26 [wandb_init.py:_pause_backend():437] pausing backend
37
- 2024-06-12 19:18:55,413 INFO MainThread:26 [wandb_init.py:_resume_backend():442] resuming backend
38
- 2024-06-12 19:19:03,775 INFO MainThread:26 [jupyter.py:save_ipynb():373] not saving jupyter notebook
39
- 2024-06-12 19:19:03,776 INFO MainThread:26 [wandb_init.py:_pause_backend():437] pausing backend
40
- 2024-06-12 19:19:03,781 INFO MainThread:26 [wandb_init.py:_resume_backend():442] resuming backend
41
- 2024-06-12 19:19:07,055 INFO MainThread:26 [jupyter.py:save_ipynb():373] not saving jupyter notebook
42
- 2024-06-12 19:19:07,056 INFO MainThread:26 [wandb_init.py:_pause_backend():437] pausing backend
43
- 2024-06-12 19:19:07,063 INFO MainThread:26 [wandb_init.py:_resume_backend():442] resuming backend
44
- 2024-06-12 19:19:08,096 INFO MainThread:26 [jupyter.py:save_ipynb():373] not saving jupyter notebook
45
- 2024-06-12 19:19:08,096 INFO MainThread:26 [wandb_init.py:_pause_backend():437] pausing backend
46
- 2024-06-12 19:19:08,101 INFO MainThread:26 [wandb_init.py:_resume_backend():442] resuming backend
47
- 2024-06-12 19:19:10,950 INFO MainThread:26 [jupyter.py:save_ipynb():373] not saving jupyter notebook
48
- 2024-06-12 19:19:10,950 INFO MainThread:26 [wandb_init.py:_pause_backend():437] pausing backend
49
- 2024-06-12 19:19:32,105 INFO MainThread:26 [wandb_init.py:_resume_backend():442] resuming backend
50
- 2024-06-12 19:19:33,147 INFO MainThread:26 [jupyter.py:save_ipynb():373] not saving jupyter notebook
51
- 2024-06-12 19:19:33,147 INFO MainThread:26 [wandb_init.py:_pause_backend():437] pausing backend
52
- 2024-06-12 19:19:33,155 INFO MainThread:26 [wandb_init.py:_resume_backend():442] resuming backend
 
1
+ 2024-06-13 09:46:45,545 INFO MainThread:35 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0
2
+ 2024-06-13 09:46:45,545 INFO MainThread:35 [wandb_setup.py:_flush():76] Configure stats pid to 35
3
+ 2024-06-13 09:46:45,545 INFO MainThread:35 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
4
+ 2024-06-13 09:46:45,545 INFO MainThread:35 [wandb_setup.py:_flush():76] Loading settings from /kaggle/working/wandb/settings
5
+ 2024-06-13 09:46:45,545 INFO MainThread:35 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
6
+ 2024-06-13 09:46:45,545 INFO MainThread:35 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
7
+ 2024-06-13 09:46:45,545 INFO MainThread:35 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program': '<python with no main file>'}
8
+ 2024-06-13 09:46:45,545 INFO MainThread:35 [wandb_setup.py:_flush():76] Applying login settings: {}
9
+ 2024-06-13 09:46:45,545 INFO MainThread:35 [wandb_setup.py:_flush():76] Applying login settings: {'api_key': '***REDACTED***'}
10
+ 2024-06-13 09:46:45,545 INFO MainThread:35 [wandb_init.py:_log_setup():520] Logging user logs to /kaggle/working/wandb/run-20240613_094645-i03sjxtv/logs/debug.log
11
+ 2024-06-13 09:46:45,545 INFO MainThread:35 [wandb_init.py:_log_setup():521] Logging internal logs to /kaggle/working/wandb/run-20240613_094645-i03sjxtv/logs/debug-internal.log
12
+ 2024-06-13 09:46:45,545 INFO MainThread:35 [wandb_init.py:_jupyter_setup():466] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x7877e40d8160>
13
+ 2024-06-13 09:46:45,546 INFO MainThread:35 [wandb_init.py:init():560] calling init triggers
14
+ 2024-06-13 09:46:45,546 INFO MainThread:35 [wandb_init.py:init():567] wandb.init called with sweep_config: {}
15
  config: {}
16
+ 2024-06-13 09:46:45,546 INFO MainThread:35 [wandb_init.py:init():610] starting backend
17
+ 2024-06-13 09:46:45,546 INFO MainThread:35 [wandb_init.py:init():614] setting up manager
18
+ 2024-06-13 09:46:45,548 INFO MainThread:35 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
19
+ 2024-06-13 09:46:45,550 INFO MainThread:35 [wandb_init.py:init():622] backend started and connected
20
+ 2024-06-13 09:46:45,564 INFO MainThread:35 [wandb_run.py:_label_probe_notebook():1328] probe notebook
21
+ 2024-06-13 09:46:45,978 INFO MainThread:35 [wandb_init.py:init():711] updated telemetry
22
+ 2024-06-13 09:46:45,981 INFO MainThread:35 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout
23
+ 2024-06-13 09:46:46,175 INFO MainThread:35 [wandb_run.py:_on_init():2396] communicating current version
24
+ 2024-06-13 09:46:46,232 INFO MainThread:35 [wandb_run.py:_on_init():2405] got version response upgrade_message: "wandb version 0.17.1 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
25
 
26
+ 2024-06-13 09:46:46,232 INFO MainThread:35 [wandb_init.py:init():795] starting run threads in backend
27
+ 2024-06-13 09:47:02,296 INFO MainThread:35 [wandb_run.py:_console_start():2374] atexit reg
28
+ 2024-06-13 09:47:02,296 INFO MainThread:35 [wandb_run.py:_redirect():2229] redirect: wrap_raw
29
+ 2024-06-13 09:47:02,297 INFO MainThread:35 [wandb_run.py:_redirect():2294] Wrapping output streams.
30
+ 2024-06-13 09:47:02,297 INFO MainThread:35 [wandb_run.py:_redirect():2319] Redirects installed.
31
+ 2024-06-13 09:47:02,304 INFO MainThread:35 [wandb_init.py:init():838] run started, returning control to user process
32
+ 2024-06-13 09:47:02,312 INFO MainThread:35 [wandb_run.py:_config_callback():1376] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 1024, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': None, 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': 0, 'eos_token_id': None, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'microsoft/deberta-v3-large', 'transformers_version': '4.41.2', 'model_type': 'deberta-v2', 'position_buckets': 256, 'norm_rel_ebd': 'layer_norm', 'share_att_key': True, 'hidden_size': 1024, 'num_hidden_layers': 24, 'num_attention_heads': 16, 'intermediate_size': 4096, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'attention_probs_dropout_prob': 0.1, 'max_position_embeddings': 512, 'type_vocab_size': 0, 'initializer_range': 0.02, 'relative_attention': True, 'max_relative_positions': -1, 'position_biased_input': False, 'pos_att_type': ['p2c', 'c2p'], 'vocab_size': 128100, 'layer_norm_eps': 1e-07, 'pooler_hidden_size': 1024, 'pooler_dropout': 0, 'pooler_hidden_act': 'gelu', 'output_dir': './', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': False, 'do_predict': False, 'eval_strategy': 'no', 'prediction_loss_only': False, 'per_device_train_batch_size': 1, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 4, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 1.41e-05, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 10, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': './runs/Jun13_09-46-29_3d0b606030ff', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'restore_callback_states_from_checkpoint': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': 'reward-train-facebook-opt350m', 'disable_tqdm': False, 'remove_unused_columns': False, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': True, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'eval_do_concat_batches': True, 'fp16_backend': 'auto', 'evaluation_strategy': 'no', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None, 'batch_eval_metrics': False}
33
+ 2024-06-13 12:04:34,011 INFO MainThread:35 [jupyter.py:save_ipynb():373] not saving jupyter notebook
34
+ 2024-06-13 12:04:34,011 INFO MainThread:35 [wandb_init.py:_pause_backend():431] pausing backend
35
+ 2024-06-13 12:25:21,829 INFO MainThread:35 [wandb_init.py:_resume_backend():436] resuming backend
36
+ 2024-06-13 12:25:24,437 INFO MainThread:35 [jupyter.py:save_ipynb():373] not saving jupyter notebook
37
+ 2024-06-13 12:25:24,437 INFO MainThread:35 [wandb_init.py:_pause_backend():431] pausing backend
38
+ 2024-06-13 12:25:24,444 INFO MainThread:35 [wandb_init.py:_resume_backend():436] resuming backend
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
wandb/run-20240613_094645-i03sjxtv/files/conda-environment.yaml ADDED
File without changes
wandb/run-20240613_094645-i03sjxtv/files/config.yaml ADDED
@@ -0,0 +1,706 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wandb_version: 1
2
+
3
+ _wandb:
4
+ desc: null
5
+ value:
6
+ python_version: 3.10.13
7
+ cli_version: 0.17.0
8
+ framework: huggingface
9
+ huggingface_version: 4.41.2
10
+ is_jupyter_run: true
11
+ is_kaggle_kernel: true
12
+ start_time: 1718272005
13
+ t:
14
+ 1:
15
+ - 1
16
+ - 2
17
+ - 3
18
+ - 5
19
+ - 11
20
+ - 12
21
+ - 49
22
+ - 51
23
+ - 53
24
+ - 55
25
+ - 71
26
+ - 84
27
+ - 98
28
+ - 105
29
+ 2:
30
+ - 1
31
+ - 2
32
+ - 3
33
+ - 5
34
+ - 11
35
+ - 12
36
+ - 49
37
+ - 51
38
+ - 53
39
+ - 55
40
+ - 71
41
+ - 84
42
+ - 98
43
+ - 105
44
+ 3:
45
+ - 7
46
+ - 13
47
+ - 23
48
+ - 62
49
+ - 66
50
+ 4: 3.10.13
51
+ 5: 0.17.0
52
+ 6: 4.41.2
53
+ 8:
54
+ - 1
55
+ - 2
56
+ - 5
57
+ 9:
58
+ 1: transformers_trainer
59
+ 13: linux-x86_64
60
+ m:
61
+ - 1: train/global_step
62
+ 6:
63
+ - 3
64
+ - 1: train/loss
65
+ 5: 1
66
+ 6:
67
+ - 1
68
+ - 1: train/grad_norm
69
+ 5: 1
70
+ 6:
71
+ - 1
72
+ - 1: train/learning_rate
73
+ 5: 1
74
+ 6:
75
+ - 1
76
+ - 1: train/epoch
77
+ 5: 1
78
+ 6:
79
+ - 1
80
+ return_dict:
81
+ desc: null
82
+ value: true
83
+ output_hidden_states:
84
+ desc: null
85
+ value: false
86
+ output_attentions:
87
+ desc: null
88
+ value: false
89
+ torchscript:
90
+ desc: null
91
+ value: false
92
+ torch_dtype:
93
+ desc: null
94
+ value: null
95
+ use_bfloat16:
96
+ desc: null
97
+ value: false
98
+ tf_legacy_loss:
99
+ desc: null
100
+ value: false
101
+ pruned_heads:
102
+ desc: null
103
+ value: {}
104
+ tie_word_embeddings:
105
+ desc: null
106
+ value: true
107
+ chunk_size_feed_forward:
108
+ desc: null
109
+ value: 0
110
+ is_encoder_decoder:
111
+ desc: null
112
+ value: false
113
+ is_decoder:
114
+ desc: null
115
+ value: false
116
+ cross_attention_hidden_size:
117
+ desc: null
118
+ value: null
119
+ add_cross_attention:
120
+ desc: null
121
+ value: false
122
+ tie_encoder_decoder:
123
+ desc: null
124
+ value: false
125
+ max_length:
126
+ desc: null
127
+ value: 1024
128
+ min_length:
129
+ desc: null
130
+ value: 0
131
+ do_sample:
132
+ desc: null
133
+ value: false
134
+ early_stopping:
135
+ desc: null
136
+ value: false
137
+ num_beams:
138
+ desc: null
139
+ value: 1
140
+ num_beam_groups:
141
+ desc: null
142
+ value: 1
143
+ diversity_penalty:
144
+ desc: null
145
+ value: 0.0
146
+ temperature:
147
+ desc: null
148
+ value: 1.0
149
+ top_k:
150
+ desc: null
151
+ value: 50
152
+ top_p:
153
+ desc: null
154
+ value: 1.0
155
+ typical_p:
156
+ desc: null
157
+ value: 1.0
158
+ repetition_penalty:
159
+ desc: null
160
+ value: 1.0
161
+ length_penalty:
162
+ desc: null
163
+ value: 1.0
164
+ no_repeat_ngram_size:
165
+ desc: null
166
+ value: 0
167
+ encoder_no_repeat_ngram_size:
168
+ desc: null
169
+ value: 0
170
+ bad_words_ids:
171
+ desc: null
172
+ value: null
173
+ num_return_sequences:
174
+ desc: null
175
+ value: 1
176
+ output_scores:
177
+ desc: null
178
+ value: false
179
+ return_dict_in_generate:
180
+ desc: null
181
+ value: false
182
+ forced_bos_token_id:
183
+ desc: null
184
+ value: null
185
+ forced_eos_token_id:
186
+ desc: null
187
+ value: null
188
+ remove_invalid_values:
189
+ desc: null
190
+ value: false
191
+ exponential_decay_length_penalty:
192
+ desc: null
193
+ value: null
194
+ suppress_tokens:
195
+ desc: null
196
+ value: null
197
+ begin_suppress_tokens:
198
+ desc: null
199
+ value: null
200
+ architectures:
201
+ desc: null
202
+ value: null
203
+ finetuning_task:
204
+ desc: null
205
+ value: null
206
+ id2label:
207
+ desc: null
208
+ value:
209
+ '0': LABEL_0
210
+ '1': LABEL_1
211
+ label2id:
212
+ desc: null
213
+ value:
214
+ LABEL_0: 0
215
+ LABEL_1: 1
216
+ tokenizer_class:
217
+ desc: null
218
+ value: null
219
+ prefix:
220
+ desc: null
221
+ value: null
222
+ bos_token_id:
223
+ desc: null
224
+ value: null
225
+ pad_token_id:
226
+ desc: null
227
+ value: 0
228
+ eos_token_id:
229
+ desc: null
230
+ value: null
231
+ sep_token_id:
232
+ desc: null
233
+ value: null
234
+ decoder_start_token_id:
235
+ desc: null
236
+ value: null
237
+ task_specific_params:
238
+ desc: null
239
+ value: null
240
+ problem_type:
241
+ desc: null
242
+ value: null
243
+ _name_or_path:
244
+ desc: null
245
+ value: microsoft/deberta-v3-large
246
+ transformers_version:
247
+ desc: null
248
+ value: 4.41.2
249
+ model_type:
250
+ desc: null
251
+ value: deberta-v2
252
+ position_buckets:
253
+ desc: null
254
+ value: 256
255
+ norm_rel_ebd:
256
+ desc: null
257
+ value: layer_norm
258
+ share_att_key:
259
+ desc: null
260
+ value: true
261
+ hidden_size:
262
+ desc: null
263
+ value: 1024
264
+ num_hidden_layers:
265
+ desc: null
266
+ value: 24
267
+ num_attention_heads:
268
+ desc: null
269
+ value: 16
270
+ intermediate_size:
271
+ desc: null
272
+ value: 4096
273
+ hidden_act:
274
+ desc: null
275
+ value: gelu
276
+ hidden_dropout_prob:
277
+ desc: null
278
+ value: 0.1
279
+ attention_probs_dropout_prob:
280
+ desc: null
281
+ value: 0.1
282
+ max_position_embeddings:
283
+ desc: null
284
+ value: 512
285
+ type_vocab_size:
286
+ desc: null
287
+ value: 0
288
+ initializer_range:
289
+ desc: null
290
+ value: 0.02
291
+ relative_attention:
292
+ desc: null
293
+ value: true
294
+ max_relative_positions:
295
+ desc: null
296
+ value: -1
297
+ position_biased_input:
298
+ desc: null
299
+ value: false
300
+ pos_att_type:
301
+ desc: null
302
+ value:
303
+ - p2c
304
+ - c2p
305
+ vocab_size:
306
+ desc: null
307
+ value: 128100
308
+ layer_norm_eps:
309
+ desc: null
310
+ value: 1.0e-07
311
+ pooler_hidden_size:
312
+ desc: null
313
+ value: 1024
314
+ pooler_dropout:
315
+ desc: null
316
+ value: 0
317
+ pooler_hidden_act:
318
+ desc: null
319
+ value: gelu
320
+ output_dir:
321
+ desc: null
322
+ value: ./
323
+ overwrite_output_dir:
324
+ desc: null
325
+ value: false
326
+ do_train:
327
+ desc: null
328
+ value: false
329
+ do_eval:
330
+ desc: null
331
+ value: false
332
+ do_predict:
333
+ desc: null
334
+ value: false
335
+ eval_strategy:
336
+ desc: null
337
+ value: 'no'
338
+ prediction_loss_only:
339
+ desc: null
340
+ value: false
341
+ per_device_train_batch_size:
342
+ desc: null
343
+ value: 1
344
+ per_device_eval_batch_size:
345
+ desc: null
346
+ value: 8
347
+ per_gpu_train_batch_size:
348
+ desc: null
349
+ value: null
350
+ per_gpu_eval_batch_size:
351
+ desc: null
352
+ value: null
353
+ gradient_accumulation_steps:
354
+ desc: null
355
+ value: 4
356
+ eval_accumulation_steps:
357
+ desc: null
358
+ value: null
359
+ eval_delay:
360
+ desc: null
361
+ value: 0
362
+ learning_rate:
363
+ desc: null
364
+ value: 1.41e-05
365
+ weight_decay:
366
+ desc: null
367
+ value: 0.0
368
+ adam_beta1:
369
+ desc: null
370
+ value: 0.9
371
+ adam_beta2:
372
+ desc: null
373
+ value: 0.999
374
+ adam_epsilon:
375
+ desc: null
376
+ value: 1.0e-08
377
+ max_grad_norm:
378
+ desc: null
379
+ value: 1.0
380
+ num_train_epochs:
381
+ desc: null
382
+ value: 10
383
+ max_steps:
384
+ desc: null
385
+ value: -1
386
+ lr_scheduler_type:
387
+ desc: null
388
+ value: linear
389
+ lr_scheduler_kwargs:
390
+ desc: null
391
+ value: {}
392
+ warmup_ratio:
393
+ desc: null
394
+ value: 0.0
395
+ warmup_steps:
396
+ desc: null
397
+ value: 0
398
+ log_level:
399
+ desc: null
400
+ value: passive
401
+ log_level_replica:
402
+ desc: null
403
+ value: warning
404
+ log_on_each_node:
405
+ desc: null
406
+ value: true
407
+ logging_dir:
408
+ desc: null
409
+ value: ./runs/Jun13_09-46-29_3d0b606030ff
410
+ logging_strategy:
411
+ desc: null
412
+ value: steps
413
+ logging_first_step:
414
+ desc: null
415
+ value: false
416
+ logging_steps:
417
+ desc: null
418
+ value: 500
419
+ logging_nan_inf_filter:
420
+ desc: null
421
+ value: true
422
+ save_strategy:
423
+ desc: null
424
+ value: steps
425
+ save_steps:
426
+ desc: null
427
+ value: 500
428
+ save_total_limit:
429
+ desc: null
430
+ value: null
431
+ save_safetensors:
432
+ desc: null
433
+ value: true
434
+ save_on_each_node:
435
+ desc: null
436
+ value: false
437
+ save_only_model:
438
+ desc: null
439
+ value: false
440
+ restore_callback_states_from_checkpoint:
441
+ desc: null
442
+ value: false
443
+ no_cuda:
444
+ desc: null
445
+ value: false
446
+ use_cpu:
447
+ desc: null
448
+ value: false
449
+ use_mps_device:
450
+ desc: null
451
+ value: false
452
+ seed:
453
+ desc: null
454
+ value: 42
455
+ data_seed:
456
+ desc: null
457
+ value: null
458
+ jit_mode_eval:
459
+ desc: null
460
+ value: false
461
+ use_ipex:
462
+ desc: null
463
+ value: false
464
+ bf16:
465
+ desc: null
466
+ value: false
467
+ fp16:
468
+ desc: null
469
+ value: false
470
+ fp16_opt_level:
471
+ desc: null
472
+ value: O1
473
+ half_precision_backend:
474
+ desc: null
475
+ value: auto
476
+ bf16_full_eval:
477
+ desc: null
478
+ value: false
479
+ fp16_full_eval:
480
+ desc: null
481
+ value: false
482
+ tf32:
483
+ desc: null
484
+ value: null
485
+ local_rank:
486
+ desc: null
487
+ value: 0
488
+ ddp_backend:
489
+ desc: null
490
+ value: null
491
+ tpu_num_cores:
492
+ desc: null
493
+ value: null
494
+ tpu_metrics_debug:
495
+ desc: null
496
+ value: false
497
+ debug:
498
+ desc: null
499
+ value: []
500
+ dataloader_drop_last:
501
+ desc: null
502
+ value: false
503
+ eval_steps:
504
+ desc: null
505
+ value: null
506
+ dataloader_num_workers:
507
+ desc: null
508
+ value: 0
509
+ dataloader_prefetch_factor:
510
+ desc: null
511
+ value: null
512
+ past_index:
513
+ desc: null
514
+ value: -1
515
+ run_name:
516
+ desc: null
517
+ value: reward-train-facebook-opt350m
518
+ disable_tqdm:
519
+ desc: null
520
+ value: false
521
+ remove_unused_columns:
522
+ desc: null
523
+ value: false
524
+ label_names:
525
+ desc: null
526
+ value: null
527
+ load_best_model_at_end:
528
+ desc: null
529
+ value: false
530
+ metric_for_best_model:
531
+ desc: null
532
+ value: null
533
+ greater_is_better:
534
+ desc: null
535
+ value: null
536
+ ignore_data_skip:
537
+ desc: null
538
+ value: false
539
+ fsdp:
540
+ desc: null
541
+ value: []
542
+ fsdp_min_num_params:
543
+ desc: null
544
+ value: 0
545
+ fsdp_config:
546
+ desc: null
547
+ value:
548
+ min_num_params: 0
549
+ xla: false
550
+ xla_fsdp_v2: false
551
+ xla_fsdp_grad_ckpt: false
552
+ fsdp_transformer_layer_cls_to_wrap:
553
+ desc: null
554
+ value: null
555
+ accelerator_config:
556
+ desc: null
557
+ value:
558
+ split_batches: false
559
+ dispatch_batches: null
560
+ even_batches: true
561
+ use_seedable_sampler: true
562
+ non_blocking: false
563
+ gradient_accumulation_kwargs: null
564
+ deepspeed:
565
+ desc: null
566
+ value: null
567
+ label_smoothing_factor:
568
+ desc: null
569
+ value: 0.0
570
+ optim:
571
+ desc: null
572
+ value: adamw_torch
573
+ optim_args:
574
+ desc: null
575
+ value: null
576
+ adafactor:
577
+ desc: null
578
+ value: false
579
+ group_by_length:
580
+ desc: null
581
+ value: false
582
+ length_column_name:
583
+ desc: null
584
+ value: length
585
+ report_to:
586
+ desc: null
587
+ value:
588
+ - tensorboard
589
+ - wandb
590
+ ddp_find_unused_parameters:
591
+ desc: null
592
+ value: null
593
+ ddp_bucket_cap_mb:
594
+ desc: null
595
+ value: null
596
+ ddp_broadcast_buffers:
597
+ desc: null
598
+ value: null
599
+ dataloader_pin_memory:
600
+ desc: null
601
+ value: true
602
+ dataloader_persistent_workers:
603
+ desc: null
604
+ value: false
605
+ skip_memory_metrics:
606
+ desc: null
607
+ value: true
608
+ use_legacy_prediction_loop:
609
+ desc: null
610
+ value: false
611
+ push_to_hub:
612
+ desc: null
613
+ value: false
614
+ resume_from_checkpoint:
615
+ desc: null
616
+ value: null
617
+ hub_model_id:
618
+ desc: null
619
+ value: null
620
+ hub_strategy:
621
+ desc: null
622
+ value: every_save
623
+ hub_token:
624
+ desc: null
625
+ value: <HUB_TOKEN>
626
+ hub_private_repo:
627
+ desc: null
628
+ value: false
629
+ hub_always_push:
630
+ desc: null
631
+ value: false
632
+ gradient_checkpointing:
633
+ desc: null
634
+ value: true
635
+ gradient_checkpointing_kwargs:
636
+ desc: null
637
+ value: null
638
+ include_inputs_for_metrics:
639
+ desc: null
640
+ value: false
641
+ eval_do_concat_batches:
642
+ desc: null
643
+ value: true
644
+ fp16_backend:
645
+ desc: null
646
+ value: auto
647
+ evaluation_strategy:
648
+ desc: null
649
+ value: 'no'
650
+ push_to_hub_model_id:
651
+ desc: null
652
+ value: null
653
+ push_to_hub_organization:
654
+ desc: null
655
+ value: null
656
+ push_to_hub_token:
657
+ desc: null
658
+ value: <PUSH_TO_HUB_TOKEN>
659
+ mp_parameters:
660
+ desc: null
661
+ value: ''
662
+ auto_find_batch_size:
663
+ desc: null
664
+ value: false
665
+ full_determinism:
666
+ desc: null
667
+ value: false
668
+ torchdynamo:
669
+ desc: null
670
+ value: null
671
+ ray_scope:
672
+ desc: null
673
+ value: last
674
+ ddp_timeout:
675
+ desc: null
676
+ value: 1800
677
+ torch_compile:
678
+ desc: null
679
+ value: false
680
+ torch_compile_backend:
681
+ desc: null
682
+ value: null
683
+ torch_compile_mode:
684
+ desc: null
685
+ value: null
686
+ dispatch_batches:
687
+ desc: null
688
+ value: null
689
+ split_batches:
690
+ desc: null
691
+ value: null
692
+ include_tokens_per_second:
693
+ desc: null
694
+ value: false
695
+ include_num_input_tokens_seen:
696
+ desc: null
697
+ value: false
698
+ neftune_noise_alpha:
699
+ desc: null
700
+ value: null
701
+ optim_target_modules:
702
+ desc: null
703
+ value: null
704
+ batch_eval_metrics:
705
+ desc: null
706
+ value: false
wandb/run-20240613_094645-i03sjxtv/files/output.log ADDED
@@ -0,0 +1,97 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ You're using a DebertaV2TokenizerFast tokenizer. Please note that with a fast tokenizer, using the `__call__` method is faster than using a method to encode the text followed by a call to the `pad` method to get a padded encoding.
2
+ /opt/conda/lib/python3.10/site-packages/transformers/tokenization_utils_base.py:2717: UserWarning: `max_length` is ignored when `padding`=`True` and there is no truncation strategy. To pad to max length, use `padding='max_length'`.
3
+ warnings.warn(
4
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
5
+ warnings.warn(
6
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:61: UserWarning: None of the inputs have requires_grad=True. Gradients will be None
7
+ warnings.warn(
8
+ Could not estimate the number of tokens of the input, floating-point operations will not be computed
9
+ /opt/conda/lib/python3.10/site-packages/huggingface_hub/file_download.py:1132: FutureWarning: `resume_download` is deprecated and will be removed in version 1.0.0. Downloads always resume when possible. If you want to force a new download, use `force_download=True`.
10
+ warnings.warn(
11
+ /opt/conda/lib/python3.10/site-packages/transformers/tokenization_utils_base.py:2717: UserWarning: `max_length` is ignored when `padding`=`True` and there is no truncation strategy. To pad to max length, use `padding='max_length'`.
12
+ warnings.warn(
13
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
14
+ warnings.warn(
15
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:61: UserWarning: None of the inputs have requires_grad=True. Gradients will be None
16
+ warnings.warn(
17
+ /opt/conda/lib/python3.10/site-packages/huggingface_hub/file_download.py:1132: FutureWarning: `resume_download` is deprecated and will be removed in version 1.0.0. Downloads always resume when possible. If you want to force a new download, use `force_download=True`.
18
+ warnings.warn(
19
+ /opt/conda/lib/python3.10/site-packages/transformers/tokenization_utils_base.py:2717: UserWarning: `max_length` is ignored when `padding`=`True` and there is no truncation strategy. To pad to max length, use `padding='max_length'`.
20
+ warnings.warn(
21
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
22
+ warnings.warn(
23
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:61: UserWarning: None of the inputs have requires_grad=True. Gradients will be None
24
+ warnings.warn(
25
+ /opt/conda/lib/python3.10/site-packages/huggingface_hub/file_download.py:1132: FutureWarning: `resume_download` is deprecated and will be removed in version 1.0.0. Downloads always resume when possible. If you want to force a new download, use `force_download=True`.
26
+ warnings.warn(
27
+ /opt/conda/lib/python3.10/site-packages/transformers/tokenization_utils_base.py:2717: UserWarning: `max_length` is ignored when `padding`=`True` and there is no truncation strategy. To pad to max length, use `padding='max_length'`.
28
+ warnings.warn(
29
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
30
+ warnings.warn(
31
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:61: UserWarning: None of the inputs have requires_grad=True. Gradients will be None
32
+ warnings.warn(
33
+ /opt/conda/lib/python3.10/site-packages/huggingface_hub/file_download.py:1132: FutureWarning: `resume_download` is deprecated and will be removed in version 1.0.0. Downloads always resume when possible. If you want to force a new download, use `force_download=True`.
34
+ warnings.warn(
35
+ /opt/conda/lib/python3.10/site-packages/transformers/tokenization_utils_base.py:2717: UserWarning: `max_length` is ignored when `padding`=`True` and there is no truncation strategy. To pad to max length, use `padding='max_length'`.
36
+ warnings.warn(
37
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
38
+ warnings.warn(
39
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:61: UserWarning: None of the inputs have requires_grad=True. Gradients will be None
40
+ warnings.warn(
41
+ /opt/conda/lib/python3.10/site-packages/huggingface_hub/file_download.py:1132: FutureWarning: `resume_download` is deprecated and will be removed in version 1.0.0. Downloads always resume when possible. If you want to force a new download, use `force_download=True`.
42
+ warnings.warn(
43
+ /opt/conda/lib/python3.10/site-packages/transformers/tokenization_utils_base.py:2717: UserWarning: `max_length` is ignored when `padding`=`True` and there is no truncation strategy. To pad to max length, use `padding='max_length'`.
44
+ warnings.warn(
45
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
46
+ warnings.warn(
47
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:61: UserWarning: None of the inputs have requires_grad=True. Gradients will be None
48
+ warnings.warn(
49
+ /opt/conda/lib/python3.10/site-packages/huggingface_hub/file_download.py:1132: FutureWarning: `resume_download` is deprecated and will be removed in version 1.0.0. Downloads always resume when possible. If you want to force a new download, use `force_download=True`.
50
+ warnings.warn(
51
+ /opt/conda/lib/python3.10/site-packages/transformers/tokenization_utils_base.py:2717: UserWarning: `max_length` is ignored when `padding`=`True` and there is no truncation strategy. To pad to max length, use `padding='max_length'`.
52
+ warnings.warn(
53
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
54
+ warnings.warn(
55
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:61: UserWarning: None of the inputs have requires_grad=True. Gradients will be None
56
+ warnings.warn(
57
+ /opt/conda/lib/python3.10/site-packages/huggingface_hub/file_download.py:1132: FutureWarning: `resume_download` is deprecated and will be removed in version 1.0.0. Downloads always resume when possible. If you want to force a new download, use `force_download=True`.
58
+ warnings.warn(
59
+ /opt/conda/lib/python3.10/site-packages/transformers/tokenization_utils_base.py:2717: UserWarning: `max_length` is ignored when `padding`=`True` and there is no truncation strategy. To pad to max length, use `padding='max_length'`.
60
+ warnings.warn(
61
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
62
+ warnings.warn(
63
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:61: UserWarning: None of the inputs have requires_grad=True. Gradients will be None
64
+ warnings.warn(
65
+ /opt/conda/lib/python3.10/site-packages/huggingface_hub/file_download.py:1132: FutureWarning: `resume_download` is deprecated and will be removed in version 1.0.0. Downloads always resume when possible. If you want to force a new download, use `force_download=True`.
66
+ warnings.warn(
67
+ /opt/conda/lib/python3.10/site-packages/transformers/tokenization_utils_base.py:2717: UserWarning: `max_length` is ignored when `padding`=`True` and there is no truncation strategy. To pad to max length, use `padding='max_length'`.
68
+ warnings.warn(
69
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
70
+ warnings.warn(
71
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:61: UserWarning: None of the inputs have requires_grad=True. Gradients will be None
72
+ warnings.warn(
73
+ /opt/conda/lib/python3.10/site-packages/huggingface_hub/file_download.py:1132: FutureWarning: `resume_download` is deprecated and will be removed in version 1.0.0. Downloads always resume when possible. If you want to force a new download, use `force_download=True`.
74
+ warnings.warn(
75
+ /opt/conda/lib/python3.10/site-packages/transformers/tokenization_utils_base.py:2717: UserWarning: `max_length` is ignored when `padding`=`True` and there is no truncation strategy. To pad to max length, use `padding='max_length'`.
76
+ warnings.warn(
77
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
78
+ warnings.warn(
79
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:61: UserWarning: None of the inputs have requires_grad=True. Gradients will be None
80
+ warnings.warn(
81
+ /opt/conda/lib/python3.10/site-packages/huggingface_hub/file_download.py:1132: FutureWarning: `resume_download` is deprecated and will be removed in version 1.0.0. Downloads always resume when possible. If you want to force a new download, use `force_download=True`.
82
+ warnings.warn(
83
+ /opt/conda/lib/python3.10/site-packages/transformers/tokenization_utils_base.py:2717: UserWarning: `max_length` is ignored when `padding`=`True` and there is no truncation strategy. To pad to max length, use `padding='max_length'`.
84
+ warnings.warn(
85
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
86
+ warnings.warn(
87
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:61: UserWarning: None of the inputs have requires_grad=True. Gradients will be None
88
+ warnings.warn(
89
+ /opt/conda/lib/python3.10/site-packages/huggingface_hub/file_download.py:1132: FutureWarning: `resume_download` is deprecated and will be removed in version 1.0.0. Downloads always resume when possible. If you want to force a new download, use `force_download=True`.
90
+ warnings.warn(
91
+ /opt/conda/lib/python3.10/site-packages/transformers/tokenization_utils_base.py:2717: UserWarning: `max_length` is ignored when `padding`=`True` and there is no truncation strategy. To pad to max length, use `padding='max_length'`.
92
+ warnings.warn(
93
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
94
+ warnings.warn(
95
+ /opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:61: UserWarning: None of the inputs have requires_grad=True. Gradients will be None
96
+ warnings.warn(
97
+ /opt/conda/lib/python3.10/site-packages/huggingface_hub/file_download.py:1132: FutureWarning: `resume_download` is deprecated and will be removed in version 1.0.0. Downloads always resume when possible. If you want to force a new download, use `force_download=True`.
wandb/run-20240613_094645-i03sjxtv/files/requirements.txt ADDED
@@ -0,0 +1,872 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Babel==2.14.0
2
+ Boruta==0.3
3
+ Brotli==1.1.0
4
+ CVXcanon==0.1.2
5
+ Cartopy==0.23.0
6
+ Cython==3.0.8
7
+ Deprecated==1.2.14
8
+ Farama-Notifications==0.0.4
9
+ Flask==3.0.3
10
+ Geohash==1.0
11
+ GitPython==3.1.41
12
+ ImageHash==4.3.1
13
+ Janome==0.5.0
14
+ Jinja2==3.1.2
15
+ LunarCalendar==0.0.9
16
+ Mako==1.3.5
17
+ Markdown==3.5.2
18
+ MarkupSafe==2.1.3
19
+ MarkupSafe==2.1.5
20
+ Pillow==9.5.0
21
+ PuLP==2.8.0
22
+ PyArabic==0.6.15
23
+ PyJWT==2.8.0
24
+ PyMeeus==0.5.12
25
+ PySocks==1.7.1
26
+ PyUpSet==0.1.1.post7
27
+ PyWavelets==1.5.0
28
+ PyYAML==6.0.1
29
+ Pygments==2.17.2
30
+ Pympler==1.0.1
31
+ QtPy==2.4.1
32
+ Rtree==1.2.0
33
+ SQLAlchemy==2.0.25
34
+ SecretStorage==3.3.3
35
+ Send2Trash==1.8.2
36
+ Shapely==1.8.5.post1
37
+ Shimmy==1.3.0
38
+ SimpleITK==2.3.1
39
+ TPOT==0.12.1
40
+ Theano-PyMC==1.1.2
41
+ Theano==1.0.5
42
+ Wand==0.6.13
43
+ Werkzeug==3.0.3
44
+ absl-py==1.4.0
45
+ accelerate==0.30.1
46
+ access==1.1.9
47
+ affine==2.4.0
48
+ aiobotocore==2.13.0
49
+ aiofiles==22.1.0
50
+ aiohttp==3.9.1
51
+ aioitertools==0.11.0
52
+ aiorwlock==1.3.0
53
+ aiosignal==1.3.1
54
+ aiosqlite==0.19.0
55
+ albumentations==1.4.0
56
+ alembic==1.13.1
57
+ altair==5.3.0
58
+ annotated-types==0.6.0
59
+ annotated-types==0.7.0
60
+ annoy==1.17.3
61
+ anyio==4.2.0
62
+ apache-beam==2.46.0
63
+ aplus==0.11.0
64
+ appdirs==1.4.4
65
+ archspec==0.2.3
66
+ argon2-cffi-bindings==21.2.0
67
+ argon2-cffi==23.1.0
68
+ array-record==0.5.0
69
+ arrow==1.3.0
70
+ arviz==0.18.0
71
+ astroid==3.2.2
72
+ astropy-iers-data==0.2024.6.3.0.31.14
73
+ astropy==6.1.0
74
+ asttokens==2.4.1
75
+ astunparse==1.6.3
76
+ async-lru==2.0.4
77
+ async-timeout==4.0.3
78
+ attrs==23.2.0
79
+ audioread==3.0.1
80
+ autopep8==2.0.4
81
+ backoff==2.2.1
82
+ bayesian-optimization==1.4.3
83
+ beatrix_jupyterlab==2023.128.151533
84
+ beautifulsoup4==4.12.2
85
+ blake3==0.2.1
86
+ bleach==6.1.0
87
+ blessed==1.20.0
88
+ blinker==1.8.2
89
+ blis==0.7.10
90
+ blosc2==2.6.2
91
+ bokeh==3.4.1
92
+ boltons==23.1.1
93
+ boto3==1.26.100
94
+ botocore==1.34.106
95
+ bq_helper==0.4.1
96
+ bqplot==0.12.43
97
+ branca==0.7.2
98
+ brewer2mpl==1.4.1
99
+ brotlipy==0.7.0
100
+ cached-property==1.5.2
101
+ cachetools==4.2.4
102
+ cachetools==5.3.2
103
+ catalogue==2.0.10
104
+ catalyst==22.4
105
+ catboost==1.2.5
106
+ category-encoders==2.6.3
107
+ certifi==2024.2.2
108
+ cesium==0.12.1
109
+ cffi==1.16.0
110
+ charset-normalizer==3.3.2
111
+ chex==0.1.86
112
+ cleverhans==4.0.0
113
+ click-plugins==1.1.1
114
+ click==8.1.7
115
+ cligj==0.7.2
116
+ cloud-tpu-client==0.10
117
+ cloud-tpu-profiler==2.4.0
118
+ cloudpathlib==0.16.0
119
+ cloudpickle==2.2.1
120
+ cloudpickle==3.0.0
121
+ cmdstanpy==1.2.3
122
+ colorama==0.4.6
123
+ colorcet==3.1.0
124
+ colorful==0.5.6
125
+ colorlog==6.8.2
126
+ colorlover==0.3.0
127
+ comm==0.2.1
128
+ conda-libmamba-solver==23.12.0
129
+ conda-package-handling==2.2.0
130
+ conda==24.5.0
131
+ conda_package_streaming==0.9.0
132
+ confection==0.1.4
133
+ contextily==1.6.0
134
+ contourpy==1.2.0
135
+ contourpy==1.2.1
136
+ convertdate==2.4.0
137
+ crcmod==1.7
138
+ cryptography==41.0.7
139
+ cuda-python==12.5.0
140
+ cudf==24.4.1
141
+ cufflinks==0.17.3
142
+ cuml==24.4.0
143
+ cupy==13.1.0
144
+ cycler==0.12.1
145
+ cymem==2.0.8
146
+ cytoolz==0.12.3
147
+ daal4py==2024.4.0
148
+ daal==2024.4.0
149
+ dacite==1.8.1
150
+ dask-cuda==24.4.0
151
+ dask-cudf==24.4.1
152
+ dask-expr==1.1.2
153
+ dask==2024.5.2
154
+ dataclasses-json==0.6.6
155
+ dataproc_jupyter_plugin==0.1.66
156
+ datasets==2.19.2
157
+ datashader==0.16.2
158
+ datatile==1.0.3
159
+ db-dtypes==1.2.0
160
+ deap==1.4.1
161
+ debugpy==1.8.0
162
+ decorator==5.1.1
163
+ deepdiff==7.0.1
164
+ defusedxml==0.7.1
165
+ deprecation==2.1.0
166
+ descartes==1.1.0
167
+ dill==0.3.8
168
+ dipy==1.9.0
169
+ distlib==0.3.8
170
+ distributed==2024.1.1
171
+ distro==1.9.0
172
+ dm-tree==0.1.8
173
+ docker-pycreds==0.4.0
174
+ docker==7.0.0
175
+ docopt==0.6.2
176
+ docstring-parser==0.15
177
+ docstring-to-markdown==0.15
178
+ docutils==0.21.2
179
+ earthengine-api==0.1.405
180
+ easydict==1.13
181
+ easyocr==1.7.1
182
+ ecos==2.0.13
183
+ eli5==0.13.0
184
+ emoji==2.12.1
185
+ en-core-web-lg==3.7.1
186
+ en-core-web-sm==3.7.1
187
+ entrypoints==0.4
188
+ ephem==4.1.5
189
+ esda==2.5.1
190
+ essentia==2.1b6.dev1110
191
+ et-xmlfile==1.1.0
192
+ etils==1.6.0
193
+ exceptiongroup==1.2.0
194
+ executing==2.0.1
195
+ explainable-ai-sdk==1.3.3
196
+ fastai==2.7.15
197
+ fastapi==0.108.0
198
+ fastavro==1.9.3
199
+ fastcore==1.5.43
200
+ fastdownload==0.0.7
201
+ fasteners==0.19
202
+ fastjsonschema==2.19.1
203
+ fastprogress==1.0.3
204
+ fastrlock==0.8.2
205
+ fasttext==0.9.2
206
+ feather-format==0.4.1
207
+ featuretools==1.31.0
208
+ filelock==3.13.1
209
+ fiona==1.9.6
210
+ fitter==1.7.0
211
+ flake8==7.0.0
212
+ flashtext==2.7
213
+ flatbuffers==23.5.26
214
+ flax==0.8.4
215
+ folium==0.16.0
216
+ fonttools==4.47.0
217
+ fonttools==4.53.0
218
+ fqdn==1.5.1
219
+ frozendict==2.4.4
220
+ frozenlist==1.4.1
221
+ fsspec==2024.3.1
222
+ fsspec==2024.5.0
223
+ funcy==2.0
224
+ fury==0.10.0
225
+ future==1.0.0
226
+ fuzzywuzzy==0.18.0
227
+ gast==0.5.4
228
+ gatspy==0.3
229
+ gcsfs==2024.3.1
230
+ gensim==4.3.2
231
+ geographiclib==2.0
232
+ geojson==3.1.0
233
+ geopandas==0.14.4
234
+ geoplot==0.5.1
235
+ geopy==2.4.1
236
+ geoviews==1.12.0
237
+ ggplot==0.11.5
238
+ giddy==2.3.5
239
+ gitdb==4.0.11
240
+ google-ai-generativelanguage==0.6.4
241
+ google-api-core==2.11.1
242
+ google-api-core==2.19.0
243
+ google-api-python-client==2.131.0
244
+ google-apitools==0.5.31
245
+ google-auth-httplib2==0.2.0
246
+ google-auth-oauthlib==1.2.0
247
+ google-auth==2.26.1
248
+ google-cloud-aiplatform==0.6.0a1
249
+ google-cloud-artifact-registry==1.10.0
250
+ google-cloud-automl==1.0.1
251
+ google-cloud-bigquery==2.34.4
252
+ google-cloud-bigtable==1.7.3
253
+ google-cloud-core==2.4.1
254
+ google-cloud-datastore==2.19.0
255
+ google-cloud-dlp==3.14.0
256
+ google-cloud-jupyter-config==0.0.5
257
+ google-cloud-language==2.13.3
258
+ google-cloud-monitoring==2.18.0
259
+ google-cloud-pubsub==2.19.0
260
+ google-cloud-pubsublite==1.9.0
261
+ google-cloud-recommendations-ai==0.7.1
262
+ google-cloud-resource-manager==1.11.0
263
+ google-cloud-spanner==3.40.1
264
+ google-cloud-storage==1.44.0
265
+ google-cloud-translate==3.12.1
266
+ google-cloud-videointelligence==2.13.3
267
+ google-cloud-vision==2.8.0
268
+ google-crc32c==1.5.0
269
+ google-generativeai==0.6.0
270
+ google-pasta==0.2.0
271
+ google-resumable-media==2.7.0
272
+ googleapis-common-protos==1.62.0
273
+ gplearn==0.4.2
274
+ gpustat==1.0.0
275
+ gpxpy==1.6.2
276
+ graphviz==0.20.3
277
+ greenlet==3.0.3
278
+ grpc-google-iam-v1==0.12.7
279
+ grpcio-status==1.48.1
280
+ grpcio-status==1.48.2
281
+ grpcio==1.59.3
282
+ grpcio==1.60.0
283
+ gviz-api==1.10.0
284
+ gym-notices==0.0.8
285
+ gym==0.26.2
286
+ gymnasium==0.29.0
287
+ h11==0.14.0
288
+ h2o==3.46.0.2
289
+ h5netcdf==1.3.0
290
+ h5py==3.10.0
291
+ haversine==2.8.1
292
+ hdfs==2.7.3
293
+ hep-ml==0.7.2
294
+ hijri-converter==2.3.1
295
+ hmmlearn==0.3.2
296
+ holidays==0.24
297
+ holoviews==1.18.3
298
+ hpsklearn==0.1.0
299
+ html5lib==1.1
300
+ htmlmin==0.1.12
301
+ httpcore==1.0.5
302
+ httplib2==0.21.0
303
+ httptools==0.6.1
304
+ httpx==0.27.0
305
+ huggingface-hub==0.23.2
306
+ hunspell==0.5.5
307
+ hydra-slayer==0.5.0
308
+ hyperopt==0.2.7
309
+ hypertools==0.8.0
310
+ idna==3.6
311
+ igraph==0.11.5
312
+ imagecodecs==2024.6.1
313
+ imageio==2.33.1
314
+ imbalanced-learn==0.12.3
315
+ imgaug==0.4.0
316
+ importlib-metadata==6.11.0
317
+ importlib-metadata==7.0.1
318
+ importlib-resources==6.1.1
319
+ inequality==1.0.1
320
+ iniconfig==2.0.0
321
+ ipydatawidgets==4.3.5
322
+ ipykernel==6.28.0
323
+ ipyleaflet==0.19.1
324
+ ipympl==0.7.0
325
+ ipython-genutils==0.2.0
326
+ ipython-genutils==0.2.0
327
+ ipython-sql==0.5.0
328
+ ipython==8.20.0
329
+ ipyvolume==0.6.3
330
+ ipyvue==1.11.1
331
+ ipyvuetify==1.9.4
332
+ ipywebrtc==0.6.0
333
+ ipywidgets==7.7.1
334
+ isoduration==20.11.0
335
+ isort==5.13.2
336
+ isoweek==1.3.3
337
+ itsdangerous==2.2.0
338
+ jaraco.classes==3.3.0
339
+ jax-jumpy==1.0.0
340
+ jax==0.4.26
341
+ jaxlib==0.4.26.dev20240504
342
+ jedi==0.19.1
343
+ jeepney==0.8.0
344
+ jieba==0.42.1
345
+ jmespath==1.0.1
346
+ joblib==1.4.2
347
+ json5==0.9.14
348
+ jsonpatch==1.33
349
+ jsonpointer==2.4
350
+ jsonschema-specifications==2023.12.1
351
+ jsonschema==4.20.0
352
+ jupyter-console==6.6.3
353
+ jupyter-events==0.9.0
354
+ jupyter-http-over-ws==0.0.8
355
+ jupyter-leaflet==0.19.1
356
+ jupyter-lsp==1.5.1
357
+ jupyter-server-mathjax==0.2.6
358
+ jupyter-ydoc==0.2.5
359
+ jupyter_client==7.4.9
360
+ jupyter_client==8.6.0
361
+ jupyter_core==5.7.1
362
+ jupyter_server==2.12.5
363
+ jupyter_server_fileid==0.9.1
364
+ jupyter_server_proxy==4.1.0
365
+ jupyter_server_terminals==0.5.1
366
+ jupyter_server_ydoc==0.8.0
367
+ jupyterlab-lsp==5.1.0
368
+ jupyterlab-widgets==3.0.9
369
+ jupyterlab==4.2.1
370
+ jupyterlab_git==0.44.0
371
+ jupyterlab_pygments==0.3.0
372
+ jupyterlab_server==2.27.2
373
+ jupytext==1.16.0
374
+ kaggle-environments==1.14.11
375
+ kaggle==1.6.14
376
+ kagglehub==0.2.5
377
+ keras-cv==0.9.0
378
+ keras-nlp==0.12.1
379
+ keras-tuner==1.4.6
380
+ keras==3.3.3
381
+ kernels-mixer==0.0.7
382
+ keyring==24.3.0
383
+ keyrings.google-artifactregistry-auth==1.1.2
384
+ kfp-pipeline-spec==0.2.2
385
+ kfp-server-api==2.0.5
386
+ kfp==2.5.0
387
+ kiwisolver==1.4.5
388
+ kmapper==2.0.1
389
+ kmodes==0.12.2
390
+ korean-lunar-calendar==0.3.1
391
+ kornia==0.7.2
392
+ kornia_rs==0.1.3
393
+ kt-legacy==1.0.5
394
+ kubernetes==26.1.0
395
+ langcodes==3.4.0
396
+ langid==1.1.6
397
+ language_data==1.2.0
398
+ lazy_loader==0.3
399
+ learntools==0.3.4
400
+ leven==1.0.4
401
+ libclang==16.0.6
402
+ libmambapy==1.5.8
403
+ libpysal==4.9.2
404
+ librosa==0.10.2.post1
405
+ lightgbm==4.2.0
406
+ lightning-utilities==0.11.2
407
+ lime==0.2.0.1
408
+ line_profiler==4.1.3
409
+ linkify-it-py==2.0.3
410
+ llvmlite==0.41.1
411
+ llvmlite==0.42.0
412
+ lml==0.1.0
413
+ locket==1.0.0
414
+ loguru==0.7.2
415
+ lxml==5.2.2
416
+ lz4==4.3.3
417
+ mamba==1.5.8
418
+ mapclassify==2.6.1
419
+ marisa-trie==1.1.0
420
+ markdown-it-py==3.0.0
421
+ marshmallow==3.21.2
422
+ matplotlib-inline==0.1.6
423
+ matplotlib-venn==0.11.10
424
+ matplotlib==3.7.5
425
+ matplotlib==3.8.4
426
+ mccabe==0.7.0
427
+ mdit-py-plugins==0.4.0
428
+ mdurl==0.1.2
429
+ memory-profiler==0.61.0
430
+ menuinst==2.0.1
431
+ mercantile==1.2.1
432
+ mgwr==2.2.1
433
+ missingno==0.5.2
434
+ mistune==0.8.4
435
+ mizani==0.11.4
436
+ ml-dtypes==0.2.0
437
+ mlcrate==0.2.0
438
+ mlens==0.2.3
439
+ mlxtend==0.23.1
440
+ mne==1.7.0
441
+ mnist==0.2.2
442
+ momepy==0.7.0
443
+ more-itertools==10.2.0
444
+ mpld3==0.5.10
445
+ mpmath==1.3.0
446
+ msgpack==1.0.7
447
+ msgpack==1.0.8
448
+ multidict==6.0.4
449
+ multimethod==1.10
450
+ multipledispatch==1.0.0
451
+ multiprocess==0.70.16
452
+ munkres==1.1.4
453
+ murmurhash==1.0.10
454
+ mypy-extensions==1.0.0
455
+ namex==0.0.8
456
+ nb-conda-kernels==2.3.1
457
+ nb_conda==2.2.1
458
+ nbclassic==1.0.0
459
+ nbclient==0.5.13
460
+ nbclient==0.9.0
461
+ nbconvert==6.4.5
462
+ nbdime==3.2.0
463
+ nbformat==5.9.2
464
+ ndindex==1.8
465
+ nest-asyncio==1.5.8
466
+ networkx==3.2.1
467
+ nibabel==5.2.1
468
+ nilearn==0.10.4
469
+ ninja==1.11.1.1
470
+ nltk==3.2.4
471
+ nose==1.3.7
472
+ notebook==6.5.4
473
+ notebook==6.5.6
474
+ notebook_executor==0.2
475
+ notebook_shim==0.2.3
476
+ numba==0.58.1
477
+ numba==0.59.1
478
+ numexpr==2.10.0
479
+ numpy==1.26.4
480
+ nvidia-ml-py==11.495.46
481
+ nvtx==0.2.10
482
+ oauth2client==4.1.3
483
+ oauthlib==3.2.2
484
+ objsize==0.6.1
485
+ odfpy==1.4.1
486
+ olefile==0.47
487
+ onnx==1.16.1
488
+ opencensus-context==0.1.3
489
+ opencensus==0.11.4
490
+ opencv-contrib-python==4.10.0.82
491
+ opencv-python-headless==4.10.0.82
492
+ opencv-python==4.10.0.82
493
+ openpyxl==3.1.3
494
+ openslide-python==1.3.1
495
+ opentelemetry-api==1.22.0
496
+ opentelemetry-exporter-otlp-proto-common==1.22.0
497
+ opentelemetry-exporter-otlp-proto-grpc==1.22.0
498
+ opentelemetry-exporter-otlp-proto-http==1.22.0
499
+ opentelemetry-exporter-otlp==1.22.0
500
+ opentelemetry-proto==1.22.0
501
+ opentelemetry-sdk==1.22.0
502
+ opentelemetry-semantic-conventions==0.43b0
503
+ opt-einsum==3.3.0
504
+ optax==0.2.2
505
+ optree==0.11.0
506
+ optuna==3.6.1
507
+ orbax-checkpoint==0.5.15
508
+ ordered-set==4.1.0
509
+ orjson==3.9.10
510
+ ortools==9.4.1874
511
+ osmnx==1.9.3
512
+ overrides==7.4.0
513
+ packaging==21.3
514
+ pandas-datareader==0.10.0
515
+ pandas-profiling==3.6.6
516
+ pandas-summary==0.2.0
517
+ pandas==2.2.1
518
+ pandas==2.2.2
519
+ pandasql==0.7.3
520
+ pandocfilters==1.5.0
521
+ panel==1.4.4
522
+ papermill==2.5.0
523
+ param==2.1.0
524
+ parso==0.8.3
525
+ partd==1.4.2
526
+ path.py==12.5.0
527
+ path==16.14.0
528
+ pathos==0.3.2
529
+ pathy==0.10.3
530
+ patsy==0.5.6
531
+ pdf2image==1.17.0
532
+ peft==0.11.1
533
+ pettingzoo==1.24.0
534
+ pexpect==4.8.0
535
+ pexpect==4.9.0
536
+ phik==0.12.4
537
+ pickleshare==0.7.5
538
+ pillow==10.3.0
539
+ pip==23.3.2
540
+ pkgutil_resolve_name==1.3.10
541
+ platformdirs==3.11.0
542
+ platformdirs==4.1.0
543
+ plotly-express==0.4.1
544
+ plotly==5.18.0
545
+ plotnine==0.13.6
546
+ pluggy==1.5.0
547
+ pointpats==2.4.0
548
+ polars==0.20.31
549
+ polyglot==16.7.4
550
+ pooch==1.8.1
551
+ pox==0.3.4
552
+ ppca==0.0.4
553
+ ppft==1.7.6.8
554
+ preprocessing==0.1.13
555
+ preshed==3.0.9
556
+ prettytable==3.9.0
557
+ progressbar2==4.4.2
558
+ prometheus-client==0.19.0
559
+ promise==2.3
560
+ prompt-toolkit==3.0.42
561
+ prompt-toolkit==3.0.43
562
+ prophet==1.1.1
563
+ proto-plus==1.23.0
564
+ protobuf==3.20.3
565
+ protobuf==4.24.4
566
+ psutil==5.9.3
567
+ psutil==5.9.7
568
+ ptyprocess==0.7.0
569
+ pudb==2024.1
570
+ pure-eval==0.2.2
571
+ py-cpuinfo==9.0.0
572
+ py-spy==0.3.14
573
+ py4j==0.10.9.7
574
+ pyLDAvis==3.4.1
575
+ pyOpenSSL==23.3.0
576
+ pyaml==24.4.0
577
+ pyarrow-hotfix==0.6
578
+ pyarrow==14.0.2
579
+ pyasn1-modules==0.3.0
580
+ pyasn1==0.5.1
581
+ pybind11==2.12.0
582
+ pyclipper==1.3.0.post5
583
+ pycodestyle==2.11.1
584
+ pycosat==0.6.6
585
+ pycparser==2.21
586
+ pycryptodome==3.20.0
587
+ pyct==0.5.0
588
+ pycuda==2024.1
589
+ pydantic==2.5.3
590
+ pydantic==2.7.2
591
+ pydantic_core==2.14.6
592
+ pydantic_core==2.18.3
593
+ pydegensac==0.1.2
594
+ pydicom==2.4.4
595
+ pydocstyle==6.3.0
596
+ pydot==1.4.2
597
+ pydub==0.25.1
598
+ pyemd==1.0.0
599
+ pyerfa==2.0.1.4
600
+ pyexcel-io==0.6.6
601
+ pyexcel-ods==0.6.0
602
+ pyflakes==3.2.0
603
+ pygltflib==1.16.2
604
+ pykalman==0.9.7
605
+ pylibraft==24.4.0
606
+ pylint==3.2.2
607
+ pymc3==3.11.4
608
+ pymongo==3.13.0
609
+ pynndescent==0.5.12
610
+ pynvjitlink==0.2.3
611
+ pynvml==11.4.1
612
+ pynvrtc==9.2
613
+ pyparsing==3.1.1
614
+ pyparsing==3.1.2
615
+ pypdf==4.2.0
616
+ pyproj==3.6.1
617
+ pysal==24.1
618
+ pyshp==2.3.1
619
+ pytesseract==0.3.10
620
+ pytest==8.2.1
621
+ python-bidi==0.4.2
622
+ python-dateutil==2.9.0.post0
623
+ python-dotenv==1.0.0
624
+ python-json-logger==2.0.7
625
+ python-louvain==0.16
626
+ python-lsp-jsonrpc==1.1.2
627
+ python-lsp-server==1.11.0
628
+ python-slugify==8.0.4
629
+ python-utils==3.8.2
630
+ pythreejs==2.4.2
631
+ pytoolconfig==1.3.1
632
+ pytools==2024.1.3
633
+ pytorch-ignite==0.5.0.post2
634
+ pytorch-lightning==2.2.5
635
+ pytz==2023.3.post1
636
+ pytz==2024.1
637
+ pyu2f==0.1.5
638
+ pyviz_comms==3.0.2
639
+ pyzmq==24.0.1
640
+ pyzmq==25.1.2
641
+ qgrid==1.3.1
642
+ qtconsole==5.5.2
643
+ quantecon==0.7.2
644
+ qudida==0.0.4
645
+ raft-dask==24.4.0
646
+ rapids-dask-dependency==24.4.1a0
647
+ rasterio==1.3.10
648
+ rasterstats==0.19.0
649
+ ray-cpp==2.9.0
650
+ ray==2.9.0
651
+ referencing==0.32.1
652
+ regex==2023.12.25
653
+ requests-oauthlib==1.3.1
654
+ requests-toolbelt==0.10.1
655
+ requests==2.32.3
656
+ retrying==1.3.3
657
+ retrying==1.3.4
658
+ rfc3339-validator==0.1.4
659
+ rfc3986-validator==0.1.1
660
+ rgf-python==3.12.0
661
+ rich-click==1.8.2
662
+ rich==13.7.0
663
+ rich==13.7.1
664
+ rmm==24.4.0
665
+ rope==1.13.0
666
+ rpds-py==0.16.2
667
+ rsa==4.9
668
+ ruamel-yaml-conda==0.15.100
669
+ ruamel.yaml.clib==0.2.7
670
+ ruamel.yaml==0.18.5
671
+ s2sphere==0.2.5
672
+ s3fs==2024.3.1
673
+ s3transfer==0.6.2
674
+ safetensors==0.4.3
675
+ scattertext==0.1.19
676
+ scikit-image==0.22.0
677
+ scikit-learn-intelex==2024.4.0
678
+ scikit-learn==1.2.2
679
+ scikit-multilearn==0.2.0
680
+ scikit-optimize==0.10.1
681
+ scikit-plot==0.3.7
682
+ scikit-surprise==1.1.4
683
+ scipy==1.11.4
684
+ scipy==1.13.1
685
+ seaborn==0.12.2
686
+ segment_anything==1.0
687
+ segregation==2.5
688
+ semver==3.0.2
689
+ sentencepiece==0.2.0
690
+ sentry-sdk==2.3.1
691
+ setproctitle==1.3.3
692
+ setuptools-git==1.2
693
+ setuptools-scm==8.1.0
694
+ setuptools==69.0.3
695
+ shap==0.44.1
696
+ shapely==2.0.4
697
+ shellingham==1.5.4
698
+ shtab==1.7.1
699
+ simpervisor==1.0.0
700
+ simplejson==3.19.2
701
+ six==1.16.0
702
+ sklearn-pandas==2.2.0
703
+ slicer==0.0.7
704
+ smart-open==6.4.0
705
+ smmap==5.0.1
706
+ sniffio==1.3.0
707
+ snowballstemmer==2.2.0
708
+ snuggs==1.4.7
709
+ sortedcontainers==2.4.0
710
+ soundfile==0.12.1
711
+ soupsieve==2.5
712
+ soxr==0.3.7
713
+ spacy-legacy==3.0.12
714
+ spacy-loggers==1.0.5
715
+ spacy==3.7.3
716
+ spaghetti==1.7.5.post1
717
+ spectral==0.23.1
718
+ spglm==1.1.0
719
+ sphinx-rtd-theme==0.2.4
720
+ spint==1.0.7
721
+ splot==1.1.5.post1
722
+ spopt==0.6.0
723
+ spreg==1.4.2
724
+ spvcm==0.3.0
725
+ sqlparse==0.4.4
726
+ squarify==0.4.3
727
+ srsly==2.4.8
728
+ stable-baselines3==2.1.0
729
+ stack-data==0.6.2
730
+ stack-data==0.6.3
731
+ stanio==0.5.0
732
+ starlette==0.32.0.post1
733
+ statsmodels==0.14.1
734
+ stemming==1.0.1
735
+ stop-words==2018.7.23
736
+ stopit==1.1.2
737
+ stumpy==1.12.0
738
+ sympy==1.12.1
739
+ tables==3.9.2
740
+ tabulate==0.9.0
741
+ tangled-up-in-unicode==0.2.0
742
+ tbb==2021.12.0
743
+ tblib==3.0.0
744
+ tenacity==8.2.3
745
+ tensorboard-data-server==0.7.2
746
+ tensorboard-plugin-profile==2.15.0
747
+ tensorboard==2.15.1
748
+ tensorboardX==2.6.2.2
749
+ tensorflow-cloud==0.1.16
750
+ tensorflow-datasets==4.9.4
751
+ tensorflow-decision-forests==1.8.1
752
+ tensorflow-estimator==2.15.0
753
+ tensorflow-hub==0.16.1
754
+ tensorflow-io-gcs-filesystem==0.35.0
755
+ tensorflow-io==0.35.0
756
+ tensorflow-metadata==0.14.0
757
+ tensorflow-probability==0.23.0
758
+ tensorflow-serving-api==2.14.1
759
+ tensorflow-text==2.15.0
760
+ tensorflow-transform==0.14.0
761
+ tensorflow==2.15.0
762
+ tensorstore==0.1.60
763
+ termcolor==2.4.0
764
+ terminado==0.18.0
765
+ testpath==0.6.0
766
+ text-unidecode==1.3
767
+ textblob==0.18.0.post0
768
+ texttable==1.7.0
769
+ tf_keras==2.15.1
770
+ tfp-nightly==0.24.0.dev0
771
+ thinc==8.2.3
772
+ threadpoolctl==3.2.0
773
+ tifffile==2023.12.9
774
+ timm==1.0.3
775
+ tinycss2==1.2.1
776
+ tobler==0.11.2
777
+ tokenizers==0.19.1
778
+ toml==0.10.2
779
+ tomli==2.0.1
780
+ tomlkit==0.12.5
781
+ toolz==0.12.1
782
+ torch==2.1.2
783
+ torchaudio==2.1.2
784
+ torchdata==0.7.1
785
+ torchinfo==1.8.0
786
+ torchmetrics==1.4.0.post0
787
+ torchtext==0.16.2
788
+ torchvision==0.16.2
789
+ tornado==6.3.3
790
+ tqdm==4.66.4
791
+ traceml==1.0.8
792
+ traitlets==5.9.0
793
+ traittypes==0.2.1
794
+ transformers==4.41.2
795
+ treelite==4.1.2
796
+ trl==0.9.4
797
+ truststore==0.8.0
798
+ trx-python==0.2.9
799
+ tsfresh==0.20.2
800
+ typeguard==4.1.5
801
+ typer==0.9.0
802
+ typer==0.9.4
803
+ types-python-dateutil==2.8.19.20240106
804
+ typing-inspect==0.9.0
805
+ typing-utils==0.1.0
806
+ typing_extensions==4.9.0
807
+ tyro==0.8.4
808
+ tzdata==2023.4
809
+ tzdata==2024.1
810
+ uc-micro-py==1.0.3
811
+ ucx-py==0.37.0
812
+ ujson==5.10.0
813
+ umap-learn==0.5.6
814
+ unicodedata2==15.1.0
815
+ update-checker==0.18.0
816
+ uri-template==1.3.0
817
+ uritemplate==3.0.1
818
+ urllib3==1.26.18
819
+ urllib3==2.1.0
820
+ urwid==2.6.12
821
+ urwid_readline==0.14
822
+ uvicorn==0.25.0
823
+ uvloop==0.19.0
824
+ vaex-astro==0.9.3
825
+ vaex-core==4.17.1
826
+ vaex-hdf5==0.14.1
827
+ vaex-jupyter==0.8.2
828
+ vaex-ml==0.18.3
829
+ vaex-server==0.9.0
830
+ vaex-viz==0.5.4
831
+ vaex==4.17.0
832
+ vec_noise==1.1.4
833
+ vecstack==0.4.0
834
+ virtualenv==20.21.0
835
+ visions==0.7.5
836
+ vowpalwabbit==9.9.0
837
+ vtk==9.3.0
838
+ wandb==0.17.0
839
+ wasabi==1.1.2
840
+ watchfiles==0.21.0
841
+ wavio==0.0.9
842
+ wcwidth==0.2.13
843
+ weasel==0.3.4
844
+ webcolors==1.13
845
+ webencodings==0.5.1
846
+ websocket-client==1.7.0
847
+ websockets==12.0
848
+ wfdb==4.1.2
849
+ whatthepatch==1.0.5
850
+ wheel==0.42.0
851
+ widgetsnbextension==3.6.6
852
+ witwidget==1.8.1
853
+ woodwork==0.31.0
854
+ wordcloud==1.9.3
855
+ wordsegment==1.3.1
856
+ wrapt==1.14.1
857
+ xarray-einstats==0.7.0
858
+ xarray==2024.5.0
859
+ xgboost==2.0.3
860
+ xvfbwrapper==0.2.9
861
+ xxhash==3.4.1
862
+ xyzservices==2024.4.0
863
+ y-py==0.6.2
864
+ yapf==0.40.2
865
+ yarl==1.9.3
866
+ yarl==1.9.4
867
+ ydata-profiling==4.6.4
868
+ yellowbrick==1.5
869
+ ypy-websocket==0.8.4
870
+ zict==3.0.0
871
+ zipp==3.17.0
872
+ zstandard==0.19.0
wandb/run-20240613_094645-i03sjxtv/files/wandb-metadata.json ADDED
@@ -0,0 +1,62 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.15.133+-x86_64-with-glibc2.31",
3
+ "python": "3.10.13",
4
+ "heartbeatAt": "2024-06-13T09:46:46.264588",
5
+ "startedAt": "2024-06-13T09:46:45.543499",
6
+ "docker": null,
7
+ "cuda": null,
8
+ "args": [],
9
+ "state": "running",
10
+ "program": "kaggle.ipynb",
11
+ "codePathLocal": null,
12
+ "root": "/kaggle/working",
13
+ "host": "3d0b606030ff",
14
+ "username": "root",
15
+ "executable": "/opt/conda/bin/python3.10",
16
+ "cpu_count": 2,
17
+ "cpu_count_logical": 4,
18
+ "cpu_freq": {
19
+ "current": 2000.164,
20
+ "min": 0.0,
21
+ "max": 0.0
22
+ },
23
+ "cpu_freq_per_core": [
24
+ {
25
+ "current": 2000.164,
26
+ "min": 0.0,
27
+ "max": 0.0
28
+ },
29
+ {
30
+ "current": 2000.164,
31
+ "min": 0.0,
32
+ "max": 0.0
33
+ },
34
+ {
35
+ "current": 2000.164,
36
+ "min": 0.0,
37
+ "max": 0.0
38
+ },
39
+ {
40
+ "current": 2000.164,
41
+ "min": 0.0,
42
+ "max": 0.0
43
+ }
44
+ ],
45
+ "disk": {
46
+ "/": {
47
+ "total": 8062.387607574463,
48
+ "used": 5690.186828613281
49
+ }
50
+ },
51
+ "gpu": "Tesla P100-PCIE-16GB",
52
+ "gpu_count": 1,
53
+ "gpu_devices": [
54
+ {
55
+ "name": "Tesla P100-PCIE-16GB",
56
+ "memory_total": 17179869184
57
+ }
58
+ ],
59
+ "memory": {
60
+ "total": 31.357563018798828
61
+ }
62
+ }
wandb/run-20240613_094645-i03sjxtv/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"train/loss": 0.6905, "train/grad_norm": 1.481561303138733, "train/learning_rate": 9.559322033898305e-07, "train/epoch": 9.995764506565015, "train/global_step": 5900, "_timestamp": 1718280274.0059853, "_runtime": 8268.45505619049, "_step": 11, "train_runtime": 8283.918, "train_samples_per_second": 2.85, "train_steps_per_second": 0.712, "total_flos": 0.0, "train_loss": 0.6911494057865466}
wandb/run-20240613_094645-i03sjxtv/logs/debug-internal.log ADDED
The diff for this file is too large to render. See raw diff
 
wandb/run-20240613_094645-i03sjxtv/logs/debug.log ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-06-13 09:46:45,545 INFO MainThread:35 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0
2
+ 2024-06-13 09:46:45,545 INFO MainThread:35 [wandb_setup.py:_flush():76] Configure stats pid to 35
3
+ 2024-06-13 09:46:45,545 INFO MainThread:35 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
4
+ 2024-06-13 09:46:45,545 INFO MainThread:35 [wandb_setup.py:_flush():76] Loading settings from /kaggle/working/wandb/settings
5
+ 2024-06-13 09:46:45,545 INFO MainThread:35 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
6
+ 2024-06-13 09:46:45,545 INFO MainThread:35 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
7
+ 2024-06-13 09:46:45,545 INFO MainThread:35 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program': '<python with no main file>'}
8
+ 2024-06-13 09:46:45,545 INFO MainThread:35 [wandb_setup.py:_flush():76] Applying login settings: {}
9
+ 2024-06-13 09:46:45,545 INFO MainThread:35 [wandb_setup.py:_flush():76] Applying login settings: {'api_key': '***REDACTED***'}
10
+ 2024-06-13 09:46:45,545 INFO MainThread:35 [wandb_init.py:_log_setup():520] Logging user logs to /kaggle/working/wandb/run-20240613_094645-i03sjxtv/logs/debug.log
11
+ 2024-06-13 09:46:45,545 INFO MainThread:35 [wandb_init.py:_log_setup():521] Logging internal logs to /kaggle/working/wandb/run-20240613_094645-i03sjxtv/logs/debug-internal.log
12
+ 2024-06-13 09:46:45,545 INFO MainThread:35 [wandb_init.py:_jupyter_setup():466] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x7877e40d8160>
13
+ 2024-06-13 09:46:45,546 INFO MainThread:35 [wandb_init.py:init():560] calling init triggers
14
+ 2024-06-13 09:46:45,546 INFO MainThread:35 [wandb_init.py:init():567] wandb.init called with sweep_config: {}
15
+ config: {}
16
+ 2024-06-13 09:46:45,546 INFO MainThread:35 [wandb_init.py:init():610] starting backend
17
+ 2024-06-13 09:46:45,546 INFO MainThread:35 [wandb_init.py:init():614] setting up manager
18
+ 2024-06-13 09:46:45,548 INFO MainThread:35 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
19
+ 2024-06-13 09:46:45,550 INFO MainThread:35 [wandb_init.py:init():622] backend started and connected
20
+ 2024-06-13 09:46:45,564 INFO MainThread:35 [wandb_run.py:_label_probe_notebook():1328] probe notebook
21
+ 2024-06-13 09:46:45,978 INFO MainThread:35 [wandb_init.py:init():711] updated telemetry
22
+ 2024-06-13 09:46:45,981 INFO MainThread:35 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout
23
+ 2024-06-13 09:46:46,175 INFO MainThread:35 [wandb_run.py:_on_init():2396] communicating current version
24
+ 2024-06-13 09:46:46,232 INFO MainThread:35 [wandb_run.py:_on_init():2405] got version response upgrade_message: "wandb version 0.17.1 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
25
+
26
+ 2024-06-13 09:46:46,232 INFO MainThread:35 [wandb_init.py:init():795] starting run threads in backend
27
+ 2024-06-13 09:47:02,296 INFO MainThread:35 [wandb_run.py:_console_start():2374] atexit reg
28
+ 2024-06-13 09:47:02,296 INFO MainThread:35 [wandb_run.py:_redirect():2229] redirect: wrap_raw
29
+ 2024-06-13 09:47:02,297 INFO MainThread:35 [wandb_run.py:_redirect():2294] Wrapping output streams.
30
+ 2024-06-13 09:47:02,297 INFO MainThread:35 [wandb_run.py:_redirect():2319] Redirects installed.
31
+ 2024-06-13 09:47:02,304 INFO MainThread:35 [wandb_init.py:init():838] run started, returning control to user process
32
+ 2024-06-13 09:47:02,312 INFO MainThread:35 [wandb_run.py:_config_callback():1376] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 1024, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': None, 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': 0, 'eos_token_id': None, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'microsoft/deberta-v3-large', 'transformers_version': '4.41.2', 'model_type': 'deberta-v2', 'position_buckets': 256, 'norm_rel_ebd': 'layer_norm', 'share_att_key': True, 'hidden_size': 1024, 'num_hidden_layers': 24, 'num_attention_heads': 16, 'intermediate_size': 4096, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'attention_probs_dropout_prob': 0.1, 'max_position_embeddings': 512, 'type_vocab_size': 0, 'initializer_range': 0.02, 'relative_attention': True, 'max_relative_positions': -1, 'position_biased_input': False, 'pos_att_type': ['p2c', 'c2p'], 'vocab_size': 128100, 'layer_norm_eps': 1e-07, 'pooler_hidden_size': 1024, 'pooler_dropout': 0, 'pooler_hidden_act': 'gelu', 'output_dir': './', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': False, 'do_predict': False, 'eval_strategy': 'no', 'prediction_loss_only': False, 'per_device_train_batch_size': 1, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 4, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 1.41e-05, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 10, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': './runs/Jun13_09-46-29_3d0b606030ff', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'restore_callback_states_from_checkpoint': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': 'reward-train-facebook-opt350m', 'disable_tqdm': False, 'remove_unused_columns': False, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': True, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'eval_do_concat_batches': True, 'fp16_backend': 'auto', 'evaluation_strategy': 'no', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None, 'batch_eval_metrics': False}
33
+ 2024-06-13 12:04:34,011 INFO MainThread:35 [jupyter.py:save_ipynb():373] not saving jupyter notebook
34
+ 2024-06-13 12:04:34,011 INFO MainThread:35 [wandb_init.py:_pause_backend():431] pausing backend
35
+ 2024-06-13 12:25:21,829 INFO MainThread:35 [wandb_init.py:_resume_backend():436] resuming backend
36
+ 2024-06-13 12:25:24,437 INFO MainThread:35 [jupyter.py:save_ipynb():373] not saving jupyter notebook
37
+ 2024-06-13 12:25:24,437 INFO MainThread:35 [wandb_init.py:_pause_backend():431] pausing backend
38
+ 2024-06-13 12:25:24,444 INFO MainThread:35 [wandb_init.py:_resume_backend():436] resuming backend
wandb/run-20240613_094645-i03sjxtv/run-i03sjxtv.wandb ADDED
Binary file (205 kB). View file