sharukat commited on
Commit
563e70d
·
verified ·
1 Parent(s): 2c0d046

Training in progress, epoch 1

Browse files
config.json CHANGED
@@ -1,9 +1,11 @@
1
  {
2
- "_name_or_path": "mmukh/SOBertBase",
 
3
  "architectures": [
4
- "MegatronBertForSequenceClassification"
5
  ],
6
  "attention_probs_dropout_prob": 0.1,
 
7
  "hidden_act": "gelu",
8
  "hidden_dropout_prob": 0.1,
9
  "hidden_size": 768,
@@ -28,17 +30,18 @@
28
  "Requesting (Additional) Resources": 1
29
  },
30
  "layer_norm_eps": 1e-12,
31
- "max_position_embeddings": 2048,
32
- "model_type": "megatron-bert",
33
  "num_attention_heads": 12,
34
  "num_hidden_layers": 12,
 
 
35
  "pad_token_id": 0,
36
  "position_embedding_type": "absolute",
37
  "problem_type": "single_label_classification",
38
- "tokenizer_type": "SentencePieceTokenizer",
39
  "torch_dtype": "float32",
40
  "transformers_version": "4.39.3",
41
  "type_vocab_size": 2,
42
  "use_cache": true,
43
- "vocab_size": 50048
44
  }
 
1
  {
2
+ "_name_or_path": "lanwuwei/BERTOverflow_stackoverflow_github",
3
+ "_num_labels": 51,
4
  "architectures": [
5
+ "BertForSequenceClassification"
6
  ],
7
  "attention_probs_dropout_prob": 0.1,
8
+ "classifier_dropout": null,
9
  "hidden_act": "gelu",
10
  "hidden_dropout_prob": 0.1,
11
  "hidden_size": 768,
 
30
  "Requesting (Additional) Resources": 1
31
  },
32
  "layer_norm_eps": 1e-12,
33
+ "max_position_embeddings": 512,
34
+ "model_type": "bert",
35
  "num_attention_heads": 12,
36
  "num_hidden_layers": 12,
37
+ "output_hidden_states": true,
38
+ "output_past": true,
39
  "pad_token_id": 0,
40
  "position_embedding_type": "absolute",
41
  "problem_type": "single_label_classification",
 
42
  "torch_dtype": "float32",
43
  "transformers_version": "4.39.3",
44
  "type_vocab_size": 2,
45
  "use_cache": true,
46
+ "vocab_size": 82000
47
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:67a4cb420958d75a3bb4a7ce6dc06955cc986e3d340535eadfc1a5407026e1a6
3
- size 502675828
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7f32bd9201f86bc9fb8c4e9d255d1b876209f641dbd6b9464fded038413d3c76
3
+ size 596114452
runs/Apr10_15-56-15_ae604480b8c9/events.out.tfevents.1712764576.ae604480b8c9.34.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c4d3c98682ad82e34eae35f85288d481380e684d780f154aa17241d6390219a
3
+ size 5294
runs/Apr10_16-01-09_ae604480b8c9/events.out.tfevents.1712764870.ae604480b8c9.34.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:580e87cbf893958dcf1a4793865ee37cd7a60b13d87cfc1dfdfc3e56cd3cb295
3
+ size 5295
runs/Apr10_16-02-54_ae604480b8c9/events.out.tfevents.1712764975.ae604480b8c9.34.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b58cceece7dc26974f0aea186027aae16c4f5b8b24ac602b6f91e367b2d9380c
3
+ size 5503
runs/Apr10_16-14-35_ae604480b8c9/events.out.tfevents.1712765676.ae604480b8c9.34.3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:094b221471e68fd9d96eba31b1362785de551207154070224f708c8299a840c4
3
+ size 5503
runs/Apr10_16-17-22_ae604480b8c9/events.out.tfevents.1712765843.ae604480b8c9.34.4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:00e26802e615e75e71a61425bab2310dfd2f53924b9f6c49f39ea6051d4c68ce
3
+ size 5966
special_tokens_map.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "[CLS]",
3
+ "cls_token": "[CLS]",
4
+ "eos_token": "[SEP]",
5
+ "mask_token": "[MASK]",
6
+ "pad_token": "[PAD]",
7
+ "sep_token": "[SEP]",
8
+ "unk_token": "[UNK]"
9
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,59 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "[UNK]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "[CLS]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "3": {
28
+ "content": "[MASK]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "4": {
36
+ "content": "[SEP]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "bos_token": "[CLS]",
45
+ "clean_up_tokenization_spaces": true,
46
+ "cls_token": "[CLS]",
47
+ "do_basic_tokenize": true,
48
+ "do_lower_case": false,
49
+ "eos_token": "[SEP]",
50
+ "mask_token": "[MASK]",
51
+ "model_max_length": 1000000000000000019884624838656,
52
+ "never_split": null,
53
+ "pad_token": "[PAD]",
54
+ "sep_token": "[SEP]",
55
+ "strip_accents": null,
56
+ "tokenize_chinese_chars": true,
57
+ "tokenizer_class": "BertTokenizer",
58
+ "unk_token": "[UNK]"
59
+ }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:89a2ed4d7d292fe5780011c7f589882560b0e4256bbecfb8bacfeeb6dd55757e
3
  size 4920
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d6eb9a293f527bc9d639d37280a5d4d9934c34913b6e09a6a6715b4bc8841456
3
  size 4920
vocab.txt ADDED
The diff for this file is too large to render. See raw diff
 
wandb/debug-internal.log CHANGED
The diff for this file is too large to render. See raw diff
 
wandb/debug.log CHANGED
@@ -1,36 +1,153 @@
1
- 2024-04-10 15:01:23,544 INFO MainThread:34 [wandb_setup.py:_flush():76] Current SDK version is 0.16.5
2
- 2024-04-10 15:01:23,544 INFO MainThread:34 [wandb_setup.py:_flush():76] Configure stats pid to 34
3
- 2024-04-10 15:01:23,544 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
4
- 2024-04-10 15:01:23,544 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from /kaggle/working/wandb/settings
5
- 2024-04-10 15:01:23,544 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
6
- 2024-04-10 15:01:23,544 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
7
- 2024-04-10 15:01:23,544 INFO MainThread:34 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program': '<python with no main file>'}
8
- 2024-04-10 15:01:23,545 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying login settings: {}
9
- 2024-04-10 15:01:23,545 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying login settings: {'api_key': '***REDACTED***'}
10
- 2024-04-10 15:01:23,545 INFO MainThread:34 [wandb_init.py:_log_setup():527] Logging user logs to /kaggle/working/wandb/run-20240410_150123-cuua991r/logs/debug.log
11
- 2024-04-10 15:01:23,545 INFO MainThread:34 [wandb_init.py:_log_setup():528] Logging internal logs to /kaggle/working/wandb/run-20240410_150123-cuua991r/logs/debug-internal.log
12
- 2024-04-10 15:01:23,545 INFO MainThread:34 [wandb_init.py:_jupyter_setup():473] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x7f00b2f33ee0>
13
- 2024-04-10 15:01:23,545 INFO MainThread:34 [wandb_init.py:init():567] calling init triggers
14
- 2024-04-10 15:01:23,545 INFO MainThread:34 [wandb_init.py:init():574] wandb.init called with sweep_config: {}
15
  config: {}
16
- 2024-04-10 15:01:23,545 INFO MainThread:34 [wandb_init.py:init():617] starting backend
17
- 2024-04-10 15:01:23,545 INFO MainThread:34 [wandb_init.py:init():621] setting up manager
18
- 2024-04-10 15:01:23,547 INFO MainThread:34 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
19
- 2024-04-10 15:01:23,550 INFO MainThread:34 [wandb_init.py:init():629] backend started and connected
20
- 2024-04-10 15:01:23,564 INFO MainThread:34 [wandb_run.py:_label_probe_notebook():1299] probe notebook
21
- 2024-04-10 15:01:25,152 INFO MainThread:34 [wandb_init.py:init():721] updated telemetry
22
- 2024-04-10 15:01:25,155 INFO MainThread:34 [wandb_init.py:init():754] communicating run to backend with 90.0 second timeout
23
- 2024-04-10 15:01:25,657 INFO MainThread:34 [wandb_run.py:_on_init():2344] communicating current version
24
- 2024-04-10 15:01:25,751 INFO MainThread:34 [wandb_run.py:_on_init():2353] got version response upgrade_message: "wandb version 0.16.6 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
25
 
26
- 2024-04-10 15:01:25,751 INFO MainThread:34 [wandb_init.py:init():805] starting run threads in backend
27
- 2024-04-10 15:01:41,746 INFO MainThread:34 [wandb_run.py:_console_start():2323] atexit reg
28
- 2024-04-10 15:01:41,746 INFO MainThread:34 [wandb_run.py:_redirect():2178] redirect: wrap_raw
29
- 2024-04-10 15:01:41,747 INFO MainThread:34 [wandb_run.py:_redirect():2243] Wrapping output streams.
30
- 2024-04-10 15:01:41,747 INFO MainThread:34 [wandb_run.py:_redirect():2268] Redirects installed.
31
- 2024-04-10 15:01:41,749 INFO MainThread:34 [wandb_init.py:init():848] run started, returning control to user process
32
- 2024-04-10 15:01:41,754 INFO MainThread:34 [wandb_run.py:_config_callback():1347] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': None, 'finetuning_task': None, 'id2label': {0: 'Documentation Replication on Other Examples', 1: 'Requesting (Additional) Resources', 2: 'Inadequate Examples', 3: 'Documentation Ambiguity', 4: 'Documentation Completeness', 5: 'Lack of Alternative Solutions/Documentation', 6: 'Documentation Replicability'}, 'label2id': {'Documentation Replication on Other Examples': 0, 'Requesting (Additional) Resources': 1, 'Inadequate Examples': 2, 'Documentation Ambiguity': 3, 'Documentation Completeness': 4, 'Lack of Alternative Solutions/Documentation': 5, 'Documentation Replicability': 6}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': 0, 'eos_token_id': None, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'mmukh/SOBertBase', 'transformers_version': '4.39.3', 'model_type': 'megatron-bert', 'tokenizer_type': 'SentencePieceTokenizer', 'vocab_size': 50048, 'hidden_size': 768, 'num_hidden_layers': 12, 'num_attention_heads': 12, 'hidden_act': 'gelu', 'intermediate_size': 3072, 'hidden_dropout_prob': 0.1, 'attention_probs_dropout_prob': 0.1, 'max_position_embeddings': 2048, 'type_vocab_size': 2, 'initializer_range': 0.02, 'layer_norm_eps': 1e-12, 'position_embedding_type': 'absolute', 'use_cache': True, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'epoch', 'prediction_loss_only': False, 'per_device_train_batch_size': 8, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 2e-05, 'weight_decay': 0.0, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 5, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Apr10_14-58-13_9088970cc926', 'logging_strategy': 'epoch', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'epoch', 'save_steps': 500, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None}
33
- 2024-04-10 15:06:23,666 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
34
- 2024-04-10 15:06:23,666 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
35
- 2024-04-10 15:06:30,822 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
36
- 2024-04-10 15:06:32,237 INFO MainThread:34 [wandb_run.py:_config_callback():1347] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': None, 'finetuning_task': None, 'id2label': {0: 'Documentation Replication on Other Examples', 1: 'Requesting (Additional) Resources', 2: 'Inadequate Examples', 3: 'Documentation Ambiguity', 4: 'Documentation Completeness', 5: 'Lack of Alternative Solutions/Documentation', 6: 'Documentation Replicability'}, 'label2id': {'Documentation Replication on Other Examples': 0, 'Requesting (Additional) Resources': 1, 'Inadequate Examples': 2, 'Documentation Ambiguity': 3, 'Documentation Completeness': 4, 'Lack of Alternative Solutions/Documentation': 5, 'Documentation Replicability': 6}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': 0, 'eos_token_id': None, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'mmukh/SOBertBase', 'transformers_version': '4.39.3', 'model_type': 'megatron-bert', 'tokenizer_type': 'SentencePieceTokenizer', 'vocab_size': 50048, 'hidden_size': 768, 'num_hidden_layers': 12, 'num_attention_heads': 12, 'hidden_act': 'gelu', 'intermediate_size': 3072, 'hidden_dropout_prob': 0.1, 'attention_probs_dropout_prob': 0.1, 'max_position_embeddings': 2048, 'type_vocab_size': 2, 'initializer_range': 0.02, 'layer_norm_eps': 1e-12, 'position_embedding_type': 'absolute', 'use_cache': True, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'epoch', 'prediction_loss_only': False, 'per_device_train_batch_size': 8, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 1e-05, 'weight_decay': 0.01, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 10, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Apr10_15-06-31_9088970cc926', 'logging_strategy': 'epoch', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'epoch', 'save_steps': 500, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None}
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-04-10 15:56:50,262 INFO MainThread:34 [wandb_setup.py:_flush():76] Current SDK version is 0.16.5
2
+ 2024-04-10 15:56:50,262 INFO MainThread:34 [wandb_setup.py:_flush():76] Configure stats pid to 34
3
+ 2024-04-10 15:56:50,263 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
4
+ 2024-04-10 15:56:50,263 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from /kaggle/working/wandb/settings
5
+ 2024-04-10 15:56:50,263 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
6
+ 2024-04-10 15:56:50,263 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
7
+ 2024-04-10 15:56:50,263 INFO MainThread:34 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program': '<python with no main file>'}
8
+ 2024-04-10 15:56:50,263 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying login settings: {}
9
+ 2024-04-10 15:56:50,263 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying login settings: {'api_key': '***REDACTED***'}
10
+ 2024-04-10 15:56:50,263 INFO MainThread:34 [wandb_init.py:_log_setup():527] Logging user logs to /kaggle/working/wandb/run-20240410_155650-albqrbhg/logs/debug.log
11
+ 2024-04-10 15:56:50,263 INFO MainThread:34 [wandb_init.py:_log_setup():528] Logging internal logs to /kaggle/working/wandb/run-20240410_155650-albqrbhg/logs/debug-internal.log
12
+ 2024-04-10 15:56:50,263 INFO MainThread:34 [wandb_init.py:_jupyter_setup():473] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x7ff5302cfca0>
13
+ 2024-04-10 15:56:50,263 INFO MainThread:34 [wandb_init.py:init():567] calling init triggers
14
+ 2024-04-10 15:56:50,263 INFO MainThread:34 [wandb_init.py:init():574] wandb.init called with sweep_config: {}
15
  config: {}
16
+ 2024-04-10 15:56:50,264 INFO MainThread:34 [wandb_init.py:init():617] starting backend
17
+ 2024-04-10 15:56:50,264 INFO MainThread:34 [wandb_init.py:init():621] setting up manager
18
+ 2024-04-10 15:56:50,265 INFO MainThread:34 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
19
+ 2024-04-10 15:56:50,268 INFO MainThread:34 [wandb_init.py:init():629] backend started and connected
20
+ 2024-04-10 15:56:50,281 INFO MainThread:34 [wandb_run.py:_label_probe_notebook():1299] probe notebook
21
+ 2024-04-10 15:56:51,015 INFO MainThread:34 [wandb_init.py:init():721] updated telemetry
22
+ 2024-04-10 15:56:51,018 INFO MainThread:34 [wandb_init.py:init():754] communicating run to backend with 90.0 second timeout
23
+ 2024-04-10 15:56:51,225 INFO MainThread:34 [wandb_run.py:_on_init():2344] communicating current version
24
+ 2024-04-10 15:56:51,292 INFO MainThread:34 [wandb_run.py:_on_init():2353] got version response upgrade_message: "wandb version 0.16.6 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
25
 
26
+ 2024-04-10 15:56:51,292 INFO MainThread:34 [wandb_init.py:init():805] starting run threads in backend
27
+ 2024-04-10 15:57:07,286 INFO MainThread:34 [wandb_run.py:_console_start():2323] atexit reg
28
+ 2024-04-10 15:57:07,286 INFO MainThread:34 [wandb_run.py:_redirect():2178] redirect: wrap_raw
29
+ 2024-04-10 15:57:07,286 INFO MainThread:34 [wandb_run.py:_redirect():2243] Wrapping output streams.
30
+ 2024-04-10 15:57:07,287 INFO MainThread:34 [wandb_run.py:_redirect():2268] Redirects installed.
31
+ 2024-04-10 15:57:07,288 INFO MainThread:34 [wandb_init.py:init():848] run started, returning control to user process
32
+ 2024-04-10 15:57:07,294 INFO MainThread:34 [wandb_run.py:_config_callback():1347] config_cb None None {'return_dict': True, 'output_hidden_states': True, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['BertModel'], 'finetuning_task': None, 'id2label': {0: 'Documentation Replication on Other Examples', 1: 'Requesting (Additional) Resources', 2: 'Inadequate Examples', 3: 'Documentation Ambiguity', 4: 'Documentation Completeness', 5: 'Lack of Alternative Solutions/Documentation', 6: 'Documentation Replicability'}, 'label2id': {'Documentation Replication on Other Examples': 0, 'Requesting (Additional) Resources': 1, 'Inadequate Examples': 2, 'Documentation Ambiguity': 3, 'Documentation Completeness': 4, 'Lack of Alternative Solutions/Documentation': 5, 'Documentation Replicability': 6}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': 0, 'eos_token_id': None, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'lanwuwei/BERTOverflow_stackoverflow_github', 'transformers_version': '4.39.3', '_num_labels': 51, 'model_type': 'bert', 'output_past': True, 'vocab_size': 82000, 'hidden_size': 768, 'num_hidden_layers': 12, 'num_attention_heads': 12, 'hidden_act': 'gelu', 'intermediate_size': 3072, 'hidden_dropout_prob': 0.1, 'attention_probs_dropout_prob': 0.1, 'max_position_embeddings': 512, 'type_vocab_size': 2, 'initializer_range': 0.02, 'layer_norm_eps': 1e-12, 'position_embedding_type': 'absolute', 'use_cache': True, 'classifier_dropout': None, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': True, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 16, 'per_device_eval_batch_size': 16, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 1e-05, 'weight_decay': 0.01, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 5, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Apr10_15-56-15_ae604480b8c9', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': 500, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': True, 'metric_for_best_model': 'loss', 'greater_is_better': False, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None}
33
+ 2024-04-10 15:58:28,784 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
34
+ 2024-04-10 15:58:28,785 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
35
+ 2024-04-10 16:01:09,249 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
36
+ 2024-04-10 16:01:10,407 INFO MainThread:34 [wandb_run.py:_config_callback():1347] config_cb None None {'return_dict': True, 'output_hidden_states': True, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['BertModel'], 'finetuning_task': None, 'id2label': {0: 'Documentation Replication on Other Examples', 1: 'Requesting (Additional) Resources', 2: 'Inadequate Examples', 3: 'Documentation Ambiguity', 4: 'Documentation Completeness', 5: 'Lack of Alternative Solutions/Documentation', 6: 'Documentation Replicability'}, 'label2id': {'Documentation Replication on Other Examples': 0, 'Requesting (Additional) Resources': 1, 'Inadequate Examples': 2, 'Documentation Ambiguity': 3, 'Documentation Completeness': 4, 'Lack of Alternative Solutions/Documentation': 5, 'Documentation Replicability': 6}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': 0, 'eos_token_id': None, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'lanwuwei/BERTOverflow_stackoverflow_github', 'transformers_version': '4.39.3', '_num_labels': 51, 'model_type': 'bert', 'output_past': True, 'vocab_size': 82000, 'hidden_size': 768, 'num_hidden_layers': 12, 'num_attention_heads': 12, 'hidden_act': 'gelu', 'intermediate_size': 3072, 'hidden_dropout_prob': 0.1, 'attention_probs_dropout_prob': 0.1, 'max_position_embeddings': 512, 'type_vocab_size': 2, 'initializer_range': 0.02, 'layer_norm_eps': 1e-12, 'position_embedding_type': 'absolute', 'use_cache': True, 'classifier_dropout': None, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 16, 'per_device_eval_batch_size': 16, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 1e-05, 'weight_decay': 0.01, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 5, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Apr10_16-01-09_ae604480b8c9', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': 500, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': True, 'metric_for_best_model': 'loss', 'greater_is_better': False, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None}
37
+ 2024-04-10 16:02:05,703 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
38
+ 2024-04-10 16:02:05,703 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
39
+ 2024-04-10 16:02:15,608 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
40
+ 2024-04-10 16:02:15,610 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
41
+ 2024-04-10 16:02:15,610 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
42
+ 2024-04-10 16:02:54,085 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
43
+ 2024-04-10 16:02:55,203 INFO MainThread:34 [wandb_run.py:_config_callback():1347] config_cb None None {'return_dict': True, 'output_hidden_states': True, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['BertModel'], 'finetuning_task': None, 'id2label': {0: 'Documentation Replication on Other Examples', 1: 'Requesting (Additional) Resources', 2: 'Inadequate Examples', 3: 'Documentation Ambiguity', 4: 'Documentation Completeness', 5: 'Lack of Alternative Solutions/Documentation', 6: 'Documentation Replicability'}, 'label2id': {'Documentation Replication on Other Examples': 0, 'Requesting (Additional) Resources': 1, 'Inadequate Examples': 2, 'Documentation Ambiguity': 3, 'Documentation Completeness': 4, 'Lack of Alternative Solutions/Documentation': 5, 'Documentation Replicability': 6}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': 0, 'eos_token_id': None, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'lanwuwei/BERTOverflow_stackoverflow_github', 'transformers_version': '4.39.3', '_num_labels': 51, 'model_type': 'bert', 'output_past': True, 'vocab_size': 82000, 'hidden_size': 768, 'num_hidden_layers': 12, 'num_attention_heads': 12, 'hidden_act': 'gelu', 'intermediate_size': 3072, 'hidden_dropout_prob': 0.1, 'attention_probs_dropout_prob': 0.1, 'max_position_embeddings': 512, 'type_vocab_size': 2, 'initializer_range': 0.02, 'layer_norm_eps': 1e-12, 'position_embedding_type': 'absolute', 'use_cache': True, 'classifier_dropout': None, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'epoch', 'prediction_loss_only': False, 'per_device_train_batch_size': 16, 'per_device_eval_batch_size': 16, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 1e-05, 'weight_decay': 0.01, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 5, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Apr10_16-02-54_ae604480b8c9', 'logging_strategy': 'epoch', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'epoch', 'save_steps': 500, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': True, 'metric_for_best_model': 'loss', 'greater_is_better': False, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None}
44
+ 2024-04-10 16:03:53,536 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
45
+ 2024-04-10 16:03:53,536 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
46
+ 2024-04-10 16:11:50,077 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
47
+ 2024-04-10 16:11:50,079 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
48
+ 2024-04-10 16:11:50,079 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
49
+ 2024-04-10 16:12:02,272 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
50
+ 2024-04-10 16:12:02,274 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
51
+ 2024-04-10 16:12:02,275 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
52
+ 2024-04-10 16:12:15,300 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
53
+ 2024-04-10 16:12:15,302 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
54
+ 2024-04-10 16:12:15,303 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
55
+ 2024-04-10 16:12:18,676 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
56
+ 2024-04-10 16:12:18,734 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
57
+ 2024-04-10 16:12:18,735 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
58
+ 2024-04-10 16:12:20,635 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
59
+ 2024-04-10 16:12:20,643 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
60
+ 2024-04-10 16:12:20,643 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
61
+ 2024-04-10 16:12:23,458 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
62
+ 2024-04-10 16:12:23,465 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
63
+ 2024-04-10 16:12:23,465 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
64
+ 2024-04-10 16:12:27,135 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
65
+ 2024-04-10 16:12:27,146 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
66
+ 2024-04-10 16:12:27,146 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
67
+ 2024-04-10 16:12:34,928 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
68
+ 2024-04-10 16:12:34,934 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
69
+ 2024-04-10 16:12:34,934 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
70
+ 2024-04-10 16:12:51,967 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
71
+ 2024-04-10 16:12:52,930 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
72
+ 2024-04-10 16:12:52,930 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
73
+ 2024-04-10 16:13:14,813 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
74
+ 2024-04-10 16:13:14,815 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
75
+ 2024-04-10 16:13:14,815 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
76
+ 2024-04-10 16:13:22,627 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
77
+ 2024-04-10 16:13:22,634 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
78
+ 2024-04-10 16:13:22,634 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
79
+ 2024-04-10 16:13:26,638 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
80
+ 2024-04-10 16:13:26,688 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
81
+ 2024-04-10 16:13:26,688 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
82
+ 2024-04-10 16:13:28,942 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
83
+ 2024-04-10 16:13:28,949 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
84
+ 2024-04-10 16:13:28,949 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
85
+ 2024-04-10 16:13:32,221 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
86
+ 2024-04-10 16:13:32,227 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
87
+ 2024-04-10 16:13:32,227 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
88
+ 2024-04-10 16:13:35,802 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
89
+ 2024-04-10 16:13:35,814 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
90
+ 2024-04-10 16:13:35,814 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
91
+ 2024-04-10 16:13:38,310 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
92
+ 2024-04-10 16:13:38,314 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
93
+ 2024-04-10 16:13:38,314 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
94
+ 2024-04-10 16:13:43,154 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
95
+ 2024-04-10 16:13:44,088 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
96
+ 2024-04-10 16:13:44,088 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
97
+ 2024-04-10 16:13:46,903 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
98
+ 2024-04-10 16:13:46,926 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
99
+ 2024-04-10 16:13:46,926 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
100
+ 2024-04-10 16:13:48,880 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
101
+ 2024-04-10 16:13:48,891 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
102
+ 2024-04-10 16:13:48,891 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
103
+ 2024-04-10 16:13:51,328 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
104
+ 2024-04-10 16:13:51,353 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
105
+ 2024-04-10 16:13:51,353 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
106
+ 2024-04-10 16:13:53,267 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
107
+ 2024-04-10 16:13:53,268 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
108
+ 2024-04-10 16:13:53,269 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
109
+ 2024-04-10 16:14:03,755 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
110
+ 2024-04-10 16:14:04,296 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
111
+ 2024-04-10 16:14:04,296 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
112
+ 2024-04-10 16:14:34,345 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
113
+ 2024-04-10 16:14:36,175 INFO MainThread:34 [wandb_run.py:_config_callback():1347] config_cb None None {'return_dict': True, 'output_hidden_states': True, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['BertModel'], 'finetuning_task': None, 'id2label': {0: 'Documentation Replication on Other Examples', 1: 'Requesting (Additional) Resources', 2: 'Inadequate Examples', 3: 'Documentation Ambiguity', 4: 'Documentation Completeness', 5: 'Lack of Alternative Solutions/Documentation', 6: 'Documentation Replicability'}, 'label2id': {'Documentation Replication on Other Examples': 0, 'Requesting (Additional) Resources': 1, 'Inadequate Examples': 2, 'Documentation Ambiguity': 3, 'Documentation Completeness': 4, 'Lack of Alternative Solutions/Documentation': 5, 'Documentation Replicability': 6}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': 0, 'eos_token_id': None, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'lanwuwei/BERTOverflow_stackoverflow_github', 'transformers_version': '4.39.3', '_num_labels': 51, 'model_type': 'bert', 'output_past': True, 'vocab_size': 82000, 'hidden_size': 768, 'num_hidden_layers': 12, 'num_attention_heads': 12, 'hidden_act': 'gelu', 'intermediate_size': 3072, 'hidden_dropout_prob': 0.1, 'attention_probs_dropout_prob': 0.1, 'max_position_embeddings': 512, 'type_vocab_size': 2, 'initializer_range': 0.02, 'layer_norm_eps': 1e-12, 'position_embedding_type': 'absolute', 'use_cache': True, 'classifier_dropout': None, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'epoch', 'prediction_loss_only': False, 'per_device_train_batch_size': 16, 'per_device_eval_batch_size': 16, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 1e-05, 'weight_decay': 0.01, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 5, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Apr10_16-14-35_ae604480b8c9', 'logging_strategy': 'epoch', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'epoch', 'save_steps': 500, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': True, 'metric_for_best_model': 'loss', 'greater_is_better': False, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None}
114
+ 2024-04-10 16:15:34,096 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
115
+ 2024-04-10 16:15:34,096 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
116
+ 2024-04-10 16:16:34,710 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
117
+ 2024-04-10 16:16:34,712 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
118
+ 2024-04-10 16:16:34,713 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
119
+ 2024-04-10 16:16:59,060 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
120
+ 2024-04-10 16:16:59,062 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
121
+ 2024-04-10 16:16:59,062 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
122
+ 2024-04-10 16:17:05,777 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
123
+ 2024-04-10 16:17:05,828 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
124
+ 2024-04-10 16:17:05,828 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
125
+ 2024-04-10 16:17:07,955 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
126
+ 2024-04-10 16:17:07,961 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
127
+ 2024-04-10 16:17:07,961 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
128
+ 2024-04-10 16:17:09,861 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
129
+ 2024-04-10 16:17:09,867 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
130
+ 2024-04-10 16:17:09,867 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
131
+ 2024-04-10 16:17:12,727 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
132
+ 2024-04-10 16:17:12,738 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
133
+ 2024-04-10 16:17:12,738 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
134
+ 2024-04-10 16:17:13,820 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
135
+ 2024-04-10 16:17:13,824 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
136
+ 2024-04-10 16:17:13,825 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
137
+ 2024-04-10 16:17:14,812 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
138
+ 2024-04-10 16:17:15,659 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
139
+ 2024-04-10 16:17:15,659 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
140
+ 2024-04-10 16:17:17,681 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
141
+ 2024-04-10 16:17:17,704 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
142
+ 2024-04-10 16:17:17,704 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
143
+ 2024-04-10 16:17:18,381 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
144
+ 2024-04-10 16:17:18,392 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
145
+ 2024-04-10 16:17:18,392 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
146
+ 2024-04-10 16:17:19,694 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
147
+ 2024-04-10 16:17:19,718 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
148
+ 2024-04-10 16:17:19,718 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
149
+ 2024-04-10 16:17:20,601 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
150
+ 2024-04-10 16:17:20,603 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
151
+ 2024-04-10 16:17:20,603 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
152
+ 2024-04-10 16:17:21,870 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
153
+ 2024-04-10 16:17:23,042 INFO MainThread:34 [wandb_run.py:_config_callback():1347] config_cb None None {'return_dict': True, 'output_hidden_states': True, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['BertModel'], 'finetuning_task': None, 'id2label': {0: 'Documentation Replication on Other Examples', 1: 'Requesting (Additional) Resources', 2: 'Inadequate Examples', 3: 'Documentation Ambiguity', 4: 'Documentation Completeness', 5: 'Lack of Alternative Solutions/Documentation', 6: 'Documentation Replicability'}, 'label2id': {'Documentation Replication on Other Examples': 0, 'Requesting (Additional) Resources': 1, 'Inadequate Examples': 2, 'Documentation Ambiguity': 3, 'Documentation Completeness': 4, 'Lack of Alternative Solutions/Documentation': 5, 'Documentation Replicability': 6}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': 0, 'eos_token_id': None, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'lanwuwei/BERTOverflow_stackoverflow_github', 'transformers_version': '4.39.3', '_num_labels': 51, 'model_type': 'bert', 'output_past': True, 'vocab_size': 82000, 'hidden_size': 768, 'num_hidden_layers': 12, 'num_attention_heads': 12, 'hidden_act': 'gelu', 'intermediate_size': 3072, 'hidden_dropout_prob': 0.1, 'attention_probs_dropout_prob': 0.1, 'max_position_embeddings': 512, 'type_vocab_size': 2, 'initializer_range': 0.02, 'layer_norm_eps': 1e-12, 'position_embedding_type': 'absolute', 'use_cache': True, 'classifier_dropout': None, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'epoch', 'prediction_loss_only': False, 'per_device_train_batch_size': 16, 'per_device_eval_batch_size': 16, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 1e-05, 'weight_decay': 0.01, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 5, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Apr10_16-17-22_ae604480b8c9', 'logging_strategy': 'epoch', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'epoch', 'save_steps': 500, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': True, 'metric_for_best_model': 'loss', 'greater_is_better': False, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None}
wandb/run-20240410_155650-albqrbhg/files/conda-environment.yaml ADDED
File without changes
wandb/run-20240410_155650-albqrbhg/files/config.yaml ADDED
@@ -0,0 +1,683 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wandb_version: 1
2
+
3
+ _wandb:
4
+ desc: null
5
+ value:
6
+ python_version: 3.10.13
7
+ cli_version: 0.16.5
8
+ framework: huggingface
9
+ huggingface_version: 4.39.3
10
+ is_jupyter_run: true
11
+ is_kaggle_kernel: true
12
+ start_time: 1712764610.0
13
+ t:
14
+ 1:
15
+ - 1
16
+ - 2
17
+ - 3
18
+ - 5
19
+ - 11
20
+ - 12
21
+ - 49
22
+ - 51
23
+ - 53
24
+ - 55
25
+ - 71
26
+ - 75
27
+ - 100
28
+ - 105
29
+ 2:
30
+ - 1
31
+ - 2
32
+ - 3
33
+ - 5
34
+ - 11
35
+ - 12
36
+ - 49
37
+ - 51
38
+ - 53
39
+ - 55
40
+ - 71
41
+ - 75
42
+ - 100
43
+ - 105
44
+ 3:
45
+ - 7
46
+ - 23
47
+ 4: 3.10.13
48
+ 5: 0.16.5
49
+ 6: 4.39.3
50
+ 8:
51
+ - 1
52
+ - 2
53
+ - 5
54
+ 9:
55
+ 1: transformers_trainer
56
+ 13: linux-x86_64
57
+ m:
58
+ - 1: train/global_step
59
+ 6:
60
+ - 3
61
+ - 1: train/loss
62
+ 5: 1
63
+ 6:
64
+ - 1
65
+ - 1: train/grad_norm
66
+ 5: 1
67
+ 6:
68
+ - 1
69
+ - 1: train/learning_rate
70
+ 5: 1
71
+ 6:
72
+ - 1
73
+ - 1: train/epoch
74
+ 5: 1
75
+ 6:
76
+ - 1
77
+ return_dict:
78
+ desc: null
79
+ value: true
80
+ output_hidden_states:
81
+ desc: null
82
+ value: true
83
+ output_attentions:
84
+ desc: null
85
+ value: false
86
+ torchscript:
87
+ desc: null
88
+ value: false
89
+ torch_dtype:
90
+ desc: null
91
+ value: null
92
+ use_bfloat16:
93
+ desc: null
94
+ value: false
95
+ tf_legacy_loss:
96
+ desc: null
97
+ value: false
98
+ pruned_heads:
99
+ desc: null
100
+ value: {}
101
+ tie_word_embeddings:
102
+ desc: null
103
+ value: true
104
+ chunk_size_feed_forward:
105
+ desc: null
106
+ value: 0
107
+ is_encoder_decoder:
108
+ desc: null
109
+ value: false
110
+ is_decoder:
111
+ desc: null
112
+ value: false
113
+ cross_attention_hidden_size:
114
+ desc: null
115
+ value: null
116
+ add_cross_attention:
117
+ desc: null
118
+ value: false
119
+ tie_encoder_decoder:
120
+ desc: null
121
+ value: false
122
+ max_length:
123
+ desc: null
124
+ value: 20
125
+ min_length:
126
+ desc: null
127
+ value: 0
128
+ do_sample:
129
+ desc: null
130
+ value: false
131
+ early_stopping:
132
+ desc: null
133
+ value: false
134
+ num_beams:
135
+ desc: null
136
+ value: 1
137
+ num_beam_groups:
138
+ desc: null
139
+ value: 1
140
+ diversity_penalty:
141
+ desc: null
142
+ value: 0.0
143
+ temperature:
144
+ desc: null
145
+ value: 1.0
146
+ top_k:
147
+ desc: null
148
+ value: 50
149
+ top_p:
150
+ desc: null
151
+ value: 1.0
152
+ typical_p:
153
+ desc: null
154
+ value: 1.0
155
+ repetition_penalty:
156
+ desc: null
157
+ value: 1.0
158
+ length_penalty:
159
+ desc: null
160
+ value: 1.0
161
+ no_repeat_ngram_size:
162
+ desc: null
163
+ value: 0
164
+ encoder_no_repeat_ngram_size:
165
+ desc: null
166
+ value: 0
167
+ bad_words_ids:
168
+ desc: null
169
+ value: null
170
+ num_return_sequences:
171
+ desc: null
172
+ value: 1
173
+ output_scores:
174
+ desc: null
175
+ value: false
176
+ return_dict_in_generate:
177
+ desc: null
178
+ value: false
179
+ forced_bos_token_id:
180
+ desc: null
181
+ value: null
182
+ forced_eos_token_id:
183
+ desc: null
184
+ value: null
185
+ remove_invalid_values:
186
+ desc: null
187
+ value: false
188
+ exponential_decay_length_penalty:
189
+ desc: null
190
+ value: null
191
+ suppress_tokens:
192
+ desc: null
193
+ value: null
194
+ begin_suppress_tokens:
195
+ desc: null
196
+ value: null
197
+ architectures:
198
+ desc: null
199
+ value:
200
+ - BertModel
201
+ finetuning_task:
202
+ desc: null
203
+ value: null
204
+ id2label:
205
+ desc: null
206
+ value:
207
+ '0': Documentation Replication on Other Examples
208
+ '1': Requesting (Additional) Resources
209
+ '2': Inadequate Examples
210
+ '3': Documentation Ambiguity
211
+ '4': Documentation Completeness
212
+ '5': Lack of Alternative Solutions/Documentation
213
+ '6': Documentation Replicability
214
+ label2id:
215
+ desc: null
216
+ value:
217
+ Documentation Replication on Other Examples: 0
218
+ Requesting (Additional) Resources: 1
219
+ Inadequate Examples: 2
220
+ Documentation Ambiguity: 3
221
+ Documentation Completeness: 4
222
+ Lack of Alternative Solutions/Documentation: 5
223
+ Documentation Replicability: 6
224
+ tokenizer_class:
225
+ desc: null
226
+ value: null
227
+ prefix:
228
+ desc: null
229
+ value: null
230
+ bos_token_id:
231
+ desc: null
232
+ value: null
233
+ pad_token_id:
234
+ desc: null
235
+ value: 0
236
+ eos_token_id:
237
+ desc: null
238
+ value: null
239
+ sep_token_id:
240
+ desc: null
241
+ value: null
242
+ decoder_start_token_id:
243
+ desc: null
244
+ value: null
245
+ task_specific_params:
246
+ desc: null
247
+ value: null
248
+ problem_type:
249
+ desc: null
250
+ value: null
251
+ _name_or_path:
252
+ desc: null
253
+ value: lanwuwei/BERTOverflow_stackoverflow_github
254
+ transformers_version:
255
+ desc: null
256
+ value: 4.39.3
257
+ _num_labels:
258
+ desc: null
259
+ value: 51
260
+ model_type:
261
+ desc: null
262
+ value: bert
263
+ output_past:
264
+ desc: null
265
+ value: true
266
+ vocab_size:
267
+ desc: null
268
+ value: 82000
269
+ hidden_size:
270
+ desc: null
271
+ value: 768
272
+ num_hidden_layers:
273
+ desc: null
274
+ value: 12
275
+ num_attention_heads:
276
+ desc: null
277
+ value: 12
278
+ hidden_act:
279
+ desc: null
280
+ value: gelu
281
+ intermediate_size:
282
+ desc: null
283
+ value: 3072
284
+ hidden_dropout_prob:
285
+ desc: null
286
+ value: 0.1
287
+ attention_probs_dropout_prob:
288
+ desc: null
289
+ value: 0.1
290
+ max_position_embeddings:
291
+ desc: null
292
+ value: 512
293
+ type_vocab_size:
294
+ desc: null
295
+ value: 2
296
+ initializer_range:
297
+ desc: null
298
+ value: 0.02
299
+ layer_norm_eps:
300
+ desc: null
301
+ value: 1.0e-12
302
+ position_embedding_type:
303
+ desc: null
304
+ value: absolute
305
+ use_cache:
306
+ desc: null
307
+ value: true
308
+ classifier_dropout:
309
+ desc: null
310
+ value: null
311
+ output_dir:
312
+ desc: null
313
+ value: /kaggle/working/
314
+ overwrite_output_dir:
315
+ desc: null
316
+ value: false
317
+ do_train:
318
+ desc: null
319
+ value: false
320
+ do_eval:
321
+ desc: null
322
+ value: true
323
+ do_predict:
324
+ desc: null
325
+ value: false
326
+ evaluation_strategy:
327
+ desc: null
328
+ value: epoch
329
+ prediction_loss_only:
330
+ desc: null
331
+ value: false
332
+ per_device_train_batch_size:
333
+ desc: null
334
+ value: 16
335
+ per_device_eval_batch_size:
336
+ desc: null
337
+ value: 16
338
+ per_gpu_train_batch_size:
339
+ desc: null
340
+ value: null
341
+ per_gpu_eval_batch_size:
342
+ desc: null
343
+ value: null
344
+ gradient_accumulation_steps:
345
+ desc: null
346
+ value: 1
347
+ eval_accumulation_steps:
348
+ desc: null
349
+ value: null
350
+ eval_delay:
351
+ desc: null
352
+ value: 0
353
+ learning_rate:
354
+ desc: null
355
+ value: 1.0e-05
356
+ weight_decay:
357
+ desc: null
358
+ value: 0.01
359
+ adam_beta1:
360
+ desc: null
361
+ value: 0.9
362
+ adam_beta2:
363
+ desc: null
364
+ value: 0.999
365
+ adam_epsilon:
366
+ desc: null
367
+ value: 1.0e-08
368
+ max_grad_norm:
369
+ desc: null
370
+ value: 1.0
371
+ num_train_epochs:
372
+ desc: null
373
+ value: 5
374
+ max_steps:
375
+ desc: null
376
+ value: -1
377
+ lr_scheduler_type:
378
+ desc: null
379
+ value: linear
380
+ lr_scheduler_kwargs:
381
+ desc: null
382
+ value: {}
383
+ warmup_ratio:
384
+ desc: null
385
+ value: 0.0
386
+ warmup_steps:
387
+ desc: null
388
+ value: 0
389
+ log_level:
390
+ desc: null
391
+ value: passive
392
+ log_level_replica:
393
+ desc: null
394
+ value: warning
395
+ log_on_each_node:
396
+ desc: null
397
+ value: true
398
+ logging_dir:
399
+ desc: null
400
+ value: /kaggle/working/runs/Apr10_16-17-22_ae604480b8c9
401
+ logging_strategy:
402
+ desc: null
403
+ value: epoch
404
+ logging_first_step:
405
+ desc: null
406
+ value: false
407
+ logging_steps:
408
+ desc: null
409
+ value: 500
410
+ logging_nan_inf_filter:
411
+ desc: null
412
+ value: true
413
+ save_strategy:
414
+ desc: null
415
+ value: epoch
416
+ save_steps:
417
+ desc: null
418
+ value: 500
419
+ save_total_limit:
420
+ desc: null
421
+ value: null
422
+ save_safetensors:
423
+ desc: null
424
+ value: true
425
+ save_on_each_node:
426
+ desc: null
427
+ value: false
428
+ save_only_model:
429
+ desc: null
430
+ value: false
431
+ no_cuda:
432
+ desc: null
433
+ value: false
434
+ use_cpu:
435
+ desc: null
436
+ value: false
437
+ use_mps_device:
438
+ desc: null
439
+ value: false
440
+ seed:
441
+ desc: null
442
+ value: 42
443
+ data_seed:
444
+ desc: null
445
+ value: null
446
+ jit_mode_eval:
447
+ desc: null
448
+ value: false
449
+ use_ipex:
450
+ desc: null
451
+ value: false
452
+ bf16:
453
+ desc: null
454
+ value: false
455
+ fp16:
456
+ desc: null
457
+ value: true
458
+ fp16_opt_level:
459
+ desc: null
460
+ value: O1
461
+ half_precision_backend:
462
+ desc: null
463
+ value: auto
464
+ bf16_full_eval:
465
+ desc: null
466
+ value: false
467
+ fp16_full_eval:
468
+ desc: null
469
+ value: false
470
+ tf32:
471
+ desc: null
472
+ value: null
473
+ local_rank:
474
+ desc: null
475
+ value: 0
476
+ ddp_backend:
477
+ desc: null
478
+ value: null
479
+ tpu_num_cores:
480
+ desc: null
481
+ value: null
482
+ tpu_metrics_debug:
483
+ desc: null
484
+ value: false
485
+ debug:
486
+ desc: null
487
+ value: []
488
+ dataloader_drop_last:
489
+ desc: null
490
+ value: false
491
+ eval_steps:
492
+ desc: null
493
+ value: null
494
+ dataloader_num_workers:
495
+ desc: null
496
+ value: 0
497
+ dataloader_prefetch_factor:
498
+ desc: null
499
+ value: null
500
+ past_index:
501
+ desc: null
502
+ value: -1
503
+ run_name:
504
+ desc: null
505
+ value: /kaggle/working/
506
+ disable_tqdm:
507
+ desc: null
508
+ value: false
509
+ remove_unused_columns:
510
+ desc: null
511
+ value: true
512
+ label_names:
513
+ desc: null
514
+ value: null
515
+ load_best_model_at_end:
516
+ desc: null
517
+ value: true
518
+ metric_for_best_model:
519
+ desc: null
520
+ value: loss
521
+ greater_is_better:
522
+ desc: null
523
+ value: false
524
+ ignore_data_skip:
525
+ desc: null
526
+ value: false
527
+ fsdp:
528
+ desc: null
529
+ value: []
530
+ fsdp_min_num_params:
531
+ desc: null
532
+ value: 0
533
+ fsdp_config:
534
+ desc: null
535
+ value:
536
+ min_num_params: 0
537
+ xla: false
538
+ xla_fsdp_v2: false
539
+ xla_fsdp_grad_ckpt: false
540
+ fsdp_transformer_layer_cls_to_wrap:
541
+ desc: null
542
+ value: null
543
+ accelerator_config:
544
+ desc: null
545
+ value:
546
+ split_batches: false
547
+ dispatch_batches: null
548
+ even_batches: true
549
+ use_seedable_sampler: true
550
+ deepspeed:
551
+ desc: null
552
+ value: null
553
+ label_smoothing_factor:
554
+ desc: null
555
+ value: 0.0
556
+ optim:
557
+ desc: null
558
+ value: adamw_torch
559
+ optim_args:
560
+ desc: null
561
+ value: null
562
+ adafactor:
563
+ desc: null
564
+ value: false
565
+ group_by_length:
566
+ desc: null
567
+ value: false
568
+ length_column_name:
569
+ desc: null
570
+ value: length
571
+ report_to:
572
+ desc: null
573
+ value:
574
+ - tensorboard
575
+ - wandb
576
+ ddp_find_unused_parameters:
577
+ desc: null
578
+ value: null
579
+ ddp_bucket_cap_mb:
580
+ desc: null
581
+ value: null
582
+ ddp_broadcast_buffers:
583
+ desc: null
584
+ value: null
585
+ dataloader_pin_memory:
586
+ desc: null
587
+ value: true
588
+ dataloader_persistent_workers:
589
+ desc: null
590
+ value: false
591
+ skip_memory_metrics:
592
+ desc: null
593
+ value: true
594
+ use_legacy_prediction_loop:
595
+ desc: null
596
+ value: false
597
+ push_to_hub:
598
+ desc: null
599
+ value: true
600
+ resume_from_checkpoint:
601
+ desc: null
602
+ value: null
603
+ hub_model_id:
604
+ desc: null
605
+ value: null
606
+ hub_strategy:
607
+ desc: null
608
+ value: every_save
609
+ hub_token:
610
+ desc: null
611
+ value: <HUB_TOKEN>
612
+ hub_private_repo:
613
+ desc: null
614
+ value: false
615
+ hub_always_push:
616
+ desc: null
617
+ value: false
618
+ gradient_checkpointing:
619
+ desc: null
620
+ value: false
621
+ gradient_checkpointing_kwargs:
622
+ desc: null
623
+ value: null
624
+ include_inputs_for_metrics:
625
+ desc: null
626
+ value: false
627
+ fp16_backend:
628
+ desc: null
629
+ value: auto
630
+ push_to_hub_model_id:
631
+ desc: null
632
+ value: null
633
+ push_to_hub_organization:
634
+ desc: null
635
+ value: null
636
+ push_to_hub_token:
637
+ desc: null
638
+ value: <PUSH_TO_HUB_TOKEN>
639
+ mp_parameters:
640
+ desc: null
641
+ value: ''
642
+ auto_find_batch_size:
643
+ desc: null
644
+ value: false
645
+ full_determinism:
646
+ desc: null
647
+ value: false
648
+ torchdynamo:
649
+ desc: null
650
+ value: null
651
+ ray_scope:
652
+ desc: null
653
+ value: last
654
+ ddp_timeout:
655
+ desc: null
656
+ value: 1800
657
+ torch_compile:
658
+ desc: null
659
+ value: false
660
+ torch_compile_backend:
661
+ desc: null
662
+ value: null
663
+ torch_compile_mode:
664
+ desc: null
665
+ value: null
666
+ dispatch_batches:
667
+ desc: null
668
+ value: null
669
+ split_batches:
670
+ desc: null
671
+ value: null
672
+ include_tokens_per_second:
673
+ desc: null
674
+ value: false
675
+ include_num_input_tokens_seen:
676
+ desc: null
677
+ value: false
678
+ neftune_noise_alpha:
679
+ desc: null
680
+ value: null
681
+ optim_target_modules:
682
+ desc: null
683
+ value: null
wandb/run-20240410_155650-albqrbhg/files/output.log ADDED
@@ -0,0 +1,87 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Some weights of BertForSequenceClassification were not initialized from the model checkpoint at lanwuwei/BERTOverflow_stackoverflow_github and are newly initialized: ['classifier.bias', 'classifier.weight']
2
+ You should probably TRAIN this model on a down-stream task to be able to use it for predictions and inference.
3
+ /opt/conda/lib/python3.10/site-packages/accelerate/accelerator.py:432: FutureWarning: Passing the following arguments to `Accelerator` is deprecated and will be removed in version 1.0 of Accelerate: dict_keys(['dispatch_batches', 'split_batches', 'even_batches', 'use_seedable_sampler']). Please pass an `accelerate.DataLoaderConfiguration` instead:
4
+ dataloader_config = DataLoaderConfiguration(dispatch_batches=None, split_batches=False, even_batches=True, use_seedable_sampler=True)
5
+ warnings.warn(
6
+ Some weights of BertForSequenceClassification were not initialized from the model checkpoint at lanwuwei/BERTOverflow_stackoverflow_github and are newly initialized: ['classifier.bias', 'classifier.weight']
7
+ You should probably TRAIN this model on a down-stream task to be able to use it for predictions and inference.
8
+ {'Documentation Replication on Other Examples': 0, 'Requesting (Additional) Resources': 1, 'Inadequate Examples': 2, 'Documentation Ambiguity': 3, 'Documentation Completeness': 4, 'Lack of Alternative Solutions/Documentation': 5, 'Documentation Replicability': 6}
9
+ label
10
+ Documentation Replication on Other Examples 232
11
+ Documentation Ambiguity 211
12
+ Requesting (Additional) Resources 162
13
+ Lack of Alternative Solutions/Documentation 154
14
+ Documentation Replicability 139
15
+ Documentation Completeness 104
16
+ Inadequate Examples 83
17
+ Name: count, dtype: int64
18
+ label
19
+ 0 232
20
+ 3 211
21
+ 1 162
22
+ 5 154
23
+ 6 139
24
+ 4 104
25
+ 2 83
26
+ Name: count, dtype: int64
27
+ /opt/conda/lib/python3.10/site-packages/ipywidgets/widgets/widget.py:503: DeprecationWarning: The `ipykernel.comm.Comm` class has been deprecated. Please use the `comm` module instead.For creating comms, use the function `from comm import create_comm`.
28
+ self.comm = Comm(**args)
29
+ {'Documentation Replication on Other Examples': 0, 'Requesting (Additional) Resources': 1, 'Inadequate Examples': 2, 'Documentation Ambiguity': 3, 'Documentation Completeness': 4, 'Lack of Alternative Solutions/Documentation': 5, 'Documentation Replicability': 6}
30
+ label
31
+ Documentation Replication on Other Examples 232
32
+ Documentation Ambiguity 211
33
+ Requesting (Additional) Resources 162
34
+ Lack of Alternative Solutions/Documentation 154
35
+ Documentation Replicability 139
36
+ Documentation Completeness 104
37
+ Inadequate Examples 83
38
+ Name: count, dtype: int64
39
+ label
40
+ 0 232
41
+ 3 211
42
+ 1 162
43
+ 5 154
44
+ 6 139
45
+ 4 104
46
+ 2 83
47
+ Name: count, dtype: int64
48
+ /opt/conda/lib/python3.10/site-packages/ipywidgets/widgets/widget.py:503: DeprecationWarning: The `ipykernel.comm.Comm` class has been deprecated. Please use the `comm` module instead.For creating comms, use the function `from comm import create_comm`.
49
+ self.comm = Comm(**args)
50
+ Class Weights: tensor([0.7858, 0.8505, 0.9236, 0.8059, 0.9046, 0.8579, 0.8717],
51
+ device='cuda:0')
52
+ Some weights of BertForSequenceClassification were not initialized from the model checkpoint at lanwuwei/BERTOverflow_stackoverflow_github and are newly initialized: ['classifier.bias', 'classifier.weight']
53
+ You should probably TRAIN this model on a down-stream task to be able to use it for predictions and inference.
54
+ Some weights of BertForSequenceClassification were not initialized from the model checkpoint at lanwuwei/BERTOverflow_stackoverflow_github and are newly initialized: ['classifier.bias', 'classifier.weight']
55
+ You should probably TRAIN this model on a down-stream task to be able to use it for predictions and inference.
56
+ /opt/conda/lib/python3.10/site-packages/accelerate/accelerator.py:432: FutureWarning: Passing the following arguments to `Accelerator` is deprecated and will be removed in version 1.0 of Accelerate: dict_keys(['dispatch_batches', 'split_batches', 'even_batches', 'use_seedable_sampler']). Please pass an `accelerate.DataLoaderConfiguration` instead:
57
+ dataloader_config = DataLoaderConfiguration(dispatch_batches=None, split_batches=False, even_batches=True, use_seedable_sampler=True)
58
+ warnings.warn(
59
+ {'Documentation Replication on Other Examples': 0, 'Requesting (Additional) Resources': 1, 'Inadequate Examples': 2, 'Documentation Ambiguity': 3, 'Documentation Completeness': 4, 'Lack of Alternative Solutions/Documentation': 5, 'Documentation Replicability': 6}
60
+ label
61
+ Documentation Replication on Other Examples 232
62
+ Documentation Ambiguity 211
63
+ Requesting (Additional) Resources 162
64
+ Lack of Alternative Solutions/Documentation 154
65
+ Documentation Replicability 139
66
+ Documentation Completeness 104
67
+ Inadequate Examples 83
68
+ Name: count, dtype: int64
69
+ label
70
+ 0 232
71
+ 3 211
72
+ 1 162
73
+ 5 154
74
+ 6 139
75
+ 4 104
76
+ 2 83
77
+ Name: count, dtype: int64
78
+ /opt/conda/lib/python3.10/site-packages/ipywidgets/widgets/widget.py:503: DeprecationWarning: The `ipykernel.comm.Comm` class has been deprecated. Please use the `comm` module instead.For creating comms, use the function `from comm import create_comm`.
79
+ self.comm = Comm(**args)
80
+ Class Weights: tensor([0.7858, 0.8505, 0.9236, 0.8059, 0.9046, 0.8579, 0.8717],
81
+ device='cuda:0')
82
+ Some weights of BertForSequenceClassification were not initialized from the model checkpoint at lanwuwei/BERTOverflow_stackoverflow_github and are newly initialized: ['classifier.bias', 'classifier.weight']
83
+ You should probably TRAIN this model on a down-stream task to be able to use it for predictions and inference.
84
+ /opt/conda/lib/python3.10/site-packages/accelerate/accelerator.py:432: FutureWarning: Passing the following arguments to `Accelerator` is deprecated and will be removed in version 1.0 of Accelerate: dict_keys(['dispatch_batches', 'split_batches', 'even_batches', 'use_seedable_sampler']). Please pass an `accelerate.DataLoaderConfiguration` instead:
85
+ dataloader_config = DataLoaderConfiguration(dispatch_batches=None, split_batches=False, even_batches=True, use_seedable_sampler=True)
86
+ warnings.warn(
87
+ /opt/conda/lib/python3.10/site-packages/sklearn/metrics/_classification.py:1344: UndefinedMetricWarning: Precision and F-score are ill-defined and being set to 0.0 in labels with no predicted samples. Use `zero_division` parameter to control this behavior.
wandb/run-20240410_155650-albqrbhg/files/requirements.txt ADDED
@@ -0,0 +1,863 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Babel==2.14.0
2
+ Boruta==0.3
3
+ Brotli==1.0.9
4
+ CVXcanon==0.1.2
5
+ Cartopy==0.22.0
6
+ Cython==3.0.8
7
+ Deprecated==1.2.14
8
+ Farama-Notifications==0.0.4
9
+ Flask==3.0.2
10
+ Geohash==1.0
11
+ GitPython==3.1.41
12
+ ImageHash==4.3.1
13
+ Janome==0.5.0
14
+ Jinja2==3.1.2
15
+ LunarCalendar==0.0.9
16
+ Mako==1.3.2
17
+ Markdown==3.5.2
18
+ MarkupSafe==2.1.3
19
+ MarkupSafe==2.1.5
20
+ Pillow==9.5.0
21
+ PuLP==2.8.0
22
+ PyArabic==0.6.15
23
+ PyJWT==2.8.0
24
+ PyMeeus==0.5.12
25
+ PySocks==1.7.1
26
+ PyUpSet==0.1.1.post7
27
+ PyWavelets==1.5.0
28
+ PyYAML==6.0.1
29
+ Pygments==2.17.2
30
+ Pympler==1.0.1
31
+ QtPy==2.4.1
32
+ Rtree==1.2.0
33
+ SQLAlchemy==2.0.25
34
+ SecretStorage==3.3.3
35
+ Send2Trash==1.8.2
36
+ Shapely==1.8.5.post1
37
+ Shimmy==1.3.0
38
+ SimpleITK==2.3.1
39
+ TPOT==0.12.1
40
+ Theano-PyMC==1.1.2
41
+ Theano==1.0.5
42
+ Wand==0.6.13
43
+ Werkzeug==3.0.2
44
+ absl-py==1.4.0
45
+ accelerate==0.28.0
46
+ access==1.1.9
47
+ affine==2.4.0
48
+ aiobotocore==2.12.2
49
+ aiofiles==22.1.0
50
+ aiohttp-cors==0.7.0
51
+ aiohttp==3.9.1
52
+ aioitertools==0.11.0
53
+ aiorwlock==1.3.0
54
+ aiosignal==1.3.1
55
+ aiosqlite==0.19.0
56
+ albumentations==1.4.0
57
+ alembic==1.13.1
58
+ altair==5.3.0
59
+ annotated-types==0.6.0
60
+ annoy==1.17.3
61
+ anyio==4.2.0
62
+ apache-beam==2.46.0
63
+ aplus==0.11.0
64
+ appdirs==1.4.4
65
+ archspec==0.2.3
66
+ argon2-cffi-bindings==21.2.0
67
+ argon2-cffi==23.1.0
68
+ array-record==0.5.0
69
+ arrow==1.3.0
70
+ arviz==0.17.1
71
+ astroid==3.1.0
72
+ astropy-iers-data==0.2024.4.1.0.33.14
73
+ astropy==6.0.1
74
+ asttokens==2.4.1
75
+ astunparse==1.6.3
76
+ async-lru==2.0.4
77
+ async-timeout==4.0.3
78
+ attrs==23.2.0
79
+ audioread==3.0.1
80
+ autopep8==2.0.4
81
+ backoff==2.2.1
82
+ bayesian-optimization==1.4.3
83
+ beatrix_jupyterlab==2023.128.151533
84
+ beautifulsoup4==4.12.2
85
+ blake3==0.2.1
86
+ bleach==6.1.0
87
+ blessed==1.20.0
88
+ blinker==1.7.0
89
+ blis==0.7.10
90
+ blosc2==2.6.0
91
+ bokeh==3.3.4
92
+ boltons==23.1.1
93
+ boto3==1.26.100
94
+ botocore==1.34.51
95
+ bq_helper==0.4.1
96
+ bqplot==0.12.43
97
+ branca==0.7.1
98
+ brewer2mpl==1.4.1
99
+ brotlipy==0.7.0
100
+ cached-property==1.5.2
101
+ cachetools==4.2.4
102
+ cachetools==5.3.2
103
+ catalogue==2.0.10
104
+ catalyst==22.4
105
+ catboost==1.2.3
106
+ category-encoders==2.6.3
107
+ certifi==2024.2.2
108
+ cesium==0.12.1
109
+ cffi==1.16.0
110
+ charset-normalizer==3.3.2
111
+ chex==0.1.86
112
+ cleverhans==4.0.0
113
+ click-plugins==1.1.1
114
+ click==8.1.7
115
+ cligj==0.7.2
116
+ cloud-tpu-client==0.10
117
+ cloud-tpu-profiler==2.4.0
118
+ cloudpathlib==0.16.0
119
+ cloudpickle==2.2.1
120
+ cloudpickle==3.0.0
121
+ cmdstanpy==1.2.2
122
+ colorama==0.4.6
123
+ colorcet==3.1.0
124
+ colorful==0.5.6
125
+ colorlog==6.8.2
126
+ colorlover==0.3.0
127
+ comm==0.2.1
128
+ conda-libmamba-solver==23.7.0
129
+ conda-package-handling==2.2.0
130
+ conda==23.7.4
131
+ conda_package_streaming==0.9.0
132
+ confection==0.1.4
133
+ contextily==1.6.0
134
+ contourpy==1.2.0
135
+ convertdate==2.4.0
136
+ crcmod==1.7
137
+ cryptography==41.0.7
138
+ cuda-python==12.4.0
139
+ cudf==23.8.0
140
+ cufflinks==0.17.3
141
+ cuml==23.8.0
142
+ cupy==13.0.0
143
+ cycler==0.12.1
144
+ cymem==2.0.8
145
+ cytoolz==0.12.3
146
+ daal4py==2024.2.0
147
+ daal==2024.2.0
148
+ dacite==1.8.1
149
+ dask-cuda==23.8.0
150
+ dask-cudf==23.8.0
151
+ dask-expr==1.0.9
152
+ dask==2024.4.0
153
+ dataclasses-json==0.6.4
154
+ dataproc_jupyter_plugin==0.1.66
155
+ datasets==2.18.0
156
+ datashader==0.16.0
157
+ datatile==1.0.3
158
+ db-dtypes==1.2.0
159
+ deap==1.4.1
160
+ debugpy==1.8.0
161
+ decorator==5.1.1
162
+ deepdiff==6.7.1
163
+ defusedxml==0.7.1
164
+ deprecation==2.1.0
165
+ descartes==1.1.0
166
+ dill==0.3.8
167
+ dipy==1.9.0
168
+ distlib==0.3.8
169
+ distributed==2023.7.1
170
+ distro==1.9.0
171
+ dm-tree==0.1.8
172
+ docker-pycreds==0.4.0
173
+ docker==7.0.0
174
+ docopt==0.6.2
175
+ docstring-parser==0.15
176
+ docstring-to-markdown==0.15
177
+ docutils==0.20.1
178
+ earthengine-api==0.1.395
179
+ easydict==1.13
180
+ easyocr==1.7.1
181
+ ecos==2.0.13
182
+ eli5==0.13.0
183
+ emoji==2.11.0
184
+ en-core-web-lg==3.7.1
185
+ en-core-web-sm==3.7.1
186
+ entrypoints==0.4
187
+ ephem==4.1.5
188
+ esda==2.5.1
189
+ essentia==2.1b6.dev1110
190
+ et-xmlfile==1.1.0
191
+ etils==1.6.0
192
+ evaluate==0.4.1
193
+ exceptiongroup==1.2.0
194
+ executing==2.0.1
195
+ explainable-ai-sdk==1.3.3
196
+ fastai==2.7.14
197
+ fastapi==0.108.0
198
+ fastavro==1.9.3
199
+ fastcore==1.5.29
200
+ fastdownload==0.0.7
201
+ fasteners==0.19
202
+ fastjsonschema==2.19.1
203
+ fastprogress==1.0.3
204
+ fastrlock==0.8.2
205
+ fasttext==0.9.2
206
+ feather-format==0.4.1
207
+ featuretools==1.30.0
208
+ filelock==3.13.1
209
+ fiona==1.9.6
210
+ fitter==1.7.0
211
+ flake8==7.0.0
212
+ flashtext==2.7
213
+ flatbuffers==23.5.26
214
+ flax==0.8.2
215
+ folium==0.16.0
216
+ fonttools==4.47.0
217
+ fonttools==4.50.0
218
+ fqdn==1.5.1
219
+ frozendict==2.4.1
220
+ frozenlist==1.4.1
221
+ fsspec==2024.2.0
222
+ fsspec==2024.3.1
223
+ funcy==2.0
224
+ fury==0.10.0
225
+ future==1.0.0
226
+ fuzzywuzzy==0.18.0
227
+ gast==0.5.4
228
+ gatspy==0.3
229
+ gcsfs==2024.2.0
230
+ gensim==4.3.2
231
+ geographiclib==2.0
232
+ geojson==3.1.0
233
+ geopandas==0.14.3
234
+ geoplot==0.5.1
235
+ geopy==2.4.1
236
+ geoviews==1.11.1
237
+ ggplot==0.11.5
238
+ giddy==2.3.5
239
+ gitdb==4.0.11
240
+ google-ai-generativelanguage==0.4.0
241
+ google-api-core==2.11.1
242
+ google-api-core==2.18.0
243
+ google-api-python-client==2.125.0
244
+ google-apitools==0.5.31
245
+ google-auth-httplib2==0.2.0
246
+ google-auth-oauthlib==1.2.0
247
+ google-auth==2.26.1
248
+ google-cloud-aiplatform==0.6.0a1
249
+ google-cloud-artifact-registry==1.10.0
250
+ google-cloud-automl==1.0.1
251
+ google-cloud-bigquery==2.34.4
252
+ google-cloud-bigtable==1.7.3
253
+ google-cloud-core==2.4.1
254
+ google-cloud-datastore==2.19.0
255
+ google-cloud-dlp==3.14.0
256
+ google-cloud-jupyter-config==0.0.5
257
+ google-cloud-language==2.13.3
258
+ google-cloud-monitoring==2.18.0
259
+ google-cloud-pubsub==2.19.0
260
+ google-cloud-pubsublite==1.9.0
261
+ google-cloud-recommendations-ai==0.7.1
262
+ google-cloud-resource-manager==1.11.0
263
+ google-cloud-spanner==3.40.1
264
+ google-cloud-storage==1.44.0
265
+ google-cloud-translate==3.12.1
266
+ google-cloud-videointelligence==2.13.3
267
+ google-cloud-vision==2.8.0
268
+ google-crc32c==1.5.0
269
+ google-generativeai==0.4.1
270
+ google-pasta==0.2.0
271
+ google-resumable-media==2.7.0
272
+ googleapis-common-protos==1.62.0
273
+ gplearn==0.4.2
274
+ gpustat==1.0.0
275
+ gpxpy==1.6.2
276
+ graphviz==0.20.3
277
+ greenlet==3.0.3
278
+ grpc-google-iam-v1==0.12.7
279
+ grpcio-status==1.48.1
280
+ grpcio-status==1.48.2
281
+ grpcio==1.51.1
282
+ grpcio==1.60.0
283
+ gviz-api==1.10.0
284
+ gym-notices==0.0.8
285
+ gym==0.26.2
286
+ gymnasium==0.29.0
287
+ h11==0.14.0
288
+ h2o==3.46.0.1
289
+ h5netcdf==1.3.0
290
+ h5py==3.10.0
291
+ haversine==2.8.1
292
+ hdfs==2.7.3
293
+ hep-ml==0.7.2
294
+ hijri-converter==2.3.1
295
+ hmmlearn==0.3.2
296
+ holidays==0.24
297
+ holoviews==1.18.3
298
+ hpsklearn==0.1.0
299
+ html5lib==1.1
300
+ htmlmin==0.1.12
301
+ httpcore==1.0.5
302
+ httplib2==0.21.0
303
+ httptools==0.6.1
304
+ httpx==0.27.0
305
+ huggingface-hub==0.22.2
306
+ hunspell==0.5.5
307
+ hydra-slayer==0.5.0
308
+ hyperopt==0.2.7
309
+ hypertools==0.8.0
310
+ idna==3.6
311
+ igraph==0.11.4
312
+ imagecodecs==2024.1.1
313
+ imageio==2.33.1
314
+ imbalanced-learn==0.12.2
315
+ imgaug==0.4.0
316
+ importlib-metadata==6.11.0
317
+ importlib-metadata==7.0.1
318
+ importlib-resources==6.1.1
319
+ inequality==1.0.1
320
+ iniconfig==2.0.0
321
+ ipydatawidgets==4.3.5
322
+ ipykernel==6.28.0
323
+ ipyleaflet==0.18.2
324
+ ipympl==0.7.0
325
+ ipython-genutils==0.2.0
326
+ ipython-genutils==0.2.0
327
+ ipython-sql==0.5.0
328
+ ipython==8.20.0
329
+ ipyvolume==0.6.3
330
+ ipyvue==1.10.2
331
+ ipyvuetify==1.9.3
332
+ ipywebrtc==0.6.0
333
+ ipywidgets==7.7.1
334
+ isoduration==20.11.0
335
+ isort==5.13.2
336
+ isoweek==1.3.3
337
+ itsdangerous==2.1.2
338
+ jaraco.classes==3.3.0
339
+ jax-jumpy==1.0.0
340
+ jax==0.4.23
341
+ jaxlib==0.4.23.dev20240116
342
+ jedi==0.19.1
343
+ jeepney==0.8.0
344
+ jieba==0.42.1
345
+ jmespath==1.0.1
346
+ joblib==1.3.2
347
+ json5==0.9.14
348
+ jsonpatch==1.33
349
+ jsonpointer==2.4
350
+ jsonschema-specifications==2023.12.1
351
+ jsonschema==4.20.0
352
+ jupyter-console==6.6.3
353
+ jupyter-events==0.9.0
354
+ jupyter-http-over-ws==0.0.8
355
+ jupyter-lsp==1.5.1
356
+ jupyter-server-mathjax==0.2.6
357
+ jupyter-ydoc==0.2.5
358
+ jupyter_client==7.4.9
359
+ jupyter_client==8.6.0
360
+ jupyter_core==5.7.1
361
+ jupyter_server==2.13.0
362
+ jupyter_server_fileid==0.9.1
363
+ jupyter_server_proxy==4.1.0
364
+ jupyter_server_terminals==0.5.1
365
+ jupyter_server_ydoc==0.8.0
366
+ jupyterlab-lsp==5.1.0
367
+ jupyterlab-widgets==3.0.9
368
+ jupyterlab==4.1.5
369
+ jupyterlab_git==0.44.0
370
+ jupyterlab_pygments==0.3.0
371
+ jupyterlab_server==2.25.2
372
+ jupytext==1.16.0
373
+ kaggle-environments==1.14.3
374
+ kaggle==1.6.8
375
+ kagglehub==0.2.2
376
+ keras-cv==0.8.2
377
+ keras-nlp==0.8.2
378
+ keras-tuner==1.4.6
379
+ keras==3.1.1
380
+ kernels-mixer==0.0.7
381
+ keyring==24.3.0
382
+ keyrings.google-artifactregistry-auth==1.1.2
383
+ kfp-pipeline-spec==0.2.2
384
+ kfp-server-api==2.0.5
385
+ kfp==2.5.0
386
+ kiwisolver==1.4.5
387
+ kmapper==2.0.1
388
+ kmodes==0.12.2
389
+ korean-lunar-calendar==0.3.1
390
+ kornia==0.7.2
391
+ kornia_rs==0.1.3
392
+ kt-legacy==1.0.5
393
+ kubernetes==26.1.0
394
+ langcodes==3.3.0
395
+ langid==1.1.6
396
+ lazy_loader==0.3
397
+ learntools==0.3.4
398
+ leven==1.0.4
399
+ libclang==16.0.6
400
+ libmambapy==1.5.0
401
+ libpysal==4.9.2
402
+ librosa==0.10.1
403
+ lightgbm==4.2.0
404
+ lightning-utilities==0.11.2
405
+ lime==0.2.0.1
406
+ line-profiler==4.1.2
407
+ linkify-it-py==2.0.3
408
+ llvmlite==0.41.1
409
+ llvmlite==0.42.0
410
+ lml==0.1.0
411
+ locket==1.0.0
412
+ loguru==0.7.2
413
+ lxml==5.2.1
414
+ lz4==4.3.3
415
+ mamba==1.5.0
416
+ mapclassify==2.6.1
417
+ markdown-it-py==3.0.0
418
+ marshmallow==3.21.1
419
+ matplotlib-inline==0.1.6
420
+ matplotlib-venn==0.11.10
421
+ matplotlib==3.7.5
422
+ matplotlib==3.8.3
423
+ mccabe==0.7.0
424
+ mdit-py-plugins==0.4.0
425
+ mdurl==0.1.2
426
+ memory-profiler==0.61.0
427
+ menuinst==2.0.1
428
+ mercantile==1.2.1
429
+ mgwr==2.2.1
430
+ missingno==0.5.2
431
+ mistune==0.8.4
432
+ mizani==0.11.1
433
+ ml-dtypes==0.2.0
434
+ mlcrate==0.2.0
435
+ mlens==0.2.3
436
+ mlxtend==0.23.1
437
+ mne==1.6.1
438
+ mnist==0.2.2
439
+ momepy==0.7.0
440
+ more-itertools==10.2.0
441
+ mpld3==0.5.10
442
+ mpmath==1.3.0
443
+ msgpack==1.0.7
444
+ multidict==6.0.4
445
+ multimethod==1.10
446
+ multipledispatch==1.0.0
447
+ multiprocess==0.70.16
448
+ munkres==1.1.4
449
+ murmurhash==1.0.10
450
+ mypy-extensions==1.0.0
451
+ namex==0.0.7
452
+ nb-conda-kernels==2.3.1
453
+ nb_conda==2.2.1
454
+ nbclassic==1.0.0
455
+ nbclient==0.5.13
456
+ nbconvert==6.4.5
457
+ nbdime==3.2.0
458
+ nbformat==5.9.2
459
+ ndindex==1.8
460
+ nest-asyncio==1.5.8
461
+ networkx==3.2.1
462
+ nibabel==5.2.1
463
+ nilearn==0.10.3
464
+ ninja==1.11.1.1
465
+ nltk==3.2.4
466
+ nose==1.3.7
467
+ notebook==6.5.4
468
+ notebook==6.5.6
469
+ notebook_executor==0.2
470
+ notebook_shim==0.2.3
471
+ numba==0.58.1
472
+ numba==0.59.1
473
+ numexpr==2.10.0
474
+ numpy==1.26.4
475
+ nvidia-ml-py==11.495.46
476
+ nvtx==0.2.10
477
+ oauth2client==4.1.3
478
+ oauthlib==3.2.2
479
+ objsize==0.6.1
480
+ odfpy==1.4.1
481
+ olefile==0.47
482
+ onnx==1.16.0
483
+ opencensus-context==0.1.3
484
+ opencensus==0.11.4
485
+ opencv-contrib-python==4.9.0.80
486
+ opencv-python-headless==4.9.0.80
487
+ opencv-python==4.9.0.80
488
+ openpyxl==3.1.2
489
+ openslide-python==1.3.1
490
+ opentelemetry-api==1.22.0
491
+ opentelemetry-exporter-otlp-proto-common==1.22.0
492
+ opentelemetry-exporter-otlp-proto-grpc==1.22.0
493
+ opentelemetry-exporter-otlp-proto-http==1.22.0
494
+ opentelemetry-exporter-otlp==1.22.0
495
+ opentelemetry-proto==1.22.0
496
+ opentelemetry-sdk==1.22.0
497
+ opentelemetry-semantic-conventions==0.43b0
498
+ opt-einsum==3.3.0
499
+ optax==0.2.2
500
+ optree==0.11.0
501
+ optuna==3.6.1
502
+ orbax-checkpoint==0.5.7
503
+ ordered-set==4.1.0
504
+ orjson==3.9.10
505
+ ortools==9.4.1874
506
+ osmnx==1.9.2
507
+ overrides==7.4.0
508
+ packaging==21.3
509
+ pandas-datareader==0.10.0
510
+ pandas-profiling==3.6.6
511
+ pandas-summary==0.2.0
512
+ pandas==2.1.4
513
+ pandas==2.2.1
514
+ pandasql==0.7.3
515
+ pandocfilters==1.5.0
516
+ panel==1.3.8
517
+ papermill==2.5.0
518
+ param==2.1.0
519
+ parso==0.8.3
520
+ partd==1.4.1
521
+ path.py==12.5.0
522
+ path==16.10.0
523
+ pathos==0.3.2
524
+ pathy==0.10.3
525
+ patsy==0.5.6
526
+ pdf2image==1.17.0
527
+ pettingzoo==1.24.0
528
+ pexpect==4.8.0
529
+ pexpect==4.9.0
530
+ phik==0.12.4
531
+ pickleshare==0.7.5
532
+ pillow==10.3.0
533
+ pip==23.3.2
534
+ pkgutil_resolve_name==1.3.10
535
+ platformdirs==4.2.0
536
+ plotly-express==0.4.1
537
+ plotly==5.18.0
538
+ plotnine==0.13.4
539
+ pluggy==1.4.0
540
+ pointpats==2.4.0
541
+ polars==0.20.18
542
+ polyglot==16.7.4
543
+ pooch==1.8.1
544
+ pox==0.3.4
545
+ ppca==0.0.4
546
+ ppft==1.7.6.8
547
+ preprocessing==0.1.13
548
+ preshed==3.0.9
549
+ prettytable==3.9.0
550
+ progressbar2==4.4.2
551
+ prometheus-client==0.19.0
552
+ promise==2.3
553
+ prompt-toolkit==3.0.42
554
+ prompt-toolkit==3.0.43
555
+ prophet==1.1.1
556
+ proto-plus==1.23.0
557
+ protobuf==3.20.3
558
+ protobuf==4.21.12
559
+ psutil==5.9.3
560
+ psutil==5.9.7
561
+ ptyprocess==0.7.0
562
+ pudb==2024.1
563
+ pure-eval==0.2.2
564
+ py-cpuinfo==9.0.0
565
+ py-spy==0.3.14
566
+ py4j==0.10.9.7
567
+ pyLDAvis==3.4.1
568
+ pyOpenSSL==23.3.0
569
+ pyaml==23.12.0
570
+ pyarrow-hotfix==0.6
571
+ pyarrow==15.0.2
572
+ pyasn1-modules==0.3.0
573
+ pyasn1==0.5.1
574
+ pybind11==2.12.0
575
+ pyclipper==1.3.0.post5
576
+ pycodestyle==2.11.1
577
+ pycosat==0.6.6
578
+ pycparser==2.21
579
+ pycryptodome==3.20.0
580
+ pyct==0.5.0
581
+ pycuda==2024.1
582
+ pydantic==2.5.3
583
+ pydantic==2.6.4
584
+ pydantic_core==2.14.6
585
+ pydantic_core==2.16.3
586
+ pydegensac==0.1.2
587
+ pydicom==2.4.4
588
+ pydocstyle==6.3.0
589
+ pydot==1.4.2
590
+ pydub==0.25.1
591
+ pyemd==1.0.0
592
+ pyerfa==2.0.1.1
593
+ pyexcel-io==0.6.6
594
+ pyexcel-ods==0.6.0
595
+ pyflakes==3.2.0
596
+ pygltflib==1.16.2
597
+ pykalman==0.9.7
598
+ pylibraft==23.8.0
599
+ pylint==3.1.0
600
+ pymc3==3.11.4
601
+ pymongo==3.13.0
602
+ pynndescent==0.5.12
603
+ pynvml==11.4.1
604
+ pynvrtc==9.2
605
+ pyparsing==3.1.1
606
+ pyparsing==3.1.2
607
+ pypdf==4.1.0
608
+ pyproj==3.6.1
609
+ pysal==24.1
610
+ pyshp==2.3.1
611
+ pytesseract==0.3.10
612
+ pytest==8.1.1
613
+ python-bidi==0.4.2
614
+ python-dateutil==2.9.0.post0
615
+ python-dotenv==1.0.0
616
+ python-json-logger==2.0.7
617
+ python-louvain==0.16
618
+ python-lsp-jsonrpc==1.1.2
619
+ python-lsp-server==1.11.0
620
+ python-slugify==8.0.4
621
+ python-utils==3.8.2
622
+ pythreejs==2.4.2
623
+ pytoolconfig==1.3.1
624
+ pytools==2024.1.1
625
+ pytorch-ignite==0.5.0.post2
626
+ pytorch-lightning==2.2.1
627
+ pytz==2023.3.post1
628
+ pytz==2024.1
629
+ pyu2f==0.1.5
630
+ pyviz_comms==3.0.2
631
+ pyzmq==24.0.1
632
+ pyzmq==25.1.2
633
+ qgrid==1.3.1
634
+ qtconsole==5.5.1
635
+ quantecon==0.7.2
636
+ qudida==0.0.4
637
+ raft-dask==23.8.0
638
+ rasterio==1.3.9
639
+ rasterstats==0.19.0
640
+ ray-cpp==2.9.0
641
+ ray==2.9.0
642
+ referencing==0.32.1
643
+ regex==2023.12.25
644
+ requests-oauthlib==1.3.1
645
+ requests-toolbelt==0.10.1
646
+ requests==2.31.0
647
+ responses==0.18.0
648
+ retrying==1.3.3
649
+ retrying==1.3.4
650
+ rfc3339-validator==0.1.4
651
+ rfc3986-validator==0.1.1
652
+ rgf-python==3.12.0
653
+ rich-click==1.7.4
654
+ rich==13.7.0
655
+ rich==13.7.1
656
+ rmm==23.8.0
657
+ rope==1.13.0
658
+ rpds-py==0.16.2
659
+ rsa==4.9
660
+ ruamel-yaml-conda==0.15.100
661
+ ruamel.yaml.clib==0.2.7
662
+ ruamel.yaml==0.17.40
663
+ s2sphere==0.2.5
664
+ s3fs==2024.2.0
665
+ s3transfer==0.6.2
666
+ safetensors==0.4.2
667
+ scattertext==0.1.19
668
+ scikit-image==0.22.0
669
+ scikit-learn-intelex==2024.2.0
670
+ scikit-learn==1.2.2
671
+ scikit-multilearn==0.2.0
672
+ scikit-optimize==0.10.1
673
+ scikit-plot==0.3.7
674
+ scikit-surprise==1.1.3
675
+ scipy==1.11.4
676
+ scipy==1.12.0
677
+ seaborn==0.12.2
678
+ segment_anything==1.0
679
+ segregation==2.5
680
+ semver==3.0.2
681
+ sentence-transformers==2.6.1
682
+ sentencepiece==0.2.0
683
+ sentry-sdk==1.44.1
684
+ setfit==1.0.3
685
+ setproctitle==1.3.3
686
+ setuptools-git==1.2
687
+ setuptools-scm==8.0.4
688
+ setuptools==69.0.3
689
+ shap==0.44.1
690
+ shapely==2.0.3
691
+ shellingham==1.5.4
692
+ simpervisor==1.0.0
693
+ simplejson==3.19.2
694
+ six==1.16.0
695
+ sklearn-pandas==2.2.0
696
+ slicer==0.0.7
697
+ smart-open==6.4.0
698
+ smmap==5.0.1
699
+ sniffio==1.3.0
700
+ snowballstemmer==2.2.0
701
+ snuggs==1.4.7
702
+ sortedcontainers==2.4.0
703
+ soundfile==0.12.1
704
+ soupsieve==2.5
705
+ soxr==0.3.7
706
+ spacy-legacy==3.0.12
707
+ spacy-loggers==1.0.5
708
+ spacy==3.7.2
709
+ spaghetti==1.7.5.post1
710
+ spectral==0.23.1
711
+ spglm==1.1.0
712
+ sphinx-rtd-theme==0.2.4
713
+ spint==1.0.7
714
+ splot==1.1.5.post1
715
+ spopt==0.6.0
716
+ spreg==1.4.2
717
+ spvcm==0.3.0
718
+ sqlparse==0.4.4
719
+ squarify==0.4.3
720
+ srsly==2.4.8
721
+ stable-baselines3==2.1.0
722
+ stack-data==0.6.2
723
+ stack-data==0.6.3
724
+ stanio==0.5.0
725
+ starlette==0.32.0.post1
726
+ statsmodels==0.14.1
727
+ stemming==1.0.1
728
+ stop-words==2018.7.23
729
+ stopit==1.1.2
730
+ stumpy==1.12.0
731
+ sympy==1.12
732
+ tables==3.9.2
733
+ tabulate==0.9.0
734
+ tangled-up-in-unicode==0.2.0
735
+ tbb==2021.12.0
736
+ tblib==3.0.0
737
+ tenacity==8.2.3
738
+ tensorboard-data-server==0.7.2
739
+ tensorboard-plugin-profile==2.15.0
740
+ tensorboard==2.15.1
741
+ tensorboardX==2.6.2.2
742
+ tensorflow-cloud==0.1.16
743
+ tensorflow-datasets==4.9.4
744
+ tensorflow-decision-forests==1.8.1
745
+ tensorflow-estimator==2.15.0
746
+ tensorflow-hub==0.16.1
747
+ tensorflow-io-gcs-filesystem==0.35.0
748
+ tensorflow-io==0.35.0
749
+ tensorflow-metadata==0.14.0
750
+ tensorflow-probability==0.23.0
751
+ tensorflow-serving-api==2.14.1
752
+ tensorflow-text==2.15.0
753
+ tensorflow-transform==0.14.0
754
+ tensorflow==2.15.0
755
+ tensorstore==0.1.56
756
+ termcolor==2.4.0
757
+ terminado==0.18.0
758
+ testpath==0.6.0
759
+ text-unidecode==1.3
760
+ textblob==0.18.0.post0
761
+ texttable==1.7.0
762
+ tf_keras==2.15.1
763
+ tfp-nightly==0.24.0.dev0
764
+ thinc==8.2.2
765
+ threadpoolctl==3.2.0
766
+ tifffile==2023.12.9
767
+ timm==0.9.16
768
+ tinycss2==1.2.1
769
+ tobler==0.11.2
770
+ tokenizers==0.15.2
771
+ toml==0.10.2
772
+ tomli==2.0.1
773
+ tomlkit==0.12.4
774
+ toolz==0.12.1
775
+ torch==2.1.2
776
+ torchaudio==2.1.2
777
+ torchdata==0.7.1
778
+ torchinfo==1.8.0
779
+ torchmetrics==1.3.2
780
+ torchtext==0.16.2
781
+ torchvision==0.16.2
782
+ tornado==6.3.3
783
+ tqdm==4.66.1
784
+ traceml==1.0.8
785
+ traitlets==5.9.0
786
+ traittypes==0.2.1
787
+ transformers==4.39.3
788
+ treelite-runtime==3.2.0
789
+ treelite==3.2.0
790
+ truststore==0.8.0
791
+ trx-python==0.2.9
792
+ tsfresh==0.20.2
793
+ typeguard==4.1.5
794
+ typer==0.9.0
795
+ typer==0.9.4
796
+ types-python-dateutil==2.8.19.20240106
797
+ typing-inspect==0.9.0
798
+ typing-utils==0.1.0
799
+ typing_extensions==4.9.0
800
+ tzdata==2023.4
801
+ uc-micro-py==1.0.3
802
+ ucx-py==0.33.0
803
+ ujson==5.9.0
804
+ umap-learn==0.5.5
805
+ unicodedata2==15.1.0
806
+ update-checker==0.18.0
807
+ uri-template==1.3.0
808
+ uritemplate==3.0.1
809
+ urllib3==1.26.18
810
+ urllib3==2.1.0
811
+ urwid==2.6.10
812
+ urwid_readline==0.14
813
+ uvicorn==0.25.0
814
+ uvloop==0.19.0
815
+ vaex-astro==0.9.3
816
+ vaex-core==4.17.1
817
+ vaex-hdf5==0.14.1
818
+ vaex-jupyter==0.8.2
819
+ vaex-ml==0.18.3
820
+ vaex-server==0.9.0
821
+ vaex-viz==0.5.4
822
+ vaex==4.17.0
823
+ vec_noise==1.1.4
824
+ vecstack==0.4.0
825
+ virtualenv==20.21.0
826
+ visions==0.7.5
827
+ vowpalwabbit==9.9.0
828
+ vtk==9.3.0
829
+ wandb==0.16.5
830
+ wasabi==1.1.2
831
+ watchfiles==0.21.0
832
+ wavio==0.0.8
833
+ wcwidth==0.2.13
834
+ weasel==0.3.4
835
+ webcolors==1.13
836
+ webencodings==0.5.1
837
+ websocket-client==1.7.0
838
+ websockets==12.0
839
+ wfdb==4.1.2
840
+ whatthepatch==1.0.5
841
+ wheel==0.42.0
842
+ widgetsnbextension==3.6.6
843
+ witwidget==1.8.1
844
+ woodwork==0.29.0
845
+ wordcloud==1.9.3
846
+ wordsegment==1.3.1
847
+ wrapt==1.14.1
848
+ xarray-einstats==0.7.0
849
+ xarray==2024.3.0
850
+ xgboost==2.0.3
851
+ xvfbwrapper==0.2.9
852
+ xxhash==3.4.1
853
+ xyzservices==2023.10.1
854
+ y-py==0.6.2
855
+ yapf==0.40.2
856
+ yarl==1.9.3
857
+ yarl==1.9.4
858
+ ydata-profiling==4.6.4
859
+ yellowbrick==1.5
860
+ ypy-websocket==0.8.4
861
+ zict==3.0.0
862
+ zipp==3.17.0
863
+ zstandard==0.22.0
wandb/run-20240410_155650-albqrbhg/files/wandb-metadata.json ADDED
@@ -0,0 +1,62 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.15.133+-x86_64-with-glibc2.31",
3
+ "python": "3.10.13",
4
+ "heartbeatAt": "2024-04-10T15:56:51.319241",
5
+ "startedAt": "2024-04-10T15:56:50.261059",
6
+ "docker": null,
7
+ "cuda": null,
8
+ "args": [],
9
+ "state": "running",
10
+ "program": "kaggle.ipynb",
11
+ "codePathLocal": null,
12
+ "root": "/kaggle/working",
13
+ "host": "ae604480b8c9",
14
+ "username": "root",
15
+ "executable": "/opt/conda/bin/python3.10",
16
+ "cpu_count": 2,
17
+ "cpu_count_logical": 4,
18
+ "cpu_freq": {
19
+ "current": 2000.192,
20
+ "min": 0.0,
21
+ "max": 0.0
22
+ },
23
+ "cpu_freq_per_core": [
24
+ {
25
+ "current": 2000.192,
26
+ "min": 0.0,
27
+ "max": 0.0
28
+ },
29
+ {
30
+ "current": 2000.192,
31
+ "min": 0.0,
32
+ "max": 0.0
33
+ },
34
+ {
35
+ "current": 2000.192,
36
+ "min": 0.0,
37
+ "max": 0.0
38
+ },
39
+ {
40
+ "current": 2000.192,
41
+ "min": 0.0,
42
+ "max": 0.0
43
+ }
44
+ ],
45
+ "disk": {
46
+ "/": {
47
+ "total": 8062.387607574463,
48
+ "used": 5563.648036956787
49
+ }
50
+ },
51
+ "gpu": "Tesla P100-PCIE-16GB",
52
+ "gpu_count": 1,
53
+ "gpu_devices": [
54
+ {
55
+ "name": "Tesla P100-PCIE-16GB",
56
+ "memory_total": 17179869184
57
+ }
58
+ ],
59
+ "memory": {
60
+ "total": 31.357559204101562
61
+ }
62
+ }
wandb/run-20240410_155650-albqrbhg/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"train/loss": 1.9459, "train/grad_norm": 6.674560546875, "train/learning_rate": 8.000000000000001e-06, "train/epoch": 1.0, "train/global_step": 59, "_timestamp": 1712765900.9632816, "_runtime": 1290.6948807239532, "_step": 3, "eval/loss": 1.9451844692230225, "eval/accuracy": 0.2112676056338028, "eval/precision": 0.030181086519114688, "eval/recall": 0.14285714285714285, "eval/f1": 0.04983388704318937, "eval/runtime": 5.393, "eval/samples_per_second": 26.33, "eval/steps_per_second": 1.669}
wandb/run-20240410_155650-albqrbhg/logs/debug-internal.log ADDED
The diff for this file is too large to render. See raw diff
 
wandb/run-20240410_155650-albqrbhg/logs/debug.log ADDED
@@ -0,0 +1,153 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-04-10 15:56:50,262 INFO MainThread:34 [wandb_setup.py:_flush():76] Current SDK version is 0.16.5
2
+ 2024-04-10 15:56:50,262 INFO MainThread:34 [wandb_setup.py:_flush():76] Configure stats pid to 34
3
+ 2024-04-10 15:56:50,263 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
4
+ 2024-04-10 15:56:50,263 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from /kaggle/working/wandb/settings
5
+ 2024-04-10 15:56:50,263 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
6
+ 2024-04-10 15:56:50,263 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
7
+ 2024-04-10 15:56:50,263 INFO MainThread:34 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program': '<python with no main file>'}
8
+ 2024-04-10 15:56:50,263 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying login settings: {}
9
+ 2024-04-10 15:56:50,263 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying login settings: {'api_key': '***REDACTED***'}
10
+ 2024-04-10 15:56:50,263 INFO MainThread:34 [wandb_init.py:_log_setup():527] Logging user logs to /kaggle/working/wandb/run-20240410_155650-albqrbhg/logs/debug.log
11
+ 2024-04-10 15:56:50,263 INFO MainThread:34 [wandb_init.py:_log_setup():528] Logging internal logs to /kaggle/working/wandb/run-20240410_155650-albqrbhg/logs/debug-internal.log
12
+ 2024-04-10 15:56:50,263 INFO MainThread:34 [wandb_init.py:_jupyter_setup():473] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x7ff5302cfca0>
13
+ 2024-04-10 15:56:50,263 INFO MainThread:34 [wandb_init.py:init():567] calling init triggers
14
+ 2024-04-10 15:56:50,263 INFO MainThread:34 [wandb_init.py:init():574] wandb.init called with sweep_config: {}
15
+ config: {}
16
+ 2024-04-10 15:56:50,264 INFO MainThread:34 [wandb_init.py:init():617] starting backend
17
+ 2024-04-10 15:56:50,264 INFO MainThread:34 [wandb_init.py:init():621] setting up manager
18
+ 2024-04-10 15:56:50,265 INFO MainThread:34 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
19
+ 2024-04-10 15:56:50,268 INFO MainThread:34 [wandb_init.py:init():629] backend started and connected
20
+ 2024-04-10 15:56:50,281 INFO MainThread:34 [wandb_run.py:_label_probe_notebook():1299] probe notebook
21
+ 2024-04-10 15:56:51,015 INFO MainThread:34 [wandb_init.py:init():721] updated telemetry
22
+ 2024-04-10 15:56:51,018 INFO MainThread:34 [wandb_init.py:init():754] communicating run to backend with 90.0 second timeout
23
+ 2024-04-10 15:56:51,225 INFO MainThread:34 [wandb_run.py:_on_init():2344] communicating current version
24
+ 2024-04-10 15:56:51,292 INFO MainThread:34 [wandb_run.py:_on_init():2353] got version response upgrade_message: "wandb version 0.16.6 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
25
+
26
+ 2024-04-10 15:56:51,292 INFO MainThread:34 [wandb_init.py:init():805] starting run threads in backend
27
+ 2024-04-10 15:57:07,286 INFO MainThread:34 [wandb_run.py:_console_start():2323] atexit reg
28
+ 2024-04-10 15:57:07,286 INFO MainThread:34 [wandb_run.py:_redirect():2178] redirect: wrap_raw
29
+ 2024-04-10 15:57:07,286 INFO MainThread:34 [wandb_run.py:_redirect():2243] Wrapping output streams.
30
+ 2024-04-10 15:57:07,287 INFO MainThread:34 [wandb_run.py:_redirect():2268] Redirects installed.
31
+ 2024-04-10 15:57:07,288 INFO MainThread:34 [wandb_init.py:init():848] run started, returning control to user process
32
+ 2024-04-10 15:57:07,294 INFO MainThread:34 [wandb_run.py:_config_callback():1347] config_cb None None {'return_dict': True, 'output_hidden_states': True, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['BertModel'], 'finetuning_task': None, 'id2label': {0: 'Documentation Replication on Other Examples', 1: 'Requesting (Additional) Resources', 2: 'Inadequate Examples', 3: 'Documentation Ambiguity', 4: 'Documentation Completeness', 5: 'Lack of Alternative Solutions/Documentation', 6: 'Documentation Replicability'}, 'label2id': {'Documentation Replication on Other Examples': 0, 'Requesting (Additional) Resources': 1, 'Inadequate Examples': 2, 'Documentation Ambiguity': 3, 'Documentation Completeness': 4, 'Lack of Alternative Solutions/Documentation': 5, 'Documentation Replicability': 6}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': 0, 'eos_token_id': None, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'lanwuwei/BERTOverflow_stackoverflow_github', 'transformers_version': '4.39.3', '_num_labels': 51, 'model_type': 'bert', 'output_past': True, 'vocab_size': 82000, 'hidden_size': 768, 'num_hidden_layers': 12, 'num_attention_heads': 12, 'hidden_act': 'gelu', 'intermediate_size': 3072, 'hidden_dropout_prob': 0.1, 'attention_probs_dropout_prob': 0.1, 'max_position_embeddings': 512, 'type_vocab_size': 2, 'initializer_range': 0.02, 'layer_norm_eps': 1e-12, 'position_embedding_type': 'absolute', 'use_cache': True, 'classifier_dropout': None, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': True, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 16, 'per_device_eval_batch_size': 16, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 1e-05, 'weight_decay': 0.01, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 5, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Apr10_15-56-15_ae604480b8c9', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': 500, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': True, 'metric_for_best_model': 'loss', 'greater_is_better': False, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None}
33
+ 2024-04-10 15:58:28,784 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
34
+ 2024-04-10 15:58:28,785 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
35
+ 2024-04-10 16:01:09,249 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
36
+ 2024-04-10 16:01:10,407 INFO MainThread:34 [wandb_run.py:_config_callback():1347] config_cb None None {'return_dict': True, 'output_hidden_states': True, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['BertModel'], 'finetuning_task': None, 'id2label': {0: 'Documentation Replication on Other Examples', 1: 'Requesting (Additional) Resources', 2: 'Inadequate Examples', 3: 'Documentation Ambiguity', 4: 'Documentation Completeness', 5: 'Lack of Alternative Solutions/Documentation', 6: 'Documentation Replicability'}, 'label2id': {'Documentation Replication on Other Examples': 0, 'Requesting (Additional) Resources': 1, 'Inadequate Examples': 2, 'Documentation Ambiguity': 3, 'Documentation Completeness': 4, 'Lack of Alternative Solutions/Documentation': 5, 'Documentation Replicability': 6}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': 0, 'eos_token_id': None, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'lanwuwei/BERTOverflow_stackoverflow_github', 'transformers_version': '4.39.3', '_num_labels': 51, 'model_type': 'bert', 'output_past': True, 'vocab_size': 82000, 'hidden_size': 768, 'num_hidden_layers': 12, 'num_attention_heads': 12, 'hidden_act': 'gelu', 'intermediate_size': 3072, 'hidden_dropout_prob': 0.1, 'attention_probs_dropout_prob': 0.1, 'max_position_embeddings': 512, 'type_vocab_size': 2, 'initializer_range': 0.02, 'layer_norm_eps': 1e-12, 'position_embedding_type': 'absolute', 'use_cache': True, 'classifier_dropout': None, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 16, 'per_device_eval_batch_size': 16, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 1e-05, 'weight_decay': 0.01, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 5, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Apr10_16-01-09_ae604480b8c9', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': 500, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': True, 'metric_for_best_model': 'loss', 'greater_is_better': False, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None}
37
+ 2024-04-10 16:02:05,703 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
38
+ 2024-04-10 16:02:05,703 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
39
+ 2024-04-10 16:02:15,608 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
40
+ 2024-04-10 16:02:15,610 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
41
+ 2024-04-10 16:02:15,610 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
42
+ 2024-04-10 16:02:54,085 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
43
+ 2024-04-10 16:02:55,203 INFO MainThread:34 [wandb_run.py:_config_callback():1347] config_cb None None {'return_dict': True, 'output_hidden_states': True, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['BertModel'], 'finetuning_task': None, 'id2label': {0: 'Documentation Replication on Other Examples', 1: 'Requesting (Additional) Resources', 2: 'Inadequate Examples', 3: 'Documentation Ambiguity', 4: 'Documentation Completeness', 5: 'Lack of Alternative Solutions/Documentation', 6: 'Documentation Replicability'}, 'label2id': {'Documentation Replication on Other Examples': 0, 'Requesting (Additional) Resources': 1, 'Inadequate Examples': 2, 'Documentation Ambiguity': 3, 'Documentation Completeness': 4, 'Lack of Alternative Solutions/Documentation': 5, 'Documentation Replicability': 6}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': 0, 'eos_token_id': None, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'lanwuwei/BERTOverflow_stackoverflow_github', 'transformers_version': '4.39.3', '_num_labels': 51, 'model_type': 'bert', 'output_past': True, 'vocab_size': 82000, 'hidden_size': 768, 'num_hidden_layers': 12, 'num_attention_heads': 12, 'hidden_act': 'gelu', 'intermediate_size': 3072, 'hidden_dropout_prob': 0.1, 'attention_probs_dropout_prob': 0.1, 'max_position_embeddings': 512, 'type_vocab_size': 2, 'initializer_range': 0.02, 'layer_norm_eps': 1e-12, 'position_embedding_type': 'absolute', 'use_cache': True, 'classifier_dropout': None, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'epoch', 'prediction_loss_only': False, 'per_device_train_batch_size': 16, 'per_device_eval_batch_size': 16, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 1e-05, 'weight_decay': 0.01, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 5, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Apr10_16-02-54_ae604480b8c9', 'logging_strategy': 'epoch', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'epoch', 'save_steps': 500, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': True, 'metric_for_best_model': 'loss', 'greater_is_better': False, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None}
44
+ 2024-04-10 16:03:53,536 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
45
+ 2024-04-10 16:03:53,536 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
46
+ 2024-04-10 16:11:50,077 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
47
+ 2024-04-10 16:11:50,079 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
48
+ 2024-04-10 16:11:50,079 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
49
+ 2024-04-10 16:12:02,272 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
50
+ 2024-04-10 16:12:02,274 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
51
+ 2024-04-10 16:12:02,275 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
52
+ 2024-04-10 16:12:15,300 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
53
+ 2024-04-10 16:12:15,302 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
54
+ 2024-04-10 16:12:15,303 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
55
+ 2024-04-10 16:12:18,676 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
56
+ 2024-04-10 16:12:18,734 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
57
+ 2024-04-10 16:12:18,735 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
58
+ 2024-04-10 16:12:20,635 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
59
+ 2024-04-10 16:12:20,643 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
60
+ 2024-04-10 16:12:20,643 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
61
+ 2024-04-10 16:12:23,458 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
62
+ 2024-04-10 16:12:23,465 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
63
+ 2024-04-10 16:12:23,465 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
64
+ 2024-04-10 16:12:27,135 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
65
+ 2024-04-10 16:12:27,146 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
66
+ 2024-04-10 16:12:27,146 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
67
+ 2024-04-10 16:12:34,928 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
68
+ 2024-04-10 16:12:34,934 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
69
+ 2024-04-10 16:12:34,934 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
70
+ 2024-04-10 16:12:51,967 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
71
+ 2024-04-10 16:12:52,930 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
72
+ 2024-04-10 16:12:52,930 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
73
+ 2024-04-10 16:13:14,813 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
74
+ 2024-04-10 16:13:14,815 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
75
+ 2024-04-10 16:13:14,815 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
76
+ 2024-04-10 16:13:22,627 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
77
+ 2024-04-10 16:13:22,634 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
78
+ 2024-04-10 16:13:22,634 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
79
+ 2024-04-10 16:13:26,638 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
80
+ 2024-04-10 16:13:26,688 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
81
+ 2024-04-10 16:13:26,688 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
82
+ 2024-04-10 16:13:28,942 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
83
+ 2024-04-10 16:13:28,949 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
84
+ 2024-04-10 16:13:28,949 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
85
+ 2024-04-10 16:13:32,221 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
86
+ 2024-04-10 16:13:32,227 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
87
+ 2024-04-10 16:13:32,227 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
88
+ 2024-04-10 16:13:35,802 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
89
+ 2024-04-10 16:13:35,814 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
90
+ 2024-04-10 16:13:35,814 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
91
+ 2024-04-10 16:13:38,310 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
92
+ 2024-04-10 16:13:38,314 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
93
+ 2024-04-10 16:13:38,314 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
94
+ 2024-04-10 16:13:43,154 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
95
+ 2024-04-10 16:13:44,088 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
96
+ 2024-04-10 16:13:44,088 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
97
+ 2024-04-10 16:13:46,903 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
98
+ 2024-04-10 16:13:46,926 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
99
+ 2024-04-10 16:13:46,926 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
100
+ 2024-04-10 16:13:48,880 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
101
+ 2024-04-10 16:13:48,891 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
102
+ 2024-04-10 16:13:48,891 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
103
+ 2024-04-10 16:13:51,328 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
104
+ 2024-04-10 16:13:51,353 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
105
+ 2024-04-10 16:13:51,353 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
106
+ 2024-04-10 16:13:53,267 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
107
+ 2024-04-10 16:13:53,268 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
108
+ 2024-04-10 16:13:53,269 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
109
+ 2024-04-10 16:14:03,755 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
110
+ 2024-04-10 16:14:04,296 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
111
+ 2024-04-10 16:14:04,296 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
112
+ 2024-04-10 16:14:34,345 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
113
+ 2024-04-10 16:14:36,175 INFO MainThread:34 [wandb_run.py:_config_callback():1347] config_cb None None {'return_dict': True, 'output_hidden_states': True, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['BertModel'], 'finetuning_task': None, 'id2label': {0: 'Documentation Replication on Other Examples', 1: 'Requesting (Additional) Resources', 2: 'Inadequate Examples', 3: 'Documentation Ambiguity', 4: 'Documentation Completeness', 5: 'Lack of Alternative Solutions/Documentation', 6: 'Documentation Replicability'}, 'label2id': {'Documentation Replication on Other Examples': 0, 'Requesting (Additional) Resources': 1, 'Inadequate Examples': 2, 'Documentation Ambiguity': 3, 'Documentation Completeness': 4, 'Lack of Alternative Solutions/Documentation': 5, 'Documentation Replicability': 6}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': 0, 'eos_token_id': None, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'lanwuwei/BERTOverflow_stackoverflow_github', 'transformers_version': '4.39.3', '_num_labels': 51, 'model_type': 'bert', 'output_past': True, 'vocab_size': 82000, 'hidden_size': 768, 'num_hidden_layers': 12, 'num_attention_heads': 12, 'hidden_act': 'gelu', 'intermediate_size': 3072, 'hidden_dropout_prob': 0.1, 'attention_probs_dropout_prob': 0.1, 'max_position_embeddings': 512, 'type_vocab_size': 2, 'initializer_range': 0.02, 'layer_norm_eps': 1e-12, 'position_embedding_type': 'absolute', 'use_cache': True, 'classifier_dropout': None, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'epoch', 'prediction_loss_only': False, 'per_device_train_batch_size': 16, 'per_device_eval_batch_size': 16, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 1e-05, 'weight_decay': 0.01, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 5, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Apr10_16-14-35_ae604480b8c9', 'logging_strategy': 'epoch', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'epoch', 'save_steps': 500, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': True, 'metric_for_best_model': 'loss', 'greater_is_better': False, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None}
114
+ 2024-04-10 16:15:34,096 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
115
+ 2024-04-10 16:15:34,096 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
116
+ 2024-04-10 16:16:34,710 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
117
+ 2024-04-10 16:16:34,712 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
118
+ 2024-04-10 16:16:34,713 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
119
+ 2024-04-10 16:16:59,060 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
120
+ 2024-04-10 16:16:59,062 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
121
+ 2024-04-10 16:16:59,062 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
122
+ 2024-04-10 16:17:05,777 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
123
+ 2024-04-10 16:17:05,828 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
124
+ 2024-04-10 16:17:05,828 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
125
+ 2024-04-10 16:17:07,955 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
126
+ 2024-04-10 16:17:07,961 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
127
+ 2024-04-10 16:17:07,961 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
128
+ 2024-04-10 16:17:09,861 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
129
+ 2024-04-10 16:17:09,867 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
130
+ 2024-04-10 16:17:09,867 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
131
+ 2024-04-10 16:17:12,727 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
132
+ 2024-04-10 16:17:12,738 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
133
+ 2024-04-10 16:17:12,738 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
134
+ 2024-04-10 16:17:13,820 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
135
+ 2024-04-10 16:17:13,824 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
136
+ 2024-04-10 16:17:13,825 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
137
+ 2024-04-10 16:17:14,812 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
138
+ 2024-04-10 16:17:15,659 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
139
+ 2024-04-10 16:17:15,659 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
140
+ 2024-04-10 16:17:17,681 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
141
+ 2024-04-10 16:17:17,704 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
142
+ 2024-04-10 16:17:17,704 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
143
+ 2024-04-10 16:17:18,381 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
144
+ 2024-04-10 16:17:18,392 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
145
+ 2024-04-10 16:17:18,392 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
146
+ 2024-04-10 16:17:19,694 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
147
+ 2024-04-10 16:17:19,718 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
148
+ 2024-04-10 16:17:19,718 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
149
+ 2024-04-10 16:17:20,601 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
150
+ 2024-04-10 16:17:20,603 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
151
+ 2024-04-10 16:17:20,603 INFO MainThread:34 [wandb_init.py:_pause_backend():438] pausing backend
152
+ 2024-04-10 16:17:21,870 INFO MainThread:34 [wandb_init.py:_resume_backend():443] resuming backend
153
+ 2024-04-10 16:17:23,042 INFO MainThread:34 [wandb_run.py:_config_callback():1347] config_cb None None {'return_dict': True, 'output_hidden_states': True, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['BertModel'], 'finetuning_task': None, 'id2label': {0: 'Documentation Replication on Other Examples', 1: 'Requesting (Additional) Resources', 2: 'Inadequate Examples', 3: 'Documentation Ambiguity', 4: 'Documentation Completeness', 5: 'Lack of Alternative Solutions/Documentation', 6: 'Documentation Replicability'}, 'label2id': {'Documentation Replication on Other Examples': 0, 'Requesting (Additional) Resources': 1, 'Inadequate Examples': 2, 'Documentation Ambiguity': 3, 'Documentation Completeness': 4, 'Lack of Alternative Solutions/Documentation': 5, 'Documentation Replicability': 6}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': 0, 'eos_token_id': None, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'lanwuwei/BERTOverflow_stackoverflow_github', 'transformers_version': '4.39.3', '_num_labels': 51, 'model_type': 'bert', 'output_past': True, 'vocab_size': 82000, 'hidden_size': 768, 'num_hidden_layers': 12, 'num_attention_heads': 12, 'hidden_act': 'gelu', 'intermediate_size': 3072, 'hidden_dropout_prob': 0.1, 'attention_probs_dropout_prob': 0.1, 'max_position_embeddings': 512, 'type_vocab_size': 2, 'initializer_range': 0.02, 'layer_norm_eps': 1e-12, 'position_embedding_type': 'absolute', 'use_cache': True, 'classifier_dropout': None, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'epoch', 'prediction_loss_only': False, 'per_device_train_batch_size': 16, 'per_device_eval_batch_size': 16, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 1e-05, 'weight_decay': 0.01, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 5, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/Apr10_16-17-22_ae604480b8c9', 'logging_strategy': 'epoch', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'epoch', 'save_steps': 500, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': True, 'metric_for_best_model': 'loss', 'greater_is_better': False, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None}
wandb/run-20240410_155650-albqrbhg/run-albqrbhg.wandb ADDED
Binary file (65.8 kB). View file