darapota commited on
Commit
a4419fd
·
verified ·
1 Parent(s): 2155c9c

Upload folder using huggingface_hub

Browse files
config.json CHANGED
@@ -1,12 +1,11 @@
1
  {
2
- "_name_or_path": "roberta-base",
3
  "architectures": [
4
- "RobertaForSequenceClassification"
5
  ],
6
  "attention_probs_dropout_prob": 0.1,
7
- "bos_token_id": 0,
8
  "classifier_dropout": null,
9
- "eos_token_id": 2,
10
  "hidden_act": "gelu",
11
  "hidden_dropout_prob": 0.1,
12
  "hidden_size": 768,
@@ -19,54 +18,37 @@
19
  "5": "LABEL_5",
20
  "6": "LABEL_6",
21
  "7": "LABEL_7",
22
- "8": "LABEL_8",
23
- "9": "LABEL_9",
24
- "10": "LABEL_10",
25
- "11": "LABEL_11",
26
- "12": "LABEL_12",
27
- "13": "LABEL_13",
28
- "14": "LABEL_14",
29
- "15": "LABEL_15",
30
- "16": "LABEL_16",
31
- "17": "LABEL_17",
32
- "18": "LABEL_18",
33
- "19": "LABEL_19"
34
  },
35
  "initializer_range": 0.02,
36
  "intermediate_size": 3072,
37
  "label2id": {
38
  "LABEL_0": 0,
39
  "LABEL_1": 1,
40
- "LABEL_10": 10,
41
- "LABEL_11": 11,
42
- "LABEL_12": 12,
43
- "LABEL_13": 13,
44
- "LABEL_14": 14,
45
- "LABEL_15": 15,
46
- "LABEL_16": 16,
47
- "LABEL_17": 17,
48
- "LABEL_18": 18,
49
- "LABEL_19": 19,
50
  "LABEL_2": 2,
51
  "LABEL_3": 3,
52
  "LABEL_4": 4,
53
  "LABEL_5": 5,
54
  "LABEL_6": 6,
55
  "LABEL_7": 7,
56
- "LABEL_8": 8,
57
- "LABEL_9": 9
58
  },
59
- "layer_norm_eps": 1e-05,
60
- "max_position_embeddings": 514,
61
- "model_type": "roberta",
62
  "num_attention_heads": 12,
63
  "num_hidden_layers": 12,
64
- "pad_token_id": 1,
 
 
 
 
 
65
  "position_embedding_type": "absolute",
66
  "problem_type": "single_label_classification",
67
  "torch_dtype": "float32",
68
  "transformers_version": "4.41.2",
69
- "type_vocab_size": 1,
70
  "use_cache": true,
71
- "vocab_size": 50265
72
  }
 
1
  {
2
+ "_name_or_path": "ai-forever/ruBert-base",
3
  "architectures": [
4
+ "BertForSequenceClassification"
5
  ],
6
  "attention_probs_dropout_prob": 0.1,
 
7
  "classifier_dropout": null,
8
+ "directionality": "bidi",
9
  "hidden_act": "gelu",
10
  "hidden_dropout_prob": 0.1,
11
  "hidden_size": 768,
 
18
  "5": "LABEL_5",
19
  "6": "LABEL_6",
20
  "7": "LABEL_7",
21
+ "8": "LABEL_8"
 
 
 
 
 
 
 
 
 
 
 
22
  },
23
  "initializer_range": 0.02,
24
  "intermediate_size": 3072,
25
  "label2id": {
26
  "LABEL_0": 0,
27
  "LABEL_1": 1,
 
 
 
 
 
 
 
 
 
 
28
  "LABEL_2": 2,
29
  "LABEL_3": 3,
30
  "LABEL_4": 4,
31
  "LABEL_5": 5,
32
  "LABEL_6": 6,
33
  "LABEL_7": 7,
34
+ "LABEL_8": 8
 
35
  },
36
+ "layer_norm_eps": 1e-12,
37
+ "max_position_embeddings": 512,
38
+ "model_type": "bert",
39
  "num_attention_heads": 12,
40
  "num_hidden_layers": 12,
41
+ "pad_token_id": 0,
42
+ "pooler_fc_size": 768,
43
+ "pooler_num_attention_heads": 12,
44
+ "pooler_num_fc_layers": 3,
45
+ "pooler_size_per_head": 128,
46
+ "pooler_type": "first_token_transform",
47
  "position_embedding_type": "absolute",
48
  "problem_type": "single_label_classification",
49
  "torch_dtype": "float32",
50
  "transformers_version": "4.41.2",
51
+ "type_vocab_size": 2,
52
  "use_cache": true,
53
+ "vocab_size": 120138
54
  }
label_encoder.joblib CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a409954c3cfbacdbd5a19099a851a93f386b0f467c894374e6e016095cbdd790
3
- size 1159
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d27c95bec371218940a143acd7bb374065280c3290f685c9427170df213816a6
3
+ size 729
logs/events.out.tfevents.1720001204.2679035f67af.34.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cf0581f8e1b0de0974fdce63596f30cc796999d73d08705716ba545e11e2b89d
3
+ size 5341
logs/events.out.tfevents.1720001560.2679035f67af.34.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6ad22dda0a8058cba10d9c98766fc055c5bde62b0ab9bc6830f650a48b31068e
3
+ size 4184
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f085b02066480476861c072551891a9355d5d48667914f1890aeab42c24d0a31
3
- size 498668192
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8943bedc389f9c0498b8625478228be76be72eb7797633ada3b0bc946cb1e889
3
+ size 713280540
special_tokens_map.json CHANGED
@@ -1,15 +1,7 @@
1
  {
2
- "bos_token": "<s>",
3
- "cls_token": "<s>",
4
- "eos_token": "</s>",
5
- "mask_token": {
6
- "content": "<mask>",
7
- "lstrip": true,
8
- "normalized": false,
9
- "rstrip": false,
10
- "single_word": false
11
- },
12
- "pad_token": "<pad>",
13
- "sep_token": "</s>",
14
- "unk_token": "<unk>"
15
  }
 
1
  {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
 
 
 
 
 
 
 
 
7
  }
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -1,57 +1,57 @@
1
  {
2
- "add_prefix_space": false,
3
  "added_tokens_decoder": {
4
  "0": {
5
- "content": "<s>",
6
  "lstrip": false,
7
- "normalized": true,
8
  "rstrip": false,
9
  "single_word": false,
10
  "special": true
11
  },
12
- "1": {
13
- "content": "<pad>",
14
  "lstrip": false,
15
- "normalized": true,
16
  "rstrip": false,
17
  "single_word": false,
18
  "special": true
19
  },
20
- "2": {
21
- "content": "</s>",
22
  "lstrip": false,
23
- "normalized": true,
24
  "rstrip": false,
25
  "single_word": false,
26
  "special": true
27
  },
28
- "3": {
29
- "content": "<unk>",
30
  "lstrip": false,
31
- "normalized": true,
32
  "rstrip": false,
33
  "single_word": false,
34
  "special": true
35
  },
36
- "50264": {
37
- "content": "<mask>",
38
- "lstrip": true,
39
  "normalized": false,
40
  "rstrip": false,
41
  "single_word": false,
42
  "special": true
43
  }
44
  },
45
- "bos_token": "<s>",
46
  "clean_up_tokenization_spaces": true,
47
- "cls_token": "<s>",
48
- "eos_token": "</s>",
49
- "errors": "replace",
50
- "mask_token": "<mask>",
51
- "model_max_length": 512,
52
- "pad_token": "<pad>",
53
- "sep_token": "</s>",
54
- "tokenizer_class": "RobertaTokenizer",
55
- "trim_offsets": true,
56
- "unk_token": "<unk>"
 
 
57
  }
 
1
  {
 
2
  "added_tokens_decoder": {
3
  "0": {
4
+ "content": "[PAD]",
5
  "lstrip": false,
6
+ "normalized": false,
7
  "rstrip": false,
8
  "single_word": false,
9
  "special": true
10
  },
11
+ "100": {
12
+ "content": "[UNK]",
13
  "lstrip": false,
14
+ "normalized": false,
15
  "rstrip": false,
16
  "single_word": false,
17
  "special": true
18
  },
19
+ "101": {
20
+ "content": "[CLS]",
21
  "lstrip": false,
22
+ "normalized": false,
23
  "rstrip": false,
24
  "single_word": false,
25
  "special": true
26
  },
27
+ "102": {
28
+ "content": "[SEP]",
29
  "lstrip": false,
30
+ "normalized": false,
31
  "rstrip": false,
32
  "single_word": false,
33
  "special": true
34
  },
35
+ "103": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
  "normalized": false,
39
  "rstrip": false,
40
  "single_word": false,
41
  "special": true
42
  }
43
  },
 
44
  "clean_up_tokenization_spaces": true,
45
+ "cls_token": "[CLS]",
46
+ "do_basic_tokenize": true,
47
+ "do_lower_case": true,
48
+ "mask_token": "[MASK]",
49
+ "model_max_length": 1000000000000000019884624838656,
50
+ "never_split": null,
51
+ "pad_token": "[PAD]",
52
+ "sep_token": "[SEP]",
53
+ "strip_accents": null,
54
+ "tokenize_chinese_chars": true,
55
+ "tokenizer_class": "BertTokenizer",
56
+ "unk_token": "[UNK]"
57
  }
vocab.txt ADDED
The diff for this file is too large to render. See raw diff
 
wandb/debug-internal.log CHANGED
The diff for this file is too large to render. See raw diff
 
wandb/debug.log CHANGED
@@ -1,35 +1,45 @@
1
- 2024-06-30 20:03:52,241 INFO MainThread:34 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0
2
- 2024-06-30 20:03:52,241 INFO MainThread:34 [wandb_setup.py:_flush():76] Configure stats pid to 34
3
- 2024-06-30 20:03:52,241 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
4
- 2024-06-30 20:03:52,241 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from /kaggle/working/wandb/settings
5
- 2024-06-30 20:03:52,241 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
6
- 2024-06-30 20:03:52,241 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
7
- 2024-06-30 20:03:52,241 INFO MainThread:34 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program': '<python with no main file>'}
8
- 2024-06-30 20:03:52,241 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying login settings: {}
9
- 2024-06-30 20:03:52,241 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying login settings: {'api_key': '***REDACTED***'}
10
- 2024-06-30 20:03:52,241 INFO MainThread:34 [wandb_init.py:_log_setup():520] Logging user logs to /kaggle/working/wandb/run-20240630_200352-ydm2qdsc/logs/debug.log
11
- 2024-06-30 20:03:52,241 INFO MainThread:34 [wandb_init.py:_log_setup():521] Logging internal logs to /kaggle/working/wandb/run-20240630_200352-ydm2qdsc/logs/debug-internal.log
12
- 2024-06-30 20:03:52,241 INFO MainThread:34 [wandb_init.py:_jupyter_setup():466] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x79acede24250>
13
- 2024-06-30 20:03:52,242 INFO MainThread:34 [wandb_init.py:init():560] calling init triggers
14
- 2024-06-30 20:03:52,242 INFO MainThread:34 [wandb_init.py:init():567] wandb.init called with sweep_config: {}
15
  config: {}
16
- 2024-06-30 20:03:52,242 INFO MainThread:34 [wandb_init.py:init():610] starting backend
17
- 2024-06-30 20:03:52,242 INFO MainThread:34 [wandb_init.py:init():614] setting up manager
18
- 2024-06-30 20:03:52,243 INFO MainThread:34 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
19
- 2024-06-30 20:03:52,246 INFO MainThread:34 [wandb_init.py:init():622] backend started and connected
20
- 2024-06-30 20:03:52,258 INFO MainThread:34 [wandb_run.py:_label_probe_notebook():1328] probe notebook
21
- 2024-06-30 20:03:52,506 INFO MainThread:34 [wandb_init.py:init():711] updated telemetry
22
- 2024-06-30 20:03:52,509 INFO MainThread:34 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout
23
- 2024-06-30 20:03:52,733 INFO MainThread:34 [wandb_run.py:_on_init():2396] communicating current version
24
- 2024-06-30 20:03:52,814 INFO MainThread:34 [wandb_run.py:_on_init():2405] got version response upgrade_message: "wandb version 0.17.3 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
25
 
26
- 2024-06-30 20:03:52,815 INFO MainThread:34 [wandb_init.py:init():795] starting run threads in backend
27
- 2024-06-30 20:04:08,848 INFO MainThread:34 [wandb_run.py:_console_start():2374] atexit reg
28
- 2024-06-30 20:04:08,848 INFO MainThread:34 [wandb_run.py:_redirect():2229] redirect: wrap_raw
29
- 2024-06-30 20:04:08,848 INFO MainThread:34 [wandb_run.py:_redirect():2294] Wrapping output streams.
30
- 2024-06-30 20:04:08,848 INFO MainThread:34 [wandb_run.py:_redirect():2319] Redirects installed.
31
- 2024-06-30 20:04:08,852 INFO MainThread:34 [wandb_init.py:init():838] run started, returning control to user process
32
- 2024-06-30 20:04:08,858 INFO MainThread:34 [wandb_run.py:_config_callback():1376] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['RobertaForMaskedLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1', 2: 'LABEL_2', 3: 'LABEL_3', 4: 'LABEL_4', 5: 'LABEL_5', 6: 'LABEL_6', 7: 'LABEL_7', 8: 'LABEL_8', 9: 'LABEL_9', 10: 'LABEL_10', 11: 'LABEL_11', 12: 'LABEL_12', 13: 'LABEL_13', 14: 'LABEL_14', 15: 'LABEL_15', 16: 'LABEL_16', 17: 'LABEL_17', 18: 'LABEL_18', 19: 'LABEL_19'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1, 'LABEL_2': 2, 'LABEL_3': 3, 'LABEL_4': 4, 'LABEL_5': 5, 'LABEL_6': 6, 'LABEL_7': 7, 'LABEL_8': 8, 'LABEL_9': 9, 'LABEL_10': 10, 'LABEL_11': 11, 'LABEL_12': 12, 'LABEL_13': 13, 'LABEL_14': 14, 'LABEL_15': 15, 'LABEL_16': 16, 'LABEL_17': 17, 'LABEL_18': 18, 'LABEL_19': 19}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 0, 'pad_token_id': 1, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'roberta-base', 'transformers_version': '4.41.2', 'model_type': 'roberta', 'vocab_size': 50265, 'hidden_size': 768, 'num_hidden_layers': 12, 'num_attention_heads': 12, 'hidden_act': 'gelu', 'intermediate_size': 3072, 'hidden_dropout_prob': 0.1, 'attention_probs_dropout_prob': 0.1, 'max_position_embeddings': 514, 'type_vocab_size': 1, 'initializer_range': 0.02, 'layer_norm_eps': 1e-05, 'position_embedding_type': 'absolute', 'use_cache': True, 'classifier_dropout': None, 'output_dir': '/kaggle/working/results', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'eval_strategy': 'epoch', 'prediction_loss_only': False, 'per_device_train_batch_size': 16, 'per_device_eval_batch_size': 16, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 2e-05, 'weight_decay': 0.01, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 6, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/logs', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'no', 'save_steps': 500, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'restore_callback_states_from_checkpoint': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/results', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'eval_do_concat_batches': True, 'fp16_backend': 'auto', 'evaluation_strategy': 'epoch', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None, 'batch_eval_metrics': False}
33
- 2024-06-30 20:31:29,741 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
34
- 2024-06-30 20:31:29,741 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
35
- 2024-06-30 20:33:36,476 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-07-03 10:07:00,447 INFO MainThread:34 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0
2
+ 2024-07-03 10:07:00,447 INFO MainThread:34 [wandb_setup.py:_flush():76] Configure stats pid to 34
3
+ 2024-07-03 10:07:00,447 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
4
+ 2024-07-03 10:07:00,447 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from /kaggle/working/wandb/settings
5
+ 2024-07-03 10:07:00,447 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
6
+ 2024-07-03 10:07:00,447 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
7
+ 2024-07-03 10:07:00,447 INFO MainThread:34 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program': '<python with no main file>'}
8
+ 2024-07-03 10:07:00,447 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying login settings: {}
9
+ 2024-07-03 10:07:00,447 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying login settings: {'api_key': '***REDACTED***'}
10
+ 2024-07-03 10:07:00,447 INFO MainThread:34 [wandb_init.py:_log_setup():520] Logging user logs to /kaggle/working/wandb/run-20240703_100700-cyvdged5/logs/debug.log
11
+ 2024-07-03 10:07:00,447 INFO MainThread:34 [wandb_init.py:_log_setup():521] Logging internal logs to /kaggle/working/wandb/run-20240703_100700-cyvdged5/logs/debug-internal.log
12
+ 2024-07-03 10:07:00,447 INFO MainThread:34 [wandb_init.py:_jupyter_setup():466] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x78e7a57abb20>
13
+ 2024-07-03 10:07:00,448 INFO MainThread:34 [wandb_init.py:init():560] calling init triggers
14
+ 2024-07-03 10:07:00,448 INFO MainThread:34 [wandb_init.py:init():567] wandb.init called with sweep_config: {}
15
  config: {}
16
+ 2024-07-03 10:07:00,448 INFO MainThread:34 [wandb_init.py:init():610] starting backend
17
+ 2024-07-03 10:07:00,448 INFO MainThread:34 [wandb_init.py:init():614] setting up manager
18
+ 2024-07-03 10:07:00,449 INFO MainThread:34 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
19
+ 2024-07-03 10:07:00,452 INFO MainThread:34 [wandb_init.py:init():622] backend started and connected
20
+ 2024-07-03 10:07:00,464 INFO MainThread:34 [wandb_run.py:_label_probe_notebook():1328] probe notebook
21
+ 2024-07-03 10:07:00,953 INFO MainThread:34 [wandb_init.py:init():711] updated telemetry
22
+ 2024-07-03 10:07:00,957 INFO MainThread:34 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout
23
+ 2024-07-03 10:07:01,536 INFO MainThread:34 [wandb_run.py:_on_init():2396] communicating current version
24
+ 2024-07-03 10:07:01,561 INFO MainThread:34 [wandb_run.py:_on_init():2405] got version response upgrade_message: "wandb version 0.17.3 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
25
 
26
+ 2024-07-03 10:07:01,562 INFO MainThread:34 [wandb_init.py:init():795] starting run threads in backend
27
+ 2024-07-03 10:07:17,603 INFO MainThread:34 [wandb_run.py:_console_start():2374] atexit reg
28
+ 2024-07-03 10:07:17,603 INFO MainThread:34 [wandb_run.py:_redirect():2229] redirect: wrap_raw
29
+ 2024-07-03 10:07:17,604 INFO MainThread:34 [wandb_run.py:_redirect():2294] Wrapping output streams.
30
+ 2024-07-03 10:07:17,604 INFO MainThread:34 [wandb_run.py:_redirect():2319] Redirects installed.
31
+ 2024-07-03 10:07:17,607 INFO MainThread:34 [wandb_init.py:init():838] run started, returning control to user process
32
+ 2024-07-03 10:07:17,615 INFO MainThread:34 [wandb_run.py:_config_callback():1376] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['BertForMaskedLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1', 2: 'LABEL_2', 3: 'LABEL_3', 4: 'LABEL_4', 5: 'LABEL_5', 6: 'LABEL_6', 7: 'LABEL_7', 8: 'LABEL_8'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1, 'LABEL_2': 2, 'LABEL_3': 3, 'LABEL_4': 4, 'LABEL_5': 5, 'LABEL_6': 6, 'LABEL_7': 7, 'LABEL_8': 8}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': 0, 'eos_token_id': None, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'ai-forever/ruBert-base', 'transformers_version': '4.41.2', 'directionality': 'bidi', 'pooler_fc_size': 768, 'pooler_num_attention_heads': 12, 'pooler_num_fc_layers': 3, 'pooler_size_per_head': 128, 'pooler_type': 'first_token_transform', 'model_type': 'bert', 'vocab_size': 120138, 'hidden_size': 768, 'num_hidden_layers': 12, 'num_attention_heads': 12, 'hidden_act': 'gelu', 'intermediate_size': 3072, 'hidden_dropout_prob': 0.1, 'attention_probs_dropout_prob': 0.1, 'max_position_embeddings': 512, 'type_vocab_size': 2, 'initializer_range': 0.02, 'layer_norm_eps': 1e-12, 'position_embedding_type': 'absolute', 'use_cache': True, 'classifier_dropout': None, 'output_dir': '/kaggle/working/results', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'eval_strategy': 'epoch', 'prediction_loss_only': False, 'per_device_train_batch_size': 16, 'per_device_eval_batch_size': 16, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 2e-05, 'weight_decay': 0.01, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 4, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/logs', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'no', 'save_steps': 500, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'restore_callback_states_from_checkpoint': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/results', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'eval_do_concat_batches': True, 'fp16_backend': 'auto', 'evaluation_strategy': 'epoch', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None, 'batch_eval_metrics': False}
33
+ 2024-07-03 10:10:33,327 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
34
+ 2024-07-03 10:10:33,327 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
35
+ 2024-07-03 10:12:28,050 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
36
+ 2024-07-03 10:12:31,153 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
37
+ 2024-07-03 10:12:31,153 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
38
+ 2024-07-03 10:12:36,070 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
39
+ 2024-07-03 10:12:36,071 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
40
+ 2024-07-03 10:12:36,071 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
41
+ 2024-07-03 10:12:39,277 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
42
+ 2024-07-03 10:12:40,028 INFO MainThread:34 [wandb_run.py:_config_callback():1376] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['BertForMaskedLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1', 2: 'LABEL_2', 3: 'LABEL_3', 4: 'LABEL_4', 5: 'LABEL_5', 6: 'LABEL_6', 7: 'LABEL_7', 8: 'LABEL_8'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1, 'LABEL_2': 2, 'LABEL_3': 3, 'LABEL_4': 4, 'LABEL_5': 5, 'LABEL_6': 6, 'LABEL_7': 7, 'LABEL_8': 8}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': 0, 'eos_token_id': None, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'ai-forever/ruBert-base', 'transformers_version': '4.41.2', 'directionality': 'bidi', 'pooler_fc_size': 768, 'pooler_num_attention_heads': 12, 'pooler_num_fc_layers': 3, 'pooler_size_per_head': 128, 'pooler_type': 'first_token_transform', 'model_type': 'bert', 'vocab_size': 120138, 'hidden_size': 768, 'num_hidden_layers': 12, 'num_attention_heads': 12, 'hidden_act': 'gelu', 'intermediate_size': 3072, 'hidden_dropout_prob': 0.1, 'attention_probs_dropout_prob': 0.1, 'max_position_embeddings': 512, 'type_vocab_size': 2, 'initializer_range': 0.02, 'layer_norm_eps': 1e-12, 'position_embedding_type': 'absolute', 'use_cache': True, 'classifier_dropout': None, 'output_dir': '/kaggle/working/results', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'eval_strategy': 'epoch', 'prediction_loss_only': False, 'per_device_train_batch_size': 16, 'per_device_eval_batch_size': 16, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 2e-05, 'weight_decay': 0.01, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 3, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/logs', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'no', 'save_steps': 500, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'restore_callback_states_from_checkpoint': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/results', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'eval_do_concat_batches': True, 'fp16_backend': 'auto', 'evaluation_strategy': 'epoch', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None, 'batch_eval_metrics': False}
43
+ 2024-07-03 10:13:55,032 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
44
+ 2024-07-03 10:13:55,032 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
45
+ 2024-07-03 10:13:58,119 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
wandb/run-20240703_100700-cyvdged5/files/conda-environment.yaml ADDED
File without changes
wandb/run-20240703_100700-cyvdged5/files/config.yaml ADDED
@@ -0,0 +1,695 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ wandb_version: 1
2
+
3
+ _wandb:
4
+ desc: null
5
+ value:
6
+ python_version: 3.10.13
7
+ cli_version: 0.17.0
8
+ framework: huggingface
9
+ huggingface_version: 4.41.2
10
+ is_jupyter_run: true
11
+ is_kaggle_kernel: true
12
+ start_time: 1720001220
13
+ t:
14
+ 1:
15
+ - 1
16
+ - 2
17
+ - 3
18
+ - 5
19
+ - 11
20
+ - 12
21
+ - 49
22
+ - 51
23
+ - 53
24
+ - 55
25
+ - 71
26
+ - 105
27
+ 2:
28
+ - 1
29
+ - 2
30
+ - 3
31
+ - 5
32
+ - 11
33
+ - 12
34
+ - 49
35
+ - 51
36
+ - 53
37
+ - 55
38
+ - 71
39
+ - 105
40
+ 3:
41
+ - 7
42
+ - 13
43
+ - 23
44
+ - 66
45
+ 4: 3.10.13
46
+ 5: 0.17.0
47
+ 6: 4.41.2
48
+ 8:
49
+ - 1
50
+ - 2
51
+ - 5
52
+ 9:
53
+ 1: transformers_trainer
54
+ 13: linux-x86_64
55
+ m:
56
+ - 1: train/global_step
57
+ 6:
58
+ - 3
59
+ return_dict:
60
+ desc: null
61
+ value: true
62
+ output_hidden_states:
63
+ desc: null
64
+ value: false
65
+ output_attentions:
66
+ desc: null
67
+ value: false
68
+ torchscript:
69
+ desc: null
70
+ value: false
71
+ torch_dtype:
72
+ desc: null
73
+ value: null
74
+ use_bfloat16:
75
+ desc: null
76
+ value: false
77
+ tf_legacy_loss:
78
+ desc: null
79
+ value: false
80
+ pruned_heads:
81
+ desc: null
82
+ value: {}
83
+ tie_word_embeddings:
84
+ desc: null
85
+ value: true
86
+ chunk_size_feed_forward:
87
+ desc: null
88
+ value: 0
89
+ is_encoder_decoder:
90
+ desc: null
91
+ value: false
92
+ is_decoder:
93
+ desc: null
94
+ value: false
95
+ cross_attention_hidden_size:
96
+ desc: null
97
+ value: null
98
+ add_cross_attention:
99
+ desc: null
100
+ value: false
101
+ tie_encoder_decoder:
102
+ desc: null
103
+ value: false
104
+ max_length:
105
+ desc: null
106
+ value: 20
107
+ min_length:
108
+ desc: null
109
+ value: 0
110
+ do_sample:
111
+ desc: null
112
+ value: false
113
+ early_stopping:
114
+ desc: null
115
+ value: false
116
+ num_beams:
117
+ desc: null
118
+ value: 1
119
+ num_beam_groups:
120
+ desc: null
121
+ value: 1
122
+ diversity_penalty:
123
+ desc: null
124
+ value: 0.0
125
+ temperature:
126
+ desc: null
127
+ value: 1.0
128
+ top_k:
129
+ desc: null
130
+ value: 50
131
+ top_p:
132
+ desc: null
133
+ value: 1.0
134
+ typical_p:
135
+ desc: null
136
+ value: 1.0
137
+ repetition_penalty:
138
+ desc: null
139
+ value: 1.0
140
+ length_penalty:
141
+ desc: null
142
+ value: 1.0
143
+ no_repeat_ngram_size:
144
+ desc: null
145
+ value: 0
146
+ encoder_no_repeat_ngram_size:
147
+ desc: null
148
+ value: 0
149
+ bad_words_ids:
150
+ desc: null
151
+ value: null
152
+ num_return_sequences:
153
+ desc: null
154
+ value: 1
155
+ output_scores:
156
+ desc: null
157
+ value: false
158
+ return_dict_in_generate:
159
+ desc: null
160
+ value: false
161
+ forced_bos_token_id:
162
+ desc: null
163
+ value: null
164
+ forced_eos_token_id:
165
+ desc: null
166
+ value: null
167
+ remove_invalid_values:
168
+ desc: null
169
+ value: false
170
+ exponential_decay_length_penalty:
171
+ desc: null
172
+ value: null
173
+ suppress_tokens:
174
+ desc: null
175
+ value: null
176
+ begin_suppress_tokens:
177
+ desc: null
178
+ value: null
179
+ architectures:
180
+ desc: null
181
+ value:
182
+ - BertForMaskedLM
183
+ finetuning_task:
184
+ desc: null
185
+ value: null
186
+ id2label:
187
+ desc: null
188
+ value:
189
+ '0': LABEL_0
190
+ '1': LABEL_1
191
+ '2': LABEL_2
192
+ '3': LABEL_3
193
+ '4': LABEL_4
194
+ '5': LABEL_5
195
+ '6': LABEL_6
196
+ '7': LABEL_7
197
+ '8': LABEL_8
198
+ label2id:
199
+ desc: null
200
+ value:
201
+ LABEL_0: 0
202
+ LABEL_1: 1
203
+ LABEL_2: 2
204
+ LABEL_3: 3
205
+ LABEL_4: 4
206
+ LABEL_5: 5
207
+ LABEL_6: 6
208
+ LABEL_7: 7
209
+ LABEL_8: 8
210
+ tokenizer_class:
211
+ desc: null
212
+ value: null
213
+ prefix:
214
+ desc: null
215
+ value: null
216
+ bos_token_id:
217
+ desc: null
218
+ value: null
219
+ pad_token_id:
220
+ desc: null
221
+ value: 0
222
+ eos_token_id:
223
+ desc: null
224
+ value: null
225
+ sep_token_id:
226
+ desc: null
227
+ value: null
228
+ decoder_start_token_id:
229
+ desc: null
230
+ value: null
231
+ task_specific_params:
232
+ desc: null
233
+ value: null
234
+ problem_type:
235
+ desc: null
236
+ value: null
237
+ _name_or_path:
238
+ desc: null
239
+ value: ai-forever/ruBert-base
240
+ transformers_version:
241
+ desc: null
242
+ value: 4.41.2
243
+ directionality:
244
+ desc: null
245
+ value: bidi
246
+ pooler_fc_size:
247
+ desc: null
248
+ value: 768
249
+ pooler_num_attention_heads:
250
+ desc: null
251
+ value: 12
252
+ pooler_num_fc_layers:
253
+ desc: null
254
+ value: 3
255
+ pooler_size_per_head:
256
+ desc: null
257
+ value: 128
258
+ pooler_type:
259
+ desc: null
260
+ value: first_token_transform
261
+ model_type:
262
+ desc: null
263
+ value: bert
264
+ vocab_size:
265
+ desc: null
266
+ value: 120138
267
+ hidden_size:
268
+ desc: null
269
+ value: 768
270
+ num_hidden_layers:
271
+ desc: null
272
+ value: 12
273
+ num_attention_heads:
274
+ desc: null
275
+ value: 12
276
+ hidden_act:
277
+ desc: null
278
+ value: gelu
279
+ intermediate_size:
280
+ desc: null
281
+ value: 3072
282
+ hidden_dropout_prob:
283
+ desc: null
284
+ value: 0.1
285
+ attention_probs_dropout_prob:
286
+ desc: null
287
+ value: 0.1
288
+ max_position_embeddings:
289
+ desc: null
290
+ value: 512
291
+ type_vocab_size:
292
+ desc: null
293
+ value: 2
294
+ initializer_range:
295
+ desc: null
296
+ value: 0.02
297
+ layer_norm_eps:
298
+ desc: null
299
+ value: 1.0e-12
300
+ position_embedding_type:
301
+ desc: null
302
+ value: absolute
303
+ use_cache:
304
+ desc: null
305
+ value: true
306
+ classifier_dropout:
307
+ desc: null
308
+ value: null
309
+ output_dir:
310
+ desc: null
311
+ value: /kaggle/working/results
312
+ overwrite_output_dir:
313
+ desc: null
314
+ value: false
315
+ do_train:
316
+ desc: null
317
+ value: false
318
+ do_eval:
319
+ desc: null
320
+ value: true
321
+ do_predict:
322
+ desc: null
323
+ value: false
324
+ eval_strategy:
325
+ desc: null
326
+ value: epoch
327
+ prediction_loss_only:
328
+ desc: null
329
+ value: false
330
+ per_device_train_batch_size:
331
+ desc: null
332
+ value: 16
333
+ per_device_eval_batch_size:
334
+ desc: null
335
+ value: 16
336
+ per_gpu_train_batch_size:
337
+ desc: null
338
+ value: null
339
+ per_gpu_eval_batch_size:
340
+ desc: null
341
+ value: null
342
+ gradient_accumulation_steps:
343
+ desc: null
344
+ value: 1
345
+ eval_accumulation_steps:
346
+ desc: null
347
+ value: null
348
+ eval_delay:
349
+ desc: null
350
+ value: 0
351
+ learning_rate:
352
+ desc: null
353
+ value: 2.0e-05
354
+ weight_decay:
355
+ desc: null
356
+ value: 0.01
357
+ adam_beta1:
358
+ desc: null
359
+ value: 0.9
360
+ adam_beta2:
361
+ desc: null
362
+ value: 0.999
363
+ adam_epsilon:
364
+ desc: null
365
+ value: 1.0e-08
366
+ max_grad_norm:
367
+ desc: null
368
+ value: 1.0
369
+ num_train_epochs:
370
+ desc: null
371
+ value: 3
372
+ max_steps:
373
+ desc: null
374
+ value: -1
375
+ lr_scheduler_type:
376
+ desc: null
377
+ value: linear
378
+ lr_scheduler_kwargs:
379
+ desc: null
380
+ value: {}
381
+ warmup_ratio:
382
+ desc: null
383
+ value: 0.0
384
+ warmup_steps:
385
+ desc: null
386
+ value: 0
387
+ log_level:
388
+ desc: null
389
+ value: passive
390
+ log_level_replica:
391
+ desc: null
392
+ value: warning
393
+ log_on_each_node:
394
+ desc: null
395
+ value: true
396
+ logging_dir:
397
+ desc: null
398
+ value: /kaggle/working/logs
399
+ logging_strategy:
400
+ desc: null
401
+ value: steps
402
+ logging_first_step:
403
+ desc: null
404
+ value: false
405
+ logging_steps:
406
+ desc: null
407
+ value: 500
408
+ logging_nan_inf_filter:
409
+ desc: null
410
+ value: true
411
+ save_strategy:
412
+ desc: null
413
+ value: 'no'
414
+ save_steps:
415
+ desc: null
416
+ value: 500
417
+ save_total_limit:
418
+ desc: null
419
+ value: null
420
+ save_safetensors:
421
+ desc: null
422
+ value: true
423
+ save_on_each_node:
424
+ desc: null
425
+ value: false
426
+ save_only_model:
427
+ desc: null
428
+ value: false
429
+ restore_callback_states_from_checkpoint:
430
+ desc: null
431
+ value: false
432
+ no_cuda:
433
+ desc: null
434
+ value: false
435
+ use_cpu:
436
+ desc: null
437
+ value: false
438
+ use_mps_device:
439
+ desc: null
440
+ value: false
441
+ seed:
442
+ desc: null
443
+ value: 42
444
+ data_seed:
445
+ desc: null
446
+ value: null
447
+ jit_mode_eval:
448
+ desc: null
449
+ value: false
450
+ use_ipex:
451
+ desc: null
452
+ value: false
453
+ bf16:
454
+ desc: null
455
+ value: false
456
+ fp16:
457
+ desc: null
458
+ value: false
459
+ fp16_opt_level:
460
+ desc: null
461
+ value: O1
462
+ half_precision_backend:
463
+ desc: null
464
+ value: auto
465
+ bf16_full_eval:
466
+ desc: null
467
+ value: false
468
+ fp16_full_eval:
469
+ desc: null
470
+ value: false
471
+ tf32:
472
+ desc: null
473
+ value: null
474
+ local_rank:
475
+ desc: null
476
+ value: 0
477
+ ddp_backend:
478
+ desc: null
479
+ value: null
480
+ tpu_num_cores:
481
+ desc: null
482
+ value: null
483
+ tpu_metrics_debug:
484
+ desc: null
485
+ value: false
486
+ debug:
487
+ desc: null
488
+ value: []
489
+ dataloader_drop_last:
490
+ desc: null
491
+ value: false
492
+ eval_steps:
493
+ desc: null
494
+ value: null
495
+ dataloader_num_workers:
496
+ desc: null
497
+ value: 0
498
+ dataloader_prefetch_factor:
499
+ desc: null
500
+ value: null
501
+ past_index:
502
+ desc: null
503
+ value: -1
504
+ run_name:
505
+ desc: null
506
+ value: /kaggle/working/results
507
+ disable_tqdm:
508
+ desc: null
509
+ value: false
510
+ remove_unused_columns:
511
+ desc: null
512
+ value: true
513
+ label_names:
514
+ desc: null
515
+ value: null
516
+ load_best_model_at_end:
517
+ desc: null
518
+ value: false
519
+ metric_for_best_model:
520
+ desc: null
521
+ value: null
522
+ greater_is_better:
523
+ desc: null
524
+ value: null
525
+ ignore_data_skip:
526
+ desc: null
527
+ value: false
528
+ fsdp:
529
+ desc: null
530
+ value: []
531
+ fsdp_min_num_params:
532
+ desc: null
533
+ value: 0
534
+ fsdp_config:
535
+ desc: null
536
+ value:
537
+ min_num_params: 0
538
+ xla: false
539
+ xla_fsdp_v2: false
540
+ xla_fsdp_grad_ckpt: false
541
+ fsdp_transformer_layer_cls_to_wrap:
542
+ desc: null
543
+ value: null
544
+ accelerator_config:
545
+ desc: null
546
+ value:
547
+ split_batches: false
548
+ dispatch_batches: null
549
+ even_batches: true
550
+ use_seedable_sampler: true
551
+ non_blocking: false
552
+ gradient_accumulation_kwargs: null
553
+ deepspeed:
554
+ desc: null
555
+ value: null
556
+ label_smoothing_factor:
557
+ desc: null
558
+ value: 0.0
559
+ optim:
560
+ desc: null
561
+ value: adamw_torch
562
+ optim_args:
563
+ desc: null
564
+ value: null
565
+ adafactor:
566
+ desc: null
567
+ value: false
568
+ group_by_length:
569
+ desc: null
570
+ value: false
571
+ length_column_name:
572
+ desc: null
573
+ value: length
574
+ report_to:
575
+ desc: null
576
+ value:
577
+ - tensorboard
578
+ - wandb
579
+ ddp_find_unused_parameters:
580
+ desc: null
581
+ value: null
582
+ ddp_bucket_cap_mb:
583
+ desc: null
584
+ value: null
585
+ ddp_broadcast_buffers:
586
+ desc: null
587
+ value: null
588
+ dataloader_pin_memory:
589
+ desc: null
590
+ value: true
591
+ dataloader_persistent_workers:
592
+ desc: null
593
+ value: false
594
+ skip_memory_metrics:
595
+ desc: null
596
+ value: true
597
+ use_legacy_prediction_loop:
598
+ desc: null
599
+ value: false
600
+ push_to_hub:
601
+ desc: null
602
+ value: false
603
+ resume_from_checkpoint:
604
+ desc: null
605
+ value: null
606
+ hub_model_id:
607
+ desc: null
608
+ value: null
609
+ hub_strategy:
610
+ desc: null
611
+ value: every_save
612
+ hub_token:
613
+ desc: null
614
+ value: <HUB_TOKEN>
615
+ hub_private_repo:
616
+ desc: null
617
+ value: false
618
+ hub_always_push:
619
+ desc: null
620
+ value: false
621
+ gradient_checkpointing:
622
+ desc: null
623
+ value: false
624
+ gradient_checkpointing_kwargs:
625
+ desc: null
626
+ value: null
627
+ include_inputs_for_metrics:
628
+ desc: null
629
+ value: false
630
+ eval_do_concat_batches:
631
+ desc: null
632
+ value: true
633
+ fp16_backend:
634
+ desc: null
635
+ value: auto
636
+ evaluation_strategy:
637
+ desc: null
638
+ value: epoch
639
+ push_to_hub_model_id:
640
+ desc: null
641
+ value: null
642
+ push_to_hub_organization:
643
+ desc: null
644
+ value: null
645
+ push_to_hub_token:
646
+ desc: null
647
+ value: <PUSH_TO_HUB_TOKEN>
648
+ mp_parameters:
649
+ desc: null
650
+ value: ''
651
+ auto_find_batch_size:
652
+ desc: null
653
+ value: false
654
+ full_determinism:
655
+ desc: null
656
+ value: false
657
+ torchdynamo:
658
+ desc: null
659
+ value: null
660
+ ray_scope:
661
+ desc: null
662
+ value: last
663
+ ddp_timeout:
664
+ desc: null
665
+ value: 1800
666
+ torch_compile:
667
+ desc: null
668
+ value: false
669
+ torch_compile_backend:
670
+ desc: null
671
+ value: null
672
+ torch_compile_mode:
673
+ desc: null
674
+ value: null
675
+ dispatch_batches:
676
+ desc: null
677
+ value: null
678
+ split_batches:
679
+ desc: null
680
+ value: null
681
+ include_tokens_per_second:
682
+ desc: null
683
+ value: false
684
+ include_num_input_tokens_seen:
685
+ desc: null
686
+ value: false
687
+ neftune_noise_alpha:
688
+ desc: null
689
+ value: null
690
+ optim_target_modules:
691
+ desc: null
692
+ value: null
693
+ batch_eval_metrics:
694
+ desc: null
695
+ value: false
wandb/run-20240703_100700-cyvdged5/files/output.log ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ Some weights of BertForSequenceClassification were not initialized from the model checkpoint at ai-forever/ruBert-base and are newly initialized: ['classifier.bias', 'classifier.weight']
2
+ You should probably TRAIN this model on a down-stream task to be able to use it for predictions and inference.
3
+ /opt/conda/lib/python3.10/site-packages/transformers/training_args.py:1474: FutureWarning: `evaluation_strategy` is deprecated and will be removed in version 4.46 of 🤗 Transformers. Use `eval_strategy` instead
4
+ warnings.warn(
5
+ /opt/conda/lib/python3.10/site-packages/datasets/load.py:759: FutureWarning: The repository for accuracy contains custom code which must be executed to correctly load the metric. You can inspect the repository content at https://raw.githubusercontent.com/huggingface/datasets/2.19.2/metrics/accuracy/accuracy.py
6
+ You can avoid this message in future by passing the argument `trust_remote_code=True`.
7
+ Passing `trust_remote_code=True` will be mandatory to load this metric from the next major release of `datasets`.
wandb/run-20240703_100700-cyvdged5/files/requirements.txt ADDED
@@ -0,0 +1,868 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Babel==2.14.0
2
+ Boruta==0.3
3
+ Brotli==1.1.0
4
+ CVXcanon==0.1.2
5
+ Cartopy==0.23.0
6
+ Cython==3.0.8
7
+ Deprecated==1.2.14
8
+ Farama-Notifications==0.0.4
9
+ Flask==3.0.3
10
+ Geohash==1.0
11
+ GitPython==3.1.41
12
+ ImageHash==4.3.1
13
+ Janome==0.5.0
14
+ Jinja2==3.1.2
15
+ LunarCalendar==0.0.9
16
+ Mako==1.3.5
17
+ Markdown==3.5.2
18
+ MarkupSafe==2.1.3
19
+ MarkupSafe==2.1.5
20
+ Pillow==9.5.0
21
+ PuLP==2.8.0
22
+ PyArabic==0.6.15
23
+ PyJWT==2.8.0
24
+ PyMeeus==0.5.12
25
+ PySocks==1.7.1
26
+ PyUpSet==0.1.1.post7
27
+ PyWavelets==1.5.0
28
+ PyYAML==6.0.1
29
+ Pygments==2.17.2
30
+ Pympler==1.0.1
31
+ QtPy==2.4.1
32
+ Rtree==1.2.0
33
+ SQLAlchemy==2.0.25
34
+ SecretStorage==3.3.3
35
+ Send2Trash==1.8.2
36
+ Shapely==1.8.5.post1
37
+ Shimmy==1.3.0
38
+ SimpleITK==2.3.1
39
+ TPOT==0.12.1
40
+ Theano-PyMC==1.1.2
41
+ Theano==1.0.5
42
+ Wand==0.6.13
43
+ Werkzeug==3.0.3
44
+ absl-py==1.4.0
45
+ accelerate==0.30.1
46
+ access==1.1.9
47
+ affine==2.4.0
48
+ aiobotocore==2.13.0
49
+ aiofiles==22.1.0
50
+ aiohttp==3.9.1
51
+ aioitertools==0.11.0
52
+ aiorwlock==1.3.0
53
+ aiosignal==1.3.1
54
+ aiosqlite==0.19.0
55
+ albumentations==1.4.0
56
+ alembic==1.13.1
57
+ altair==5.3.0
58
+ annotated-types==0.6.0
59
+ annotated-types==0.7.0
60
+ annoy==1.17.3
61
+ anyio==4.2.0
62
+ apache-beam==2.46.0
63
+ aplus==0.11.0
64
+ appdirs==1.4.4
65
+ archspec==0.2.3
66
+ argon2-cffi-bindings==21.2.0
67
+ argon2-cffi==23.1.0
68
+ array-record==0.5.0
69
+ arrow==1.3.0
70
+ arviz==0.18.0
71
+ astroid==3.2.2
72
+ astropy-iers-data==0.2024.6.3.0.31.14
73
+ astropy==6.1.0
74
+ asttokens==2.4.1
75
+ astunparse==1.6.3
76
+ async-lru==2.0.4
77
+ async-timeout==4.0.3
78
+ attrs==23.2.0
79
+ audioread==3.0.1
80
+ autopep8==2.0.4
81
+ backoff==2.2.1
82
+ bayesian-optimization==1.4.3
83
+ beatrix_jupyterlab==2023.128.151533
84
+ beautifulsoup4==4.12.2
85
+ blake3==0.2.1
86
+ bleach==6.1.0
87
+ blessed==1.20.0
88
+ blinker==1.8.2
89
+ blis==0.7.10
90
+ blosc2==2.6.2
91
+ bokeh==3.4.1
92
+ boltons==23.1.1
93
+ boto3==1.26.100
94
+ botocore==1.34.106
95
+ bq_helper==0.4.1
96
+ bqplot==0.12.43
97
+ branca==0.7.2
98
+ brewer2mpl==1.4.1
99
+ brotlipy==0.7.0
100
+ cached-property==1.5.2
101
+ cachetools==4.2.4
102
+ cachetools==5.3.2
103
+ catalogue==2.0.10
104
+ catalyst==22.4
105
+ catboost==1.2.5
106
+ category-encoders==2.6.3
107
+ certifi==2024.2.2
108
+ cesium==0.12.1
109
+ cffi==1.16.0
110
+ charset-normalizer==3.3.2
111
+ chex==0.1.86
112
+ cleverhans==4.0.0
113
+ click-plugins==1.1.1
114
+ click==8.1.7
115
+ cligj==0.7.2
116
+ cloud-tpu-client==0.10
117
+ cloud-tpu-profiler==2.4.0
118
+ cloudpathlib==0.16.0
119
+ cloudpickle==2.2.1
120
+ cloudpickle==3.0.0
121
+ cmdstanpy==1.2.3
122
+ colorama==0.4.6
123
+ colorcet==3.1.0
124
+ colorful==0.5.6
125
+ colorlog==6.8.2
126
+ colorlover==0.3.0
127
+ comm==0.2.1
128
+ conda-libmamba-solver==23.12.0
129
+ conda-package-handling==2.2.0
130
+ conda==24.5.0
131
+ conda_package_streaming==0.9.0
132
+ confection==0.1.4
133
+ contextily==1.6.0
134
+ contourpy==1.2.0
135
+ contourpy==1.2.1
136
+ convertdate==2.4.0
137
+ crcmod==1.7
138
+ cryptography==41.0.7
139
+ cuda-python==12.5.0
140
+ cudf==24.4.1
141
+ cufflinks==0.17.3
142
+ cuml==24.4.0
143
+ cupy==13.1.0
144
+ cycler==0.12.1
145
+ cymem==2.0.8
146
+ cytoolz==0.12.3
147
+ daal4py==2024.4.0
148
+ daal==2024.4.0
149
+ dacite==1.8.1
150
+ dask-cuda==24.4.0
151
+ dask-cudf==24.4.1
152
+ dask-expr==1.1.2
153
+ dask==2024.5.2
154
+ dataclasses-json==0.6.6
155
+ dataproc_jupyter_plugin==0.1.66
156
+ datasets==2.19.2
157
+ datashader==0.16.2
158
+ datatile==1.0.3
159
+ db-dtypes==1.2.0
160
+ deap==1.4.1
161
+ debugpy==1.8.0
162
+ decorator==5.1.1
163
+ deepdiff==7.0.1
164
+ defusedxml==0.7.1
165
+ deprecation==2.1.0
166
+ descartes==1.1.0
167
+ dill==0.3.8
168
+ dipy==1.9.0
169
+ distlib==0.3.8
170
+ distributed==2024.1.1
171
+ distro==1.9.0
172
+ dm-tree==0.1.8
173
+ docker-pycreds==0.4.0
174
+ docker==7.0.0
175
+ docopt==0.6.2
176
+ docstring-parser==0.15
177
+ docstring-to-markdown==0.15
178
+ docutils==0.21.2
179
+ earthengine-api==0.1.405
180
+ easydict==1.13
181
+ easyocr==1.7.1
182
+ ecos==2.0.13
183
+ eli5==0.13.0
184
+ emoji==2.12.1
185
+ en-core-web-lg==3.7.1
186
+ en-core-web-sm==3.7.1
187
+ entrypoints==0.4
188
+ ephem==4.1.5
189
+ esda==2.5.1
190
+ essentia==2.1b6.dev1110
191
+ et-xmlfile==1.1.0
192
+ etils==1.6.0
193
+ exceptiongroup==1.2.0
194
+ executing==2.0.1
195
+ explainable-ai-sdk==1.3.3
196
+ fastai==2.7.15
197
+ fastapi==0.108.0
198
+ fastavro==1.9.3
199
+ fastcore==1.5.43
200
+ fastdownload==0.0.7
201
+ fasteners==0.19
202
+ fastjsonschema==2.19.1
203
+ fastprogress==1.0.3
204
+ fastrlock==0.8.2
205
+ fasttext==0.9.2
206
+ feather-format==0.4.1
207
+ featuretools==1.31.0
208
+ filelock==3.13.1
209
+ fiona==1.9.6
210
+ fitter==1.7.0
211
+ flake8==7.0.0
212
+ flashtext==2.7
213
+ flatbuffers==23.5.26
214
+ flax==0.8.4
215
+ folium==0.16.0
216
+ fonttools==4.47.0
217
+ fonttools==4.53.0
218
+ fqdn==1.5.1
219
+ frozendict==2.4.4
220
+ frozenlist==1.4.1
221
+ fsspec==2024.3.1
222
+ fsspec==2024.5.0
223
+ funcy==2.0
224
+ fury==0.10.0
225
+ future==1.0.0
226
+ fuzzywuzzy==0.18.0
227
+ gast==0.5.4
228
+ gatspy==0.3
229
+ gcsfs==2024.3.1
230
+ gensim==4.3.2
231
+ geographiclib==2.0
232
+ geojson==3.1.0
233
+ geopandas==0.14.4
234
+ geoplot==0.5.1
235
+ geopy==2.4.1
236
+ geoviews==1.12.0
237
+ ggplot==0.11.5
238
+ giddy==2.3.5
239
+ gitdb==4.0.11
240
+ google-ai-generativelanguage==0.6.4
241
+ google-api-core==2.11.1
242
+ google-api-core==2.19.0
243
+ google-api-python-client==2.131.0
244
+ google-apitools==0.5.31
245
+ google-auth-httplib2==0.2.0
246
+ google-auth-oauthlib==1.2.0
247
+ google-auth==2.26.1
248
+ google-cloud-aiplatform==0.6.0a1
249
+ google-cloud-artifact-registry==1.10.0
250
+ google-cloud-automl==1.0.1
251
+ google-cloud-bigquery==2.34.4
252
+ google-cloud-bigtable==1.7.3
253
+ google-cloud-core==2.4.1
254
+ google-cloud-datastore==2.19.0
255
+ google-cloud-dlp==3.14.0
256
+ google-cloud-jupyter-config==0.0.5
257
+ google-cloud-language==2.13.3
258
+ google-cloud-monitoring==2.18.0
259
+ google-cloud-pubsub==2.19.0
260
+ google-cloud-pubsublite==1.9.0
261
+ google-cloud-recommendations-ai==0.7.1
262
+ google-cloud-resource-manager==1.11.0
263
+ google-cloud-spanner==3.40.1
264
+ google-cloud-storage==1.44.0
265
+ google-cloud-translate==3.12.1
266
+ google-cloud-videointelligence==2.13.3
267
+ google-cloud-vision==2.8.0
268
+ google-crc32c==1.5.0
269
+ google-generativeai==0.6.0
270
+ google-pasta==0.2.0
271
+ google-resumable-media==2.7.0
272
+ googleapis-common-protos==1.62.0
273
+ gplearn==0.4.2
274
+ gpustat==1.0.0
275
+ gpxpy==1.6.2
276
+ graphviz==0.20.3
277
+ greenlet==3.0.3
278
+ grpc-google-iam-v1==0.12.7
279
+ grpcio-status==1.48.1
280
+ grpcio-status==1.48.2
281
+ grpcio==1.59.3
282
+ grpcio==1.60.0
283
+ gviz-api==1.10.0
284
+ gym-notices==0.0.8
285
+ gym==0.26.2
286
+ gymnasium==0.29.0
287
+ h11==0.14.0
288
+ h2o==3.46.0.2
289
+ h5netcdf==1.3.0
290
+ h5py==3.10.0
291
+ haversine==2.8.1
292
+ hdfs==2.7.3
293
+ hep-ml==0.7.2
294
+ hijri-converter==2.3.1
295
+ hmmlearn==0.3.2
296
+ holidays==0.24
297
+ holoviews==1.18.3
298
+ hpsklearn==0.1.0
299
+ html5lib==1.1
300
+ htmlmin==0.1.12
301
+ httpcore==1.0.5
302
+ httplib2==0.21.0
303
+ httptools==0.6.1
304
+ httpx==0.27.0
305
+ huggingface-hub==0.23.2
306
+ hunspell==0.5.5
307
+ hydra-slayer==0.5.0
308
+ hyperopt==0.2.7
309
+ hypertools==0.8.0
310
+ idna==3.6
311
+ igraph==0.11.5
312
+ imagecodecs==2024.6.1
313
+ imageio==2.33.1
314
+ imbalanced-learn==0.12.3
315
+ imgaug==0.4.0
316
+ importlib-metadata==6.11.0
317
+ importlib-metadata==7.0.1
318
+ importlib-resources==6.1.1
319
+ inequality==1.0.1
320
+ iniconfig==2.0.0
321
+ ipydatawidgets==4.3.5
322
+ ipykernel==6.28.0
323
+ ipyleaflet==0.19.1
324
+ ipympl==0.7.0
325
+ ipython-genutils==0.2.0
326
+ ipython-genutils==0.2.0
327
+ ipython-sql==0.5.0
328
+ ipython==8.20.0
329
+ ipyvolume==0.6.3
330
+ ipyvue==1.11.1
331
+ ipyvuetify==1.9.4
332
+ ipywebrtc==0.6.0
333
+ ipywidgets==7.7.1
334
+ isoduration==20.11.0
335
+ isort==5.13.2
336
+ isoweek==1.3.3
337
+ itsdangerous==2.2.0
338
+ jaraco.classes==3.3.0
339
+ jax-jumpy==1.0.0
340
+ jax==0.4.26
341
+ jaxlib==0.4.26.dev20240504
342
+ jedi==0.19.1
343
+ jeepney==0.8.0
344
+ jieba==0.42.1
345
+ jmespath==1.0.1
346
+ joblib==1.4.2
347
+ json5==0.9.14
348
+ jsonpatch==1.33
349
+ jsonpointer==2.4
350
+ jsonschema-specifications==2023.12.1
351
+ jsonschema==4.20.0
352
+ jupyter-console==6.6.3
353
+ jupyter-events==0.9.0
354
+ jupyter-http-over-ws==0.0.8
355
+ jupyter-leaflet==0.19.1
356
+ jupyter-lsp==1.5.1
357
+ jupyter-server-mathjax==0.2.6
358
+ jupyter-ydoc==0.2.5
359
+ jupyter_client==7.4.9
360
+ jupyter_client==8.6.0
361
+ jupyter_core==5.7.1
362
+ jupyter_server==2.12.5
363
+ jupyter_server_fileid==0.9.1
364
+ jupyter_server_proxy==4.1.0
365
+ jupyter_server_terminals==0.5.1
366
+ jupyter_server_ydoc==0.8.0
367
+ jupyterlab-lsp==5.1.0
368
+ jupyterlab-widgets==3.0.9
369
+ jupyterlab==4.2.1
370
+ jupyterlab_git==0.44.0
371
+ jupyterlab_pygments==0.3.0
372
+ jupyterlab_server==2.27.2
373
+ jupytext==1.16.0
374
+ kaggle-environments==1.14.11
375
+ kaggle==1.6.14
376
+ kagglehub==0.2.5
377
+ keras-cv==0.9.0
378
+ keras-nlp==0.12.1
379
+ keras-tuner==1.4.6
380
+ keras==3.3.3
381
+ kernels-mixer==0.0.7
382
+ keyring==24.3.0
383
+ keyrings.google-artifactregistry-auth==1.1.2
384
+ kfp-pipeline-spec==0.2.2
385
+ kfp-server-api==2.0.5
386
+ kfp==2.5.0
387
+ kiwisolver==1.4.5
388
+ kmapper==2.0.1
389
+ kmodes==0.12.2
390
+ korean-lunar-calendar==0.3.1
391
+ kornia==0.7.2
392
+ kornia_rs==0.1.3
393
+ kt-legacy==1.0.5
394
+ kubernetes==26.1.0
395
+ langcodes==3.4.0
396
+ langid==1.1.6
397
+ language_data==1.2.0
398
+ lazy_loader==0.3
399
+ learntools==0.3.4
400
+ leven==1.0.4
401
+ libclang==16.0.6
402
+ libmambapy==1.5.8
403
+ libpysal==4.9.2
404
+ librosa==0.10.2.post1
405
+ lightgbm==4.2.0
406
+ lightning-utilities==0.11.2
407
+ lime==0.2.0.1
408
+ line_profiler==4.1.3
409
+ linkify-it-py==2.0.3
410
+ llvmlite==0.41.1
411
+ llvmlite==0.42.0
412
+ lml==0.1.0
413
+ locket==1.0.0
414
+ loguru==0.7.2
415
+ lxml==5.2.2
416
+ lz4==4.3.3
417
+ mamba==1.5.8
418
+ mapclassify==2.6.1
419
+ marisa-trie==1.1.0
420
+ markdown-it-py==3.0.0
421
+ marshmallow==3.21.2
422
+ matplotlib-inline==0.1.6
423
+ matplotlib-venn==0.11.10
424
+ matplotlib==3.7.5
425
+ matplotlib==3.8.4
426
+ mccabe==0.7.0
427
+ mdit-py-plugins==0.4.0
428
+ mdurl==0.1.2
429
+ memory-profiler==0.61.0
430
+ menuinst==2.0.1
431
+ mercantile==1.2.1
432
+ mgwr==2.2.1
433
+ missingno==0.5.2
434
+ mistune==0.8.4
435
+ mizani==0.11.4
436
+ ml-dtypes==0.2.0
437
+ mlcrate==0.2.0
438
+ mlens==0.2.3
439
+ mlxtend==0.23.1
440
+ mne==1.7.0
441
+ mnist==0.2.2
442
+ momepy==0.7.0
443
+ more-itertools==10.2.0
444
+ mpld3==0.5.10
445
+ mpmath==1.3.0
446
+ msgpack==1.0.7
447
+ msgpack==1.0.8
448
+ multidict==6.0.4
449
+ multimethod==1.10
450
+ multipledispatch==1.0.0
451
+ multiprocess==0.70.16
452
+ munkres==1.1.4
453
+ murmurhash==1.0.10
454
+ mypy-extensions==1.0.0
455
+ namex==0.0.8
456
+ nb-conda-kernels==2.3.1
457
+ nb_conda==2.2.1
458
+ nbclassic==1.0.0
459
+ nbclient==0.5.13
460
+ nbclient==0.9.0
461
+ nbconvert==6.4.5
462
+ nbdime==3.2.0
463
+ nbformat==5.9.2
464
+ ndindex==1.8
465
+ nest-asyncio==1.5.8
466
+ networkx==3.2.1
467
+ nibabel==5.2.1
468
+ nilearn==0.10.4
469
+ ninja==1.11.1.1
470
+ nltk==3.2.4
471
+ nose==1.3.7
472
+ notebook==6.5.4
473
+ notebook==6.5.6
474
+ notebook_executor==0.2
475
+ notebook_shim==0.2.3
476
+ numba==0.58.1
477
+ numba==0.59.1
478
+ numexpr==2.10.0
479
+ numpy==1.26.4
480
+ nvidia-ml-py==11.495.46
481
+ nvtx==0.2.10
482
+ oauth2client==4.1.3
483
+ oauthlib==3.2.2
484
+ objsize==0.6.1
485
+ odfpy==1.4.1
486
+ olefile==0.47
487
+ onnx==1.16.1
488
+ opencensus-context==0.1.3
489
+ opencensus==0.11.4
490
+ opencv-contrib-python==4.10.0.82
491
+ opencv-python-headless==4.10.0.82
492
+ opencv-python==4.10.0.82
493
+ openpyxl==3.1.3
494
+ openslide-python==1.3.1
495
+ opentelemetry-api==1.22.0
496
+ opentelemetry-exporter-otlp-proto-common==1.22.0
497
+ opentelemetry-exporter-otlp-proto-grpc==1.22.0
498
+ opentelemetry-exporter-otlp-proto-http==1.22.0
499
+ opentelemetry-exporter-otlp==1.22.0
500
+ opentelemetry-proto==1.22.0
501
+ opentelemetry-sdk==1.22.0
502
+ opentelemetry-semantic-conventions==0.43b0
503
+ opt-einsum==3.3.0
504
+ optax==0.2.2
505
+ optree==0.11.0
506
+ optuna==3.6.1
507
+ orbax-checkpoint==0.5.15
508
+ ordered-set==4.1.0
509
+ orjson==3.9.10
510
+ ortools==9.4.1874
511
+ osmnx==1.9.3
512
+ overrides==7.4.0
513
+ packaging==21.3
514
+ pandas-datareader==0.10.0
515
+ pandas-profiling==3.6.6
516
+ pandas-summary==0.2.0
517
+ pandas==2.2.1
518
+ pandas==2.2.2
519
+ pandasql==0.7.3
520
+ pandocfilters==1.5.0
521
+ panel==1.4.4
522
+ papermill==2.5.0
523
+ param==2.1.0
524
+ parso==0.8.3
525
+ partd==1.4.2
526
+ path.py==12.5.0
527
+ path==16.14.0
528
+ pathos==0.3.2
529
+ pathy==0.10.3
530
+ patsy==0.5.6
531
+ pdf2image==1.17.0
532
+ pettingzoo==1.24.0
533
+ pexpect==4.8.0
534
+ pexpect==4.9.0
535
+ phik==0.12.4
536
+ pickleshare==0.7.5
537
+ pillow==10.3.0
538
+ pip==23.3.2
539
+ pkgutil_resolve_name==1.3.10
540
+ platformdirs==3.11.0
541
+ platformdirs==4.1.0
542
+ plotly-express==0.4.1
543
+ plotly==5.18.0
544
+ plotnine==0.13.6
545
+ pluggy==1.5.0
546
+ pointpats==2.4.0
547
+ polars==0.20.31
548
+ polyglot==16.7.4
549
+ pooch==1.8.1
550
+ pox==0.3.4
551
+ ppca==0.0.4
552
+ ppft==1.7.6.8
553
+ preprocessing==0.1.13
554
+ preshed==3.0.9
555
+ prettytable==3.9.0
556
+ progressbar2==4.4.2
557
+ prometheus-client==0.19.0
558
+ promise==2.3
559
+ prompt-toolkit==3.0.42
560
+ prompt-toolkit==3.0.43
561
+ prophet==1.1.1
562
+ proto-plus==1.23.0
563
+ protobuf==3.20.3
564
+ protobuf==4.24.4
565
+ psutil==5.9.3
566
+ psutil==5.9.7
567
+ ptyprocess==0.7.0
568
+ pudb==2024.1
569
+ pure-eval==0.2.2
570
+ py-cpuinfo==9.0.0
571
+ py-spy==0.3.14
572
+ py4j==0.10.9.7
573
+ pyLDAvis==3.4.1
574
+ pyOpenSSL==23.3.0
575
+ pyaml==24.4.0
576
+ pyarrow-hotfix==0.6
577
+ pyarrow==14.0.2
578
+ pyasn1-modules==0.3.0
579
+ pyasn1==0.5.1
580
+ pybind11==2.12.0
581
+ pyclipper==1.3.0.post5
582
+ pycodestyle==2.11.1
583
+ pycosat==0.6.6
584
+ pycparser==2.21
585
+ pycryptodome==3.20.0
586
+ pyct==0.5.0
587
+ pycuda==2024.1
588
+ pydantic==2.5.3
589
+ pydantic==2.7.2
590
+ pydantic_core==2.14.6
591
+ pydantic_core==2.18.3
592
+ pydegensac==0.1.2
593
+ pydicom==2.4.4
594
+ pydocstyle==6.3.0
595
+ pydot==1.4.2
596
+ pydub==0.25.1
597
+ pyemd==1.0.0
598
+ pyerfa==2.0.1.4
599
+ pyexcel-io==0.6.6
600
+ pyexcel-ods==0.6.0
601
+ pyflakes==3.2.0
602
+ pygltflib==1.16.2
603
+ pykalman==0.9.7
604
+ pylibraft==24.4.0
605
+ pylint==3.2.2
606
+ pymc3==3.11.4
607
+ pymongo==3.13.0
608
+ pynndescent==0.5.12
609
+ pynvjitlink==0.2.3
610
+ pynvml==11.4.1
611
+ pynvrtc==9.2
612
+ pyparsing==3.1.1
613
+ pyparsing==3.1.2
614
+ pypdf==4.2.0
615
+ pyproj==3.6.1
616
+ pysal==24.1
617
+ pyshp==2.3.1
618
+ pytesseract==0.3.10
619
+ pytest==8.2.1
620
+ python-bidi==0.4.2
621
+ python-dateutil==2.9.0.post0
622
+ python-dotenv==1.0.0
623
+ python-json-logger==2.0.7
624
+ python-louvain==0.16
625
+ python-lsp-jsonrpc==1.1.2
626
+ python-lsp-server==1.11.0
627
+ python-slugify==8.0.4
628
+ python-utils==3.8.2
629
+ pythreejs==2.4.2
630
+ pytoolconfig==1.3.1
631
+ pytools==2024.1.3
632
+ pytorch-ignite==0.5.0.post2
633
+ pytorch-lightning==2.2.5
634
+ pytz==2023.3.post1
635
+ pytz==2024.1
636
+ pyu2f==0.1.5
637
+ pyviz_comms==3.0.2
638
+ pyzmq==24.0.1
639
+ pyzmq==25.1.2
640
+ qgrid==1.3.1
641
+ qtconsole==5.5.2
642
+ quantecon==0.7.2
643
+ qudida==0.0.4
644
+ raft-dask==24.4.0
645
+ rapids-dask-dependency==24.4.1a0
646
+ rasterio==1.3.10
647
+ rasterstats==0.19.0
648
+ ray-cpp==2.9.0
649
+ ray==2.9.0
650
+ referencing==0.32.1
651
+ regex==2023.12.25
652
+ requests-oauthlib==1.3.1
653
+ requests-toolbelt==0.10.1
654
+ requests==2.32.3
655
+ retrying==1.3.3
656
+ retrying==1.3.4
657
+ rfc3339-validator==0.1.4
658
+ rfc3986-validator==0.1.1
659
+ rgf-python==3.12.0
660
+ rich-click==1.8.2
661
+ rich==13.7.0
662
+ rich==13.7.1
663
+ rmm==24.4.0
664
+ rope==1.13.0
665
+ rpds-py==0.16.2
666
+ rsa==4.9
667
+ ruamel-yaml-conda==0.15.100
668
+ ruamel.yaml.clib==0.2.7
669
+ ruamel.yaml==0.18.5
670
+ s2sphere==0.2.5
671
+ s3fs==2024.3.1
672
+ s3transfer==0.6.2
673
+ safetensors==0.4.3
674
+ scattertext==0.1.19
675
+ scikit-image==0.22.0
676
+ scikit-learn-intelex==2024.4.0
677
+ scikit-learn==1.2.2
678
+ scikit-multilearn==0.2.0
679
+ scikit-optimize==0.10.1
680
+ scikit-plot==0.3.7
681
+ scikit-surprise==1.1.4
682
+ scipy==1.11.4
683
+ scipy==1.13.1
684
+ seaborn==0.12.2
685
+ segment_anything==1.0
686
+ segregation==2.5
687
+ semver==3.0.2
688
+ sentencepiece==0.2.0
689
+ sentry-sdk==2.3.1
690
+ setproctitle==1.3.3
691
+ setuptools-git==1.2
692
+ setuptools-scm==8.1.0
693
+ setuptools==69.0.3
694
+ shap==0.44.1
695
+ shapely==2.0.4
696
+ shellingham==1.5.4
697
+ simpervisor==1.0.0
698
+ simplejson==3.19.2
699
+ six==1.16.0
700
+ sklearn-pandas==2.2.0
701
+ slicer==0.0.7
702
+ smart-open==6.4.0
703
+ smmap==5.0.1
704
+ sniffio==1.3.0
705
+ snowballstemmer==2.2.0
706
+ snuggs==1.4.7
707
+ sortedcontainers==2.4.0
708
+ soundfile==0.12.1
709
+ soupsieve==2.5
710
+ soxr==0.3.7
711
+ spacy-legacy==3.0.12
712
+ spacy-loggers==1.0.5
713
+ spacy==3.7.3
714
+ spaghetti==1.7.5.post1
715
+ spectral==0.23.1
716
+ spglm==1.1.0
717
+ sphinx-rtd-theme==0.2.4
718
+ spint==1.0.7
719
+ splot==1.1.5.post1
720
+ spopt==0.6.0
721
+ spreg==1.4.2
722
+ spvcm==0.3.0
723
+ sqlparse==0.4.4
724
+ squarify==0.4.3
725
+ srsly==2.4.8
726
+ stable-baselines3==2.1.0
727
+ stack-data==0.6.2
728
+ stack-data==0.6.3
729
+ stanio==0.5.0
730
+ starlette==0.32.0.post1
731
+ statsmodels==0.14.1
732
+ stemming==1.0.1
733
+ stop-words==2018.7.23
734
+ stopit==1.1.2
735
+ stumpy==1.12.0
736
+ sympy==1.12.1
737
+ tables==3.9.2
738
+ tabulate==0.9.0
739
+ tangled-up-in-unicode==0.2.0
740
+ tbb==2021.12.0
741
+ tblib==3.0.0
742
+ tenacity==8.2.3
743
+ tensorboard-data-server==0.7.2
744
+ tensorboard-plugin-profile==2.15.0
745
+ tensorboard==2.15.1
746
+ tensorboardX==2.6.2.2
747
+ tensorflow-cloud==0.1.16
748
+ tensorflow-datasets==4.9.4
749
+ tensorflow-decision-forests==1.8.1
750
+ tensorflow-estimator==2.15.0
751
+ tensorflow-hub==0.16.1
752
+ tensorflow-io-gcs-filesystem==0.35.0
753
+ tensorflow-io==0.35.0
754
+ tensorflow-metadata==0.14.0
755
+ tensorflow-probability==0.23.0
756
+ tensorflow-serving-api==2.14.1
757
+ tensorflow-text==2.15.0
758
+ tensorflow-transform==0.14.0
759
+ tensorflow==2.15.0
760
+ tensorstore==0.1.60
761
+ termcolor==2.4.0
762
+ terminado==0.18.0
763
+ testpath==0.6.0
764
+ text-unidecode==1.3
765
+ textblob==0.18.0.post0
766
+ texttable==1.7.0
767
+ tf_keras==2.15.1
768
+ tfp-nightly==0.24.0.dev0
769
+ thinc==8.2.3
770
+ threadpoolctl==3.2.0
771
+ tifffile==2023.12.9
772
+ timm==1.0.3
773
+ tinycss2==1.2.1
774
+ tobler==0.11.2
775
+ tokenizers==0.19.1
776
+ toml==0.10.2
777
+ tomli==2.0.1
778
+ tomlkit==0.12.5
779
+ toolz==0.12.1
780
+ torch==2.1.2
781
+ torchaudio==2.1.2
782
+ torchdata==0.7.1
783
+ torchinfo==1.8.0
784
+ torchmetrics==1.4.0.post0
785
+ torchtext==0.16.2
786
+ torchvision==0.16.2
787
+ tornado==6.3.3
788
+ tqdm==4.66.4
789
+ traceml==1.0.8
790
+ traitlets==5.9.0
791
+ traittypes==0.2.1
792
+ transformers==4.41.2
793
+ treelite==4.1.2
794
+ truststore==0.8.0
795
+ trx-python==0.2.9
796
+ tsfresh==0.20.2
797
+ typeguard==4.1.5
798
+ typer==0.9.0
799
+ typer==0.9.4
800
+ types-python-dateutil==2.8.19.20240106
801
+ typing-inspect==0.9.0
802
+ typing-utils==0.1.0
803
+ typing_extensions==4.9.0
804
+ tzdata==2023.4
805
+ tzdata==2024.1
806
+ uc-micro-py==1.0.3
807
+ ucx-py==0.37.0
808
+ ujson==5.10.0
809
+ umap-learn==0.5.6
810
+ unicodedata2==15.1.0
811
+ update-checker==0.18.0
812
+ uri-template==1.3.0
813
+ uritemplate==3.0.1
814
+ urllib3==1.26.18
815
+ urllib3==2.1.0
816
+ urwid==2.6.12
817
+ urwid_readline==0.14
818
+ uvicorn==0.25.0
819
+ uvloop==0.19.0
820
+ vaex-astro==0.9.3
821
+ vaex-core==4.17.1
822
+ vaex-hdf5==0.14.1
823
+ vaex-jupyter==0.8.2
824
+ vaex-ml==0.18.3
825
+ vaex-server==0.9.0
826
+ vaex-viz==0.5.4
827
+ vaex==4.17.0
828
+ vec_noise==1.1.4
829
+ vecstack==0.4.0
830
+ virtualenv==20.21.0
831
+ visions==0.7.5
832
+ vowpalwabbit==9.9.0
833
+ vtk==9.3.0
834
+ wandb==0.17.0
835
+ wasabi==1.1.2
836
+ watchfiles==0.21.0
837
+ wavio==0.0.9
838
+ wcwidth==0.2.13
839
+ weasel==0.3.4
840
+ webcolors==1.13
841
+ webencodings==0.5.1
842
+ websocket-client==1.7.0
843
+ websockets==12.0
844
+ wfdb==4.1.2
845
+ whatthepatch==1.0.5
846
+ wheel==0.42.0
847
+ widgetsnbextension==3.6.6
848
+ witwidget==1.8.1
849
+ woodwork==0.31.0
850
+ wordcloud==1.9.3
851
+ wordsegment==1.3.1
852
+ wrapt==1.14.1
853
+ xarray-einstats==0.7.0
854
+ xarray==2024.5.0
855
+ xgboost==2.0.3
856
+ xvfbwrapper==0.2.9
857
+ xxhash==3.4.1
858
+ xyzservices==2024.4.0
859
+ y-py==0.6.2
860
+ yapf==0.40.2
861
+ yarl==1.9.3
862
+ yarl==1.9.4
863
+ ydata-profiling==4.6.4
864
+ yellowbrick==1.5
865
+ ypy-websocket==0.8.4
866
+ zict==3.0.0
867
+ zipp==3.17.0
868
+ zstandard==0.19.0
wandb/run-20240703_100700-cyvdged5/files/wandb-metadata.json ADDED
@@ -0,0 +1,62 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.15.133+-x86_64-with-glibc2.31",
3
+ "python": "3.10.13",
4
+ "heartbeatAt": "2024-07-03T10:07:01.631454",
5
+ "startedAt": "2024-07-03T10:07:00.445303",
6
+ "docker": null,
7
+ "cuda": null,
8
+ "args": [],
9
+ "state": "running",
10
+ "program": "kaggle.ipynb",
11
+ "codePathLocal": null,
12
+ "root": "/kaggle/working",
13
+ "host": "2679035f67af",
14
+ "username": "root",
15
+ "executable": "/opt/conda/bin/python3.10",
16
+ "cpu_count": 2,
17
+ "cpu_count_logical": 4,
18
+ "cpu_freq": {
19
+ "current": 2000.21,
20
+ "min": 0.0,
21
+ "max": 0.0
22
+ },
23
+ "cpu_freq_per_core": [
24
+ {
25
+ "current": 2000.21,
26
+ "min": 0.0,
27
+ "max": 0.0
28
+ },
29
+ {
30
+ "current": 2000.21,
31
+ "min": 0.0,
32
+ "max": 0.0
33
+ },
34
+ {
35
+ "current": 2000.21,
36
+ "min": 0.0,
37
+ "max": 0.0
38
+ },
39
+ {
40
+ "current": 2000.21,
41
+ "min": 0.0,
42
+ "max": 0.0
43
+ }
44
+ ],
45
+ "disk": {
46
+ "/": {
47
+ "total": 8062.387607574463,
48
+ "used": 5690.048397064209
49
+ }
50
+ },
51
+ "gpu": "Tesla P100-PCIE-16GB",
52
+ "gpu_count": 1,
53
+ "gpu_devices": [
54
+ {
55
+ "name": "Tesla P100-PCIE-16GB",
56
+ "memory_total": 17179869184
57
+ }
58
+ ],
59
+ "memory": {
60
+ "total": 31.357555389404297
61
+ }
62
+ }
wandb/run-20240703_100700-cyvdged5/logs/debug-internal.log ADDED
@@ -0,0 +1,300 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-07-03 10:07:00,452 INFO StreamThr :153 [internal.py:wandb_internal():85] W&B internal server running at pid: 153, started at: 2024-07-03 10:07:00.451696
2
+ 2024-07-03 10:07:00,453 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status
3
+ 2024-07-03 10:07:00,954 INFO WriterThread:153 [datastore.py:open_for_write():87] open: /kaggle/working/wandb/run-20240703_100700-cyvdged5/run-cyvdged5.wandb
4
+ 2024-07-03 10:07:00,955 DEBUG SenderThread:153 [sender.py:send():378] send: header
5
+ 2024-07-03 10:07:00,958 DEBUG SenderThread:153 [sender.py:send():378] send: run
6
+ 2024-07-03 10:07:01,527 INFO SenderThread:153 [dir_watcher.py:__init__():211] watching files in: /kaggle/working/wandb/run-20240703_100700-cyvdged5/files
7
+ 2024-07-03 10:07:01,527 INFO SenderThread:153 [sender.py:_start_run_threads():1123] run started: cyvdged5 with start time 1720001220.452695
8
+ 2024-07-03 10:07:01,537 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: check_version
9
+ 2024-07-03 10:07:01,537 DEBUG SenderThread:153 [sender.py:send_request():405] send_request: check_version
10
+ 2024-07-03 10:07:01,607 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: run_start
11
+ 2024-07-03 10:07:01,618 DEBUG HandlerThread:153 [system_info.py:__init__():26] System info init
12
+ 2024-07-03 10:07:01,618 DEBUG HandlerThread:153 [system_info.py:__init__():41] System info init done
13
+ 2024-07-03 10:07:01,618 INFO HandlerThread:153 [system_monitor.py:start():194] Starting system monitor
14
+ 2024-07-03 10:07:01,618 INFO SystemMonitor:153 [system_monitor.py:_start():158] Starting system asset monitoring threads
15
+ 2024-07-03 10:07:01,618 INFO HandlerThread:153 [system_monitor.py:probe():214] Collecting system info
16
+ 2024-07-03 10:07:01,619 INFO SystemMonitor:153 [interfaces.py:start():188] Started cpu monitoring
17
+ 2024-07-03 10:07:01,620 INFO SystemMonitor:153 [interfaces.py:start():188] Started disk monitoring
18
+ 2024-07-03 10:07:01,621 INFO SystemMonitor:153 [interfaces.py:start():188] Started gpu monitoring
19
+ 2024-07-03 10:07:01,622 INFO SystemMonitor:153 [interfaces.py:start():188] Started memory monitoring
20
+ 2024-07-03 10:07:01,623 INFO SystemMonitor:153 [interfaces.py:start():188] Started network monitoring
21
+ 2024-07-03 10:07:01,631 DEBUG HandlerThread:153 [system_info.py:probe():150] Probing system
22
+ 2024-07-03 10:07:01,633 DEBUG HandlerThread:153 [gitlib.py:_init_repo():56] git repository is invalid
23
+ 2024-07-03 10:07:01,633 DEBUG HandlerThread:153 [system_info.py:probe():198] Probing system done
24
+ 2024-07-03 10:07:01,633 DEBUG HandlerThread:153 [system_monitor.py:probe():223] {'os': 'Linux-5.15.133+-x86_64-with-glibc2.31', 'python': '3.10.13', 'heartbeatAt': '2024-07-03T10:07:01.631454', 'startedAt': '2024-07-03T10:07:00.445303', 'docker': None, 'cuda': None, 'args': (), 'state': 'running', 'program': 'kaggle.ipynb', 'codePathLocal': None, 'root': '/kaggle/working', 'host': '2679035f67af', 'username': 'root', 'executable': '/opt/conda/bin/python3.10', 'cpu_count': 2, 'cpu_count_logical': 4, 'cpu_freq': {'current': 2000.21, 'min': 0.0, 'max': 0.0}, 'cpu_freq_per_core': [{'current': 2000.21, 'min': 0.0, 'max': 0.0}, {'current': 2000.21, 'min': 0.0, 'max': 0.0}, {'current': 2000.21, 'min': 0.0, 'max': 0.0}, {'current': 2000.21, 'min': 0.0, 'max': 0.0}], 'disk': {'/': {'total': 8062.387607574463, 'used': 5690.048397064209}}, 'gpu': 'Tesla P100-PCIE-16GB', 'gpu_count': 1, 'gpu_devices': [{'name': 'Tesla P100-PCIE-16GB', 'memory_total': 17179869184}], 'memory': {'total': 31.357555389404297}}
25
+ 2024-07-03 10:07:01,633 INFO HandlerThread:153 [system_monitor.py:probe():224] Finished collecting system info
26
+ 2024-07-03 10:07:01,633 INFO HandlerThread:153 [system_monitor.py:probe():227] Publishing system info
27
+ 2024-07-03 10:07:01,633 DEBUG HandlerThread:153 [system_info.py:_save_conda():207] Saving list of conda packages installed into the current environment
28
+ 2024-07-03 10:07:02,529 INFO Thread-12 :153 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240703_100700-cyvdged5/files/conda-environment.yaml
29
+ 2024-07-03 10:07:16,647 ERROR HandlerThread:153 [system_info.py:_save_conda():221] Error saving conda packages: Command '['conda', 'env', 'export']' timed out after 15 seconds
30
+ Traceback (most recent call last):
31
+ File "/opt/conda/lib/python3.10/site-packages/wandb/sdk/internal/system/system_info.py", line 214, in _save_conda
32
+ subprocess.call(
33
+ File "/opt/conda/lib/python3.10/subprocess.py", line 347, in call
34
+ return p.wait(timeout=timeout)
35
+ File "/opt/conda/lib/python3.10/subprocess.py", line 1209, in wait
36
+ return self._wait(timeout=timeout)
37
+ File "/opt/conda/lib/python3.10/subprocess.py", line 1951, in _wait
38
+ raise TimeoutExpired(self.args, timeout)
39
+ subprocess.TimeoutExpired: Command '['conda', 'env', 'export']' timed out after 15 seconds
40
+ 2024-07-03 10:07:16,649 DEBUG HandlerThread:153 [system_info.py:_save_conda():222] Saving conda packages done
41
+ 2024-07-03 10:07:16,650 INFO HandlerThread:153 [system_monitor.py:probe():229] Finished publishing system info
42
+ 2024-07-03 10:07:16,658 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
43
+ 2024-07-03 10:07:16,658 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: keepalive
44
+ 2024-07-03 10:07:16,658 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
45
+ 2024-07-03 10:07:16,658 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: keepalive
46
+ 2024-07-03 10:07:16,658 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
47
+ 2024-07-03 10:07:16,658 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: keepalive
48
+ 2024-07-03 10:07:16,659 DEBUG SenderThread:153 [sender.py:send():378] send: files
49
+ 2024-07-03 10:07:16,659 INFO SenderThread:153 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now
50
+ 2024-07-03 10:07:17,436 INFO wandb-upload_0:153 [upload_job.py:push():130] Uploaded file /tmp/tmpftsekjz4wandb/0br4hl77-wandb-metadata.json
51
+ 2024-07-03 10:07:17,532 INFO Thread-12 :153 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240703_100700-cyvdged5/files/wandb-metadata.json
52
+ 2024-07-03 10:07:17,602 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: python_packages
53
+ 2024-07-03 10:07:17,603 DEBUG SenderThread:153 [sender.py:send_request():405] send_request: python_packages
54
+ 2024-07-03 10:07:17,605 DEBUG SenderThread:153 [sender.py:send():378] send: telemetry
55
+ 2024-07-03 10:07:17,606 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: stop_status
56
+ 2024-07-03 10:07:17,607 DEBUG SenderThread:153 [sender.py:send_request():405] send_request: stop_status
57
+ 2024-07-03 10:07:17,901 DEBUG SenderThread:153 [sender.py:send():378] send: config
58
+ 2024-07-03 10:07:17,903 DEBUG SenderThread:153 [sender.py:send():378] send: metric
59
+ 2024-07-03 10:07:17,904 DEBUG SenderThread:153 [sender.py:send():378] send: telemetry
60
+ 2024-07-03 10:07:17,904 DEBUG SenderThread:153 [sender.py:send():378] send: telemetry
61
+ 2024-07-03 10:07:17,904 DEBUG SenderThread:153 [sender.py:send():378] send: metric
62
+ 2024-07-03 10:07:17,904 WARNING SenderThread:153 [sender.py:send_metric():1340] Seen metric with glob (shouldn't happen)
63
+ 2024-07-03 10:07:17,904 DEBUG SenderThread:153 [sender.py:send():378] send: telemetry
64
+ 2024-07-03 10:07:18,532 INFO Thread-12 :153 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240703_100700-cyvdged5/files/requirements.txt
65
+ 2024-07-03 10:07:20,905 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
66
+ 2024-07-03 10:07:25,906 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
67
+ 2024-07-03 10:07:30,914 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
68
+ 2024-07-03 10:07:31,536 INFO Thread-12 :153 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240703_100700-cyvdged5/files/config.yaml
69
+ 2024-07-03 10:07:32,603 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: stop_status
70
+ 2024-07-03 10:07:32,604 DEBUG SenderThread:153 [sender.py:send_request():405] send_request: stop_status
71
+ 2024-07-03 10:07:36,885 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
72
+ 2024-07-03 10:07:41,885 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
73
+ 2024-07-03 10:07:46,886 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
74
+ 2024-07-03 10:07:47,603 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: stop_status
75
+ 2024-07-03 10:07:47,604 DEBUG SenderThread:153 [sender.py:send_request():405] send_request: stop_status
76
+ 2024-07-03 10:07:52,842 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
77
+ 2024-07-03 10:07:57,843 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
78
+ 2024-07-03 10:08:01,623 DEBUG SystemMonitor:153 [system_monitor.py:_start():172] Starting system metrics aggregation loop
79
+ 2024-07-03 10:08:01,624 DEBUG SenderThread:153 [sender.py:send():378] send: stats
80
+ 2024-07-03 10:08:02,604 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: stop_status
81
+ 2024-07-03 10:08:02,604 DEBUG SenderThread:153 [sender.py:send_request():405] send_request: stop_status
82
+ 2024-07-03 10:08:02,903 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
83
+ 2024-07-03 10:08:07,904 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
84
+ 2024-07-03 10:08:12,905 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
85
+ 2024-07-03 10:08:17,604 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: stop_status
86
+ 2024-07-03 10:08:17,604 DEBUG SenderThread:153 [sender.py:send_request():405] send_request: stop_status
87
+ 2024-07-03 10:08:18,887 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
88
+ 2024-07-03 10:08:23,888 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
89
+ 2024-07-03 10:08:28,889 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
90
+ 2024-07-03 10:08:31,625 DEBUG SenderThread:153 [sender.py:send():378] send: stats
91
+ 2024-07-03 10:08:32,604 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: stop_status
92
+ 2024-07-03 10:08:32,605 DEBUG SenderThread:153 [sender.py:send_request():405] send_request: stop_status
93
+ 2024-07-03 10:08:34,859 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
94
+ 2024-07-03 10:08:39,860 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
95
+ 2024-07-03 10:08:44,861 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
96
+ 2024-07-03 10:08:47,604 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: stop_status
97
+ 2024-07-03 10:08:47,605 DEBUG SenderThread:153 [sender.py:send_request():405] send_request: stop_status
98
+ 2024-07-03 10:08:49,909 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
99
+ 2024-07-03 10:08:54,910 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
100
+ 2024-07-03 10:08:59,911 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
101
+ 2024-07-03 10:09:01,626 DEBUG SenderThread:153 [sender.py:send():378] send: stats
102
+ 2024-07-03 10:09:02,604 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: stop_status
103
+ 2024-07-03 10:09:02,605 DEBUG SenderThread:153 [sender.py:send_request():405] send_request: stop_status
104
+ 2024-07-03 10:09:04,982 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
105
+ 2024-07-03 10:09:09,982 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
106
+ 2024-07-03 10:09:14,983 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
107
+ 2024-07-03 10:09:17,605 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: stop_status
108
+ 2024-07-03 10:09:17,605 DEBUG SenderThread:153 [sender.py:send_request():405] send_request: stop_status
109
+ 2024-07-03 10:09:20,857 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
110
+ 2024-07-03 10:09:25,858 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
111
+ 2024-07-03 10:09:30,859 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
112
+ 2024-07-03 10:09:31,627 DEBUG SenderThread:153 [sender.py:send():378] send: stats
113
+ 2024-07-03 10:09:32,604 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: stop_status
114
+ 2024-07-03 10:09:32,605 DEBUG SenderThread:153 [sender.py:send_request():405] send_request: stop_status
115
+ 2024-07-03 10:09:36,837 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
116
+ 2024-07-03 10:09:41,838 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
117
+ 2024-07-03 10:09:46,839 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
118
+ 2024-07-03 10:09:47,657 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: stop_status
119
+ 2024-07-03 10:09:47,657 DEBUG SenderThread:153 [sender.py:send_request():405] send_request: stop_status
120
+ 2024-07-03 10:09:51,939 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
121
+ 2024-07-03 10:09:56,940 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
122
+ 2024-07-03 10:10:01,627 DEBUG SenderThread:153 [sender.py:send():378] send: stats
123
+ 2024-07-03 10:10:02,628 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
124
+ 2024-07-03 10:10:02,737 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: stop_status
125
+ 2024-07-03 10:10:02,737 DEBUG SenderThread:153 [sender.py:send_request():405] send_request: stop_status
126
+ 2024-07-03 10:10:07,966 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
127
+ 2024-07-03 10:10:12,967 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
128
+ 2024-07-03 10:10:17,737 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: stop_status
129
+ 2024-07-03 10:10:17,737 DEBUG SenderThread:153 [sender.py:send_request():405] send_request: stop_status
130
+ 2024-07-03 10:10:18,023 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
131
+ 2024-07-03 10:10:23,024 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
132
+ 2024-07-03 10:10:28,026 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
133
+ 2024-07-03 10:10:31,628 DEBUG SenderThread:153 [sender.py:send():378] send: stats
134
+ 2024-07-03 10:10:32,946 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: stop_status
135
+ 2024-07-03 10:10:32,946 DEBUG SenderThread:153 [sender.py:send_request():405] send_request: stop_status
136
+ 2024-07-03 10:10:33,251 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
137
+ 2024-07-03 10:10:33,327 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: pause
138
+ 2024-07-03 10:10:33,328 INFO HandlerThread:153 [handler.py:handle_request_pause():724] stopping system metrics thread
139
+ 2024-07-03 10:10:33,328 INFO HandlerThread:153 [system_monitor.py:finish():203] Stopping system monitor
140
+ 2024-07-03 10:10:33,328 DEBUG SystemMonitor:153 [system_monitor.py:_start():179] Finished system metrics aggregation loop
141
+ 2024-07-03 10:10:33,328 DEBUG SystemMonitor:153 [system_monitor.py:_start():183] Publishing last batch of metrics
142
+ 2024-07-03 10:10:33,333 INFO HandlerThread:153 [interfaces.py:finish():200] Joined cpu monitor
143
+ 2024-07-03 10:10:33,334 INFO HandlerThread:153 [interfaces.py:finish():200] Joined disk monitor
144
+ 2024-07-03 10:10:33,336 INFO HandlerThread:153 [interfaces.py:finish():200] Joined gpu monitor
145
+ 2024-07-03 10:10:33,336 INFO HandlerThread:153 [interfaces.py:finish():200] Joined memory monitor
146
+ 2024-07-03 10:10:33,336 INFO HandlerThread:153 [interfaces.py:finish():200] Joined network monitor
147
+ 2024-07-03 10:10:33,337 DEBUG SenderThread:153 [sender.py:send():378] send: stats
148
+ 2024-07-03 10:10:38,338 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
149
+ 2024-07-03 10:10:43,339 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
150
+ 2024-07-03 10:10:47,746 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: stop_status
151
+ 2024-07-03 10:10:47,747 DEBUG SenderThread:153 [sender.py:send_request():405] send_request: stop_status
152
+ 2024-07-03 10:10:49,049 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
153
+ 2024-07-03 10:10:54,050 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
154
+ 2024-07-03 10:10:59,051 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
155
+ 2024-07-03 10:11:02,747 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: stop_status
156
+ 2024-07-03 10:11:02,747 DEBUG SenderThread:153 [sender.py:send_request():405] send_request: stop_status
157
+ 2024-07-03 10:11:05,034 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
158
+ 2024-07-03 10:11:10,034 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
159
+ 2024-07-03 10:11:15,036 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
160
+ 2024-07-03 10:11:17,747 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: stop_status
161
+ 2024-07-03 10:11:17,747 DEBUG SenderThread:153 [sender.py:send_request():405] send_request: stop_status
162
+ 2024-07-03 10:11:20,051 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
163
+ 2024-07-03 10:11:25,052 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
164
+ 2024-07-03 10:11:30,053 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
165
+ 2024-07-03 10:11:32,747 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: stop_status
166
+ 2024-07-03 10:11:32,747 DEBUG SenderThread:153 [sender.py:send_request():405] send_request: stop_status
167
+ 2024-07-03 10:11:35,993 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
168
+ 2024-07-03 10:11:40,994 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
169
+ 2024-07-03 10:11:45,995 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
170
+ 2024-07-03 10:11:47,747 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: stop_status
171
+ 2024-07-03 10:11:47,747 DEBUG SenderThread:153 [sender.py:send_request():405] send_request: stop_status
172
+ 2024-07-03 10:11:51,968 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
173
+ 2024-07-03 10:11:56,970 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
174
+ 2024-07-03 10:12:01,971 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
175
+ 2024-07-03 10:12:02,747 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: stop_status
176
+ 2024-07-03 10:12:02,748 DEBUG SenderThread:153 [sender.py:send_request():405] send_request: stop_status
177
+ 2024-07-03 10:12:07,023 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
178
+ 2024-07-03 10:12:12,024 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
179
+ 2024-07-03 10:12:17,025 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
180
+ 2024-07-03 10:12:17,747 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: stop_status
181
+ 2024-07-03 10:12:17,748 DEBUG SenderThread:153 [sender.py:send_request():405] send_request: stop_status
182
+ 2024-07-03 10:12:22,062 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
183
+ 2024-07-03 10:12:27,064 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
184
+ 2024-07-03 10:12:28,052 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: resume
185
+ 2024-07-03 10:12:28,052 INFO HandlerThread:153 [handler.py:handle_request_resume():715] starting system metrics thread
186
+ 2024-07-03 10:12:28,052 INFO HandlerThread:153 [system_monitor.py:start():194] Starting system monitor
187
+ 2024-07-03 10:12:28,052 INFO SystemMonitor:153 [system_monitor.py:_start():158] Starting system asset monitoring threads
188
+ 2024-07-03 10:12:28,053 INFO SystemMonitor:153 [interfaces.py:start():188] Started cpu monitoring
189
+ 2024-07-03 10:12:28,054 INFO SystemMonitor:153 [interfaces.py:start():188] Started disk monitoring
190
+ 2024-07-03 10:12:28,056 INFO SystemMonitor:153 [interfaces.py:start():188] Started gpu monitoring
191
+ 2024-07-03 10:12:28,056 INFO SystemMonitor:153 [interfaces.py:start():188] Started memory monitoring
192
+ 2024-07-03 10:12:28,057 INFO SystemMonitor:153 [interfaces.py:start():188] Started network monitoring
193
+ 2024-07-03 10:12:30,640 INFO Thread-12 :153 [dir_watcher.py:_on_file_created():271] file/dir created: /kaggle/working/wandb/run-20240703_100700-cyvdged5/files/output.log
194
+ 2024-07-03 10:12:31,154 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: pause
195
+ 2024-07-03 10:12:31,154 INFO HandlerThread:153 [handler.py:handle_request_pause():724] stopping system metrics thread
196
+ 2024-07-03 10:12:31,154 INFO HandlerThread:153 [system_monitor.py:finish():203] Stopping system monitor
197
+ 2024-07-03 10:12:31,154 DEBUG SystemMonitor:153 [system_monitor.py:_start():172] Starting system metrics aggregation loop
198
+ 2024-07-03 10:12:31,154 DEBUG SystemMonitor:153 [system_monitor.py:_start():179] Finished system metrics aggregation loop
199
+ 2024-07-03 10:12:31,155 DEBUG SystemMonitor:153 [system_monitor.py:_start():183] Publishing last batch of metrics
200
+ 2024-07-03 10:12:31,156 INFO HandlerThread:153 [interfaces.py:finish():200] Joined cpu monitor
201
+ 2024-07-03 10:12:31,156 INFO HandlerThread:153 [interfaces.py:finish():200] Joined disk monitor
202
+ 2024-07-03 10:12:31,161 INFO HandlerThread:153 [interfaces.py:finish():200] Joined gpu monitor
203
+ 2024-07-03 10:12:31,161 INFO HandlerThread:153 [interfaces.py:finish():200] Joined memory monitor
204
+ 2024-07-03 10:12:31,161 INFO HandlerThread:153 [interfaces.py:finish():200] Joined network monitor
205
+ 2024-07-03 10:12:31,162 DEBUG SenderThread:153 [sender.py:send():378] send: stats
206
+ 2024-07-03 10:12:32,162 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
207
+ 2024-07-03 10:12:32,641 INFO Thread-12 :153 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240703_100700-cyvdged5/files/output.log
208
+ 2024-07-03 10:12:32,747 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: stop_status
209
+ 2024-07-03 10:12:32,748 DEBUG SenderThread:153 [sender.py:send_request():405] send_request: stop_status
210
+ 2024-07-03 10:12:36,070 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: resume
211
+ 2024-07-03 10:12:36,070 INFO HandlerThread:153 [handler.py:handle_request_resume():715] starting system metrics thread
212
+ 2024-07-03 10:12:36,070 INFO HandlerThread:153 [system_monitor.py:start():194] Starting system monitor
213
+ 2024-07-03 10:12:36,071 INFO SystemMonitor:153 [system_monitor.py:_start():158] Starting system asset monitoring threads
214
+ 2024-07-03 10:12:36,071 INFO SystemMonitor:153 [interfaces.py:start():188] Started cpu monitoring
215
+ 2024-07-03 10:12:36,073 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: pause
216
+ 2024-07-03 10:12:36,074 INFO HandlerThread:153 [handler.py:handle_request_pause():724] stopping system metrics thread
217
+ 2024-07-03 10:12:36,074 INFO HandlerThread:153 [system_monitor.py:finish():203] Stopping system monitor
218
+ 2024-07-03 10:12:36,074 INFO SystemMonitor:153 [interfaces.py:start():188] Started disk monitoring
219
+ 2024-07-03 10:12:36,074 INFO HandlerThread:153 [interfaces.py:finish():200] Joined cpu monitor
220
+ 2024-07-03 10:12:36,075 DEBUG SystemMonitor:153 [system_monitor.py:_start():172] Starting system metrics aggregation loop
221
+ 2024-07-03 10:12:36,075 DEBUG SystemMonitor:153 [system_monitor.py:_start():179] Finished system metrics aggregation loop
222
+ 2024-07-03 10:12:36,075 DEBUG SystemMonitor:153 [system_monitor.py:_start():183] Publishing last batch of metrics
223
+ 2024-07-03 10:12:36,077 INFO HandlerThread:153 [interfaces.py:finish():200] Joined disk monitor
224
+ 2024-07-03 10:12:36,078 DEBUG SenderThread:153 [sender.py:send():378] send: stats
225
+ 2024-07-03 10:12:38,078 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
226
+ 2024-07-03 10:12:39,278 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: resume
227
+ 2024-07-03 10:12:39,278 INFO HandlerThread:153 [handler.py:handle_request_resume():715] starting system metrics thread
228
+ 2024-07-03 10:12:39,278 INFO HandlerThread:153 [system_monitor.py:start():194] Starting system monitor
229
+ 2024-07-03 10:12:39,278 INFO SystemMonitor:153 [system_monitor.py:_start():158] Starting system asset monitoring threads
230
+ 2024-07-03 10:12:39,279 INFO SystemMonitor:153 [interfaces.py:start():188] Started cpu monitoring
231
+ 2024-07-03 10:12:39,280 INFO SystemMonitor:153 [interfaces.py:start():188] Started disk monitoring
232
+ 2024-07-03 10:12:39,282 INFO SystemMonitor:153 [interfaces.py:start():188] Started gpu monitoring
233
+ 2024-07-03 10:12:39,282 INFO SystemMonitor:153 [interfaces.py:start():188] Started memory monitoring
234
+ 2024-07-03 10:12:39,284 INFO SystemMonitor:153 [interfaces.py:start():188] Started network monitoring
235
+ 2024-07-03 10:12:40,032 DEBUG SenderThread:153 [sender.py:send():378] send: config
236
+ 2024-07-03 10:12:40,034 DEBUG SenderThread:153 [sender.py:send():378] send: metric
237
+ 2024-07-03 10:12:40,034 DEBUG SenderThread:153 [sender.py:send():378] send: metric
238
+ 2024-07-03 10:12:40,034 WARNING SenderThread:153 [sender.py:send_metric():1340] Seen metric with glob (shouldn't happen)
239
+ 2024-07-03 10:12:44,040 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
240
+ 2024-07-03 10:12:44,645 INFO Thread-12 :153 [dir_watcher.py:_on_file_modified():288] file/dir modified: /kaggle/working/wandb/run-20240703_100700-cyvdged5/files/config.yaml
241
+ 2024-07-03 10:12:47,748 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: stop_status
242
+ 2024-07-03 10:12:47,748 DEBUG SenderThread:153 [sender.py:send_request():405] send_request: stop_status
243
+ 2024-07-03 10:12:50,014 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
244
+ 2024-07-03 10:12:55,015 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
245
+ 2024-07-03 10:13:00,016 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
246
+ 2024-07-03 10:13:02,748 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: stop_status
247
+ 2024-07-03 10:13:02,748 DEBUG SenderThread:153 [sender.py:send_request():405] send_request: stop_status
248
+ 2024-07-03 10:13:06,006 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
249
+ 2024-07-03 10:13:11,007 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
250
+ 2024-07-03 10:13:16,008 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
251
+ 2024-07-03 10:13:17,748 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: stop_status
252
+ 2024-07-03 10:13:17,748 DEBUG SenderThread:153 [sender.py:send_request():405] send_request: stop_status
253
+ 2024-07-03 10:13:22,000 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
254
+ 2024-07-03 10:13:27,001 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
255
+ 2024-07-03 10:13:32,002 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
256
+ 2024-07-03 10:13:32,748 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: stop_status
257
+ 2024-07-03 10:13:32,748 DEBUG SenderThread:153 [sender.py:send_request():405] send_request: stop_status
258
+ 2024-07-03 10:13:37,018 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
259
+ 2024-07-03 10:13:39,284 DEBUG SystemMonitor:153 [system_monitor.py:_start():172] Starting system metrics aggregation loop
260
+ 2024-07-03 10:13:39,285 DEBUG SenderThread:153 [sender.py:send():378] send: stats
261
+ 2024-07-03 10:13:42,286 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
262
+ 2024-07-03 10:13:47,288 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
263
+ 2024-07-03 10:13:47,748 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: stop_status
264
+ 2024-07-03 10:13:47,748 DEBUG SenderThread:153 [sender.py:send_request():405] send_request: stop_status
265
+ 2024-07-03 10:13:52,987 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
266
+ 2024-07-03 10:13:55,033 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: pause
267
+ 2024-07-03 10:13:55,033 INFO HandlerThread:153 [handler.py:handle_request_pause():724] stopping system metrics thread
268
+ 2024-07-03 10:13:55,033 INFO HandlerThread:153 [system_monitor.py:finish():203] Stopping system monitor
269
+ 2024-07-03 10:13:55,033 DEBUG SystemMonitor:153 [system_monitor.py:_start():179] Finished system metrics aggregation loop
270
+ 2024-07-03 10:13:55,033 DEBUG SystemMonitor:153 [system_monitor.py:_start():183] Publishing last batch of metrics
271
+ 2024-07-03 10:13:55,034 INFO HandlerThread:153 [interfaces.py:finish():200] Joined cpu monitor
272
+ 2024-07-03 10:13:55,035 INFO HandlerThread:153 [interfaces.py:finish():200] Joined disk monitor
273
+ 2024-07-03 10:13:55,041 INFO HandlerThread:153 [interfaces.py:finish():200] Joined gpu monitor
274
+ 2024-07-03 10:13:55,041 INFO HandlerThread:153 [interfaces.py:finish():200] Joined memory monitor
275
+ 2024-07-03 10:13:55,041 INFO HandlerThread:153 [interfaces.py:finish():200] Joined network monitor
276
+ 2024-07-03 10:13:55,042 DEBUG SenderThread:153 [sender.py:send():378] send: stats
277
+ 2024-07-03 10:13:58,043 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
278
+ 2024-07-03 10:13:58,120 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: resume
279
+ 2024-07-03 10:13:58,120 INFO HandlerThread:153 [handler.py:handle_request_resume():715] starting system metrics thread
280
+ 2024-07-03 10:13:58,120 INFO HandlerThread:153 [system_monitor.py:start():194] Starting system monitor
281
+ 2024-07-03 10:13:58,120 INFO SystemMonitor:153 [system_monitor.py:_start():158] Starting system asset monitoring threads
282
+ 2024-07-03 10:13:58,121 INFO SystemMonitor:153 [interfaces.py:start():188] Started cpu monitoring
283
+ 2024-07-03 10:13:58,121 INFO SystemMonitor:153 [interfaces.py:start():188] Started disk monitoring
284
+ 2024-07-03 10:13:58,122 INFO SystemMonitor:153 [interfaces.py:start():188] Started gpu monitoring
285
+ 2024-07-03 10:13:58,125 INFO SystemMonitor:153 [interfaces.py:start():188] Started memory monitoring
286
+ 2024-07-03 10:13:58,126 INFO SystemMonitor:153 [interfaces.py:start():188] Started network monitoring
287
+ 2024-07-03 10:14:02,749 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: stop_status
288
+ 2024-07-03 10:14:02,749 DEBUG SenderThread:153 [sender.py:send_request():405] send_request: stop_status
289
+ 2024-07-03 10:14:03,238 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
290
+ 2024-07-03 10:14:08,239 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
291
+ 2024-07-03 10:14:13,240 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
292
+ 2024-07-03 10:14:17,748 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: stop_status
293
+ 2024-07-03 10:14:17,749 DEBUG SenderThread:153 [sender.py:send_request():405] send_request: stop_status
294
+ 2024-07-03 10:14:19,050 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
295
+ 2024-07-03 10:14:24,051 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
296
+ 2024-07-03 10:14:29,053 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
297
+ 2024-07-03 10:14:32,748 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: stop_status
298
+ 2024-07-03 10:14:32,749 DEBUG SenderThread:153 [sender.py:send_request():405] send_request: stop_status
299
+ 2024-07-03 10:14:34,055 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
300
+ 2024-07-03 10:14:39,056 DEBUG HandlerThread:153 [handler.py:handle_request():158] handle_request: status_report
wandb/run-20240703_100700-cyvdged5/logs/debug.log ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-07-03 10:07:00,447 INFO MainThread:34 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0
2
+ 2024-07-03 10:07:00,447 INFO MainThread:34 [wandb_setup.py:_flush():76] Configure stats pid to 34
3
+ 2024-07-03 10:07:00,447 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
4
+ 2024-07-03 10:07:00,447 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from /kaggle/working/wandb/settings
5
+ 2024-07-03 10:07:00,447 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
6
+ 2024-07-03 10:07:00,447 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
7
+ 2024-07-03 10:07:00,447 INFO MainThread:34 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program': '<python with no main file>'}
8
+ 2024-07-03 10:07:00,447 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying login settings: {}
9
+ 2024-07-03 10:07:00,447 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying login settings: {'api_key': '***REDACTED***'}
10
+ 2024-07-03 10:07:00,447 INFO MainThread:34 [wandb_init.py:_log_setup():520] Logging user logs to /kaggle/working/wandb/run-20240703_100700-cyvdged5/logs/debug.log
11
+ 2024-07-03 10:07:00,447 INFO MainThread:34 [wandb_init.py:_log_setup():521] Logging internal logs to /kaggle/working/wandb/run-20240703_100700-cyvdged5/logs/debug-internal.log
12
+ 2024-07-03 10:07:00,447 INFO MainThread:34 [wandb_init.py:_jupyter_setup():466] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x78e7a57abb20>
13
+ 2024-07-03 10:07:00,448 INFO MainThread:34 [wandb_init.py:init():560] calling init triggers
14
+ 2024-07-03 10:07:00,448 INFO MainThread:34 [wandb_init.py:init():567] wandb.init called with sweep_config: {}
15
+ config: {}
16
+ 2024-07-03 10:07:00,448 INFO MainThread:34 [wandb_init.py:init():610] starting backend
17
+ 2024-07-03 10:07:00,448 INFO MainThread:34 [wandb_init.py:init():614] setting up manager
18
+ 2024-07-03 10:07:00,449 INFO MainThread:34 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
19
+ 2024-07-03 10:07:00,452 INFO MainThread:34 [wandb_init.py:init():622] backend started and connected
20
+ 2024-07-03 10:07:00,464 INFO MainThread:34 [wandb_run.py:_label_probe_notebook():1328] probe notebook
21
+ 2024-07-03 10:07:00,953 INFO MainThread:34 [wandb_init.py:init():711] updated telemetry
22
+ 2024-07-03 10:07:00,957 INFO MainThread:34 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout
23
+ 2024-07-03 10:07:01,536 INFO MainThread:34 [wandb_run.py:_on_init():2396] communicating current version
24
+ 2024-07-03 10:07:01,561 INFO MainThread:34 [wandb_run.py:_on_init():2405] got version response upgrade_message: "wandb version 0.17.3 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
25
+
26
+ 2024-07-03 10:07:01,562 INFO MainThread:34 [wandb_init.py:init():795] starting run threads in backend
27
+ 2024-07-03 10:07:17,603 INFO MainThread:34 [wandb_run.py:_console_start():2374] atexit reg
28
+ 2024-07-03 10:07:17,603 INFO MainThread:34 [wandb_run.py:_redirect():2229] redirect: wrap_raw
29
+ 2024-07-03 10:07:17,604 INFO MainThread:34 [wandb_run.py:_redirect():2294] Wrapping output streams.
30
+ 2024-07-03 10:07:17,604 INFO MainThread:34 [wandb_run.py:_redirect():2319] Redirects installed.
31
+ 2024-07-03 10:07:17,607 INFO MainThread:34 [wandb_init.py:init():838] run started, returning control to user process
32
+ 2024-07-03 10:07:17,615 INFO MainThread:34 [wandb_run.py:_config_callback():1376] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['BertForMaskedLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1', 2: 'LABEL_2', 3: 'LABEL_3', 4: 'LABEL_4', 5: 'LABEL_5', 6: 'LABEL_6', 7: 'LABEL_7', 8: 'LABEL_8'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1, 'LABEL_2': 2, 'LABEL_3': 3, 'LABEL_4': 4, 'LABEL_5': 5, 'LABEL_6': 6, 'LABEL_7': 7, 'LABEL_8': 8}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': 0, 'eos_token_id': None, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'ai-forever/ruBert-base', 'transformers_version': '4.41.2', 'directionality': 'bidi', 'pooler_fc_size': 768, 'pooler_num_attention_heads': 12, 'pooler_num_fc_layers': 3, 'pooler_size_per_head': 128, 'pooler_type': 'first_token_transform', 'model_type': 'bert', 'vocab_size': 120138, 'hidden_size': 768, 'num_hidden_layers': 12, 'num_attention_heads': 12, 'hidden_act': 'gelu', 'intermediate_size': 3072, 'hidden_dropout_prob': 0.1, 'attention_probs_dropout_prob': 0.1, 'max_position_embeddings': 512, 'type_vocab_size': 2, 'initializer_range': 0.02, 'layer_norm_eps': 1e-12, 'position_embedding_type': 'absolute', 'use_cache': True, 'classifier_dropout': None, 'output_dir': '/kaggle/working/results', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'eval_strategy': 'epoch', 'prediction_loss_only': False, 'per_device_train_batch_size': 16, 'per_device_eval_batch_size': 16, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 2e-05, 'weight_decay': 0.01, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 4, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/logs', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'no', 'save_steps': 500, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'restore_callback_states_from_checkpoint': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/results', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'eval_do_concat_batches': True, 'fp16_backend': 'auto', 'evaluation_strategy': 'epoch', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None, 'batch_eval_metrics': False}
33
+ 2024-07-03 10:10:33,327 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
34
+ 2024-07-03 10:10:33,327 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
35
+ 2024-07-03 10:12:28,050 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
36
+ 2024-07-03 10:12:31,153 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
37
+ 2024-07-03 10:12:31,153 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
38
+ 2024-07-03 10:12:36,070 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
39
+ 2024-07-03 10:12:36,071 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
40
+ 2024-07-03 10:12:36,071 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
41
+ 2024-07-03 10:12:39,277 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
42
+ 2024-07-03 10:12:40,028 INFO MainThread:34 [wandb_run.py:_config_callback():1376] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['BertForMaskedLM'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1', 2: 'LABEL_2', 3: 'LABEL_3', 4: 'LABEL_4', 5: 'LABEL_5', 6: 'LABEL_6', 7: 'LABEL_7', 8: 'LABEL_8'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1, 'LABEL_2': 2, 'LABEL_3': 3, 'LABEL_4': 4, 'LABEL_5': 5, 'LABEL_6': 6, 'LABEL_7': 7, 'LABEL_8': 8}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': None, 'pad_token_id': 0, 'eos_token_id': None, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'ai-forever/ruBert-base', 'transformers_version': '4.41.2', 'directionality': 'bidi', 'pooler_fc_size': 768, 'pooler_num_attention_heads': 12, 'pooler_num_fc_layers': 3, 'pooler_size_per_head': 128, 'pooler_type': 'first_token_transform', 'model_type': 'bert', 'vocab_size': 120138, 'hidden_size': 768, 'num_hidden_layers': 12, 'num_attention_heads': 12, 'hidden_act': 'gelu', 'intermediate_size': 3072, 'hidden_dropout_prob': 0.1, 'attention_probs_dropout_prob': 0.1, 'max_position_embeddings': 512, 'type_vocab_size': 2, 'initializer_range': 0.02, 'layer_norm_eps': 1e-12, 'position_embedding_type': 'absolute', 'use_cache': True, 'classifier_dropout': None, 'output_dir': '/kaggle/working/results', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'eval_strategy': 'epoch', 'prediction_loss_only': False, 'per_device_train_batch_size': 16, 'per_device_eval_batch_size': 16, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 2e-05, 'weight_decay': 0.01, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 3, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 0, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/logs', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'no', 'save_steps': 500, 'save_total_limit': None, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'restore_callback_states_from_checkpoint': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': False, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': None, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/results', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': False, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': False, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': False, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'eval_do_concat_batches': True, 'fp16_backend': 'auto', 'evaluation_strategy': 'epoch', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None, 'batch_eval_metrics': False}
43
+ 2024-07-03 10:13:55,032 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
44
+ 2024-07-03 10:13:55,032 INFO MainThread:34 [wandb_init.py:_pause_backend():431] pausing backend
45
+ 2024-07-03 10:13:58,119 INFO MainThread:34 [wandb_init.py:_resume_backend():436] resuming backend
wandb/run-20240703_100700-cyvdged5/run-cyvdged5.wandb ADDED
Binary file (20.7 kB). View file