Training in progress, step 500
Browse files- config.json +118 -0
- model.safetensors +3 -0
- preprocessor_config.json +9 -0
- runs/May24_20-24-49_890e8b3ca76b/events.out.tfevents.1716582397.890e8b3ca76b.34.0 +3 -0
- runs/May24_20-31-07_890e8b3ca76b/events.out.tfevents.1716582710.890e8b3ca76b.34.1 +3 -0
- runs/May24_20-34-06_890e8b3ca76b/events.out.tfevents.1716582886.890e8b3ca76b.34.2 +3 -0
- runs/May24_20-40-48_890e8b3ca76b/events.out.tfevents.1716583290.890e8b3ca76b.34.3 +3 -0
- runs/May24_20-42-26_890e8b3ca76b/events.out.tfevents.1716583387.890e8b3ca76b.34.4 +3 -0
- runs/May24_20-48-13_890e8b3ca76b/events.out.tfevents.1716583720.890e8b3ca76b.34.5 +3 -0
- training_args.bin +3 -0
- vocab.json +1 -0
- wandb/debug-internal.log +0 -0
- wandb/debug.log +217 -0
- wandb/run-20240524_202737-n1w0kmmv/files/conda-environment.yaml +0 -0
- wandb/run-20240524_202737-n1w0kmmv/files/config.yaml +866 -0
- wandb/run-20240524_202737-n1w0kmmv/files/output.log +171 -0
- wandb/run-20240524_202737-n1w0kmmv/files/requirements.txt +862 -0
- wandb/run-20240524_202737-n1w0kmmv/files/wandb-metadata.json +62 -0
- wandb/run-20240524_202737-n1w0kmmv/files/wandb-summary.json +1 -0
- wandb/run-20240524_202737-n1w0kmmv/logs/debug-internal.log +0 -0
- wandb/run-20240524_202737-n1w0kmmv/logs/debug.log +217 -0
- wandb/run-20240524_202737-n1w0kmmv/run-n1w0kmmv.wandb +0 -0
config.json
ADDED
|
@@ -0,0 +1,118 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_name_or_path": "facebook/wav2vec2-base",
|
| 3 |
+
"activation_dropout": 0.0,
|
| 4 |
+
"adapter_attn_dim": null,
|
| 5 |
+
"adapter_kernel_size": 3,
|
| 6 |
+
"adapter_stride": 2,
|
| 7 |
+
"add_adapter": false,
|
| 8 |
+
"apply_spec_augment": true,
|
| 9 |
+
"architectures": [
|
| 10 |
+
"Wav2Vec2ForCTC"
|
| 11 |
+
],
|
| 12 |
+
"attention_dropout": 0.1,
|
| 13 |
+
"bos_token_id": 1,
|
| 14 |
+
"classifier_proj_size": 256,
|
| 15 |
+
"codevector_dim": 256,
|
| 16 |
+
"contrastive_logits_temperature": 0.1,
|
| 17 |
+
"conv_bias": false,
|
| 18 |
+
"conv_dim": [
|
| 19 |
+
512,
|
| 20 |
+
512,
|
| 21 |
+
512,
|
| 22 |
+
512,
|
| 23 |
+
512,
|
| 24 |
+
512,
|
| 25 |
+
512
|
| 26 |
+
],
|
| 27 |
+
"conv_kernel": [
|
| 28 |
+
10,
|
| 29 |
+
3,
|
| 30 |
+
3,
|
| 31 |
+
3,
|
| 32 |
+
3,
|
| 33 |
+
2,
|
| 34 |
+
2
|
| 35 |
+
],
|
| 36 |
+
"conv_stride": [
|
| 37 |
+
5,
|
| 38 |
+
2,
|
| 39 |
+
2,
|
| 40 |
+
2,
|
| 41 |
+
2,
|
| 42 |
+
2,
|
| 43 |
+
2
|
| 44 |
+
],
|
| 45 |
+
"ctc_loss_reduction": "sum",
|
| 46 |
+
"ctc_zero_infinity": false,
|
| 47 |
+
"diversity_loss_weight": 0.1,
|
| 48 |
+
"do_stable_layer_norm": false,
|
| 49 |
+
"eos_token_id": 2,
|
| 50 |
+
"feat_extract_activation": "gelu",
|
| 51 |
+
"feat_extract_norm": "group",
|
| 52 |
+
"feat_proj_dropout": 0.1,
|
| 53 |
+
"feat_quantizer_dropout": 0.0,
|
| 54 |
+
"final_dropout": 0.0,
|
| 55 |
+
"freeze_feat_extract_train": true,
|
| 56 |
+
"hidden_act": "gelu",
|
| 57 |
+
"hidden_dropout": 0.1,
|
| 58 |
+
"hidden_size": 768,
|
| 59 |
+
"initializer_range": 0.02,
|
| 60 |
+
"intermediate_size": 3072,
|
| 61 |
+
"layer_norm_eps": 1e-05,
|
| 62 |
+
"layerdrop": 0.0,
|
| 63 |
+
"mask_channel_length": 10,
|
| 64 |
+
"mask_channel_min_space": 1,
|
| 65 |
+
"mask_channel_other": 0.0,
|
| 66 |
+
"mask_channel_prob": 0.0,
|
| 67 |
+
"mask_channel_selection": "static",
|
| 68 |
+
"mask_feature_length": 10,
|
| 69 |
+
"mask_feature_min_masks": 0,
|
| 70 |
+
"mask_feature_prob": 0.0,
|
| 71 |
+
"mask_time_length": 10,
|
| 72 |
+
"mask_time_min_masks": 2,
|
| 73 |
+
"mask_time_min_space": 1,
|
| 74 |
+
"mask_time_other": 0.0,
|
| 75 |
+
"mask_time_prob": 0.05,
|
| 76 |
+
"mask_time_selection": "static",
|
| 77 |
+
"model_type": "wav2vec2",
|
| 78 |
+
"no_mask_channel_overlap": false,
|
| 79 |
+
"no_mask_time_overlap": false,
|
| 80 |
+
"num_adapter_layers": 3,
|
| 81 |
+
"num_attention_heads": 12,
|
| 82 |
+
"num_codevector_groups": 2,
|
| 83 |
+
"num_codevectors_per_group": 320,
|
| 84 |
+
"num_conv_pos_embedding_groups": 16,
|
| 85 |
+
"num_conv_pos_embeddings": 128,
|
| 86 |
+
"num_feat_extract_layers": 7,
|
| 87 |
+
"num_hidden_layers": 12,
|
| 88 |
+
"num_negatives": 100,
|
| 89 |
+
"output_hidden_size": 768,
|
| 90 |
+
"pad_token_id": 0,
|
| 91 |
+
"proj_codevector_dim": 256,
|
| 92 |
+
"tdnn_dilation": [
|
| 93 |
+
1,
|
| 94 |
+
2,
|
| 95 |
+
3,
|
| 96 |
+
1,
|
| 97 |
+
1
|
| 98 |
+
],
|
| 99 |
+
"tdnn_dim": [
|
| 100 |
+
512,
|
| 101 |
+
512,
|
| 102 |
+
512,
|
| 103 |
+
512,
|
| 104 |
+
1500
|
| 105 |
+
],
|
| 106 |
+
"tdnn_kernel": [
|
| 107 |
+
5,
|
| 108 |
+
3,
|
| 109 |
+
3,
|
| 110 |
+
1,
|
| 111 |
+
1
|
| 112 |
+
],
|
| 113 |
+
"torch_dtype": "float32",
|
| 114 |
+
"transformers_version": "4.39.3",
|
| 115 |
+
"use_weighted_layer_sum": false,
|
| 116 |
+
"vocab_size": 100000000000000,
|
| 117 |
+
"xvector_output_dim": 512
|
| 118 |
+
}
|
model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ffef46ba0d91b26720ade5a0764b8eb00dd4cd8b026b05ad7fdf73b5bba92b4e
|
| 3 |
+
size 377611120
|
preprocessor_config.json
ADDED
|
@@ -0,0 +1,9 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"do_normalize": true,
|
| 3 |
+
"feature_extractor_type": "Wav2Vec2FeatureExtractor",
|
| 4 |
+
"feature_size": 1,
|
| 5 |
+
"padding_side": "right",
|
| 6 |
+
"padding_value": 0.0,
|
| 7 |
+
"return_attention_mask": false,
|
| 8 |
+
"sampling_rate": 16000
|
| 9 |
+
}
|
runs/May24_20-24-49_890e8b3ca76b/events.out.tfevents.1716582397.890e8b3ca76b.34.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2ac75e4115e01a183622b5727806489948ea4e87bf47a2c8f2b5e1a6c5093191
|
| 3 |
+
size 6307
|
runs/May24_20-31-07_890e8b3ca76b/events.out.tfevents.1716582710.890e8b3ca76b.34.1
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:60e1e37f65bdcdc419a35133bd27dc8b7b05882f02164c87954e2907a0730392
|
| 3 |
+
size 6307
|
runs/May24_20-34-06_890e8b3ca76b/events.out.tfevents.1716582886.890e8b3ca76b.34.2
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1ad92b4a9897c1c7d2d22b7271691629c322b0f513667db5be7bde20694cc1ac
|
| 3 |
+
size 6307
|
runs/May24_20-40-48_890e8b3ca76b/events.out.tfevents.1716583290.890e8b3ca76b.34.3
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6420e7a899ff2207f5231157234ae31754f5ffed0d5dbd936bd905f10d3b0e25
|
| 3 |
+
size 6307
|
runs/May24_20-42-26_890e8b3ca76b/events.out.tfevents.1716583387.890e8b3ca76b.34.4
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e83eaf9821d5367c3e81a47f83ef3e3da22c3eb4e21e6a5edaa39136b4cd55fb
|
| 3 |
+
size 6307
|
runs/May24_20-48-13_890e8b3ca76b/events.out.tfevents.1716583720.890e8b3ca76b.34.5
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5809329e6401c44df0ad3d49be3fd6b3e399b208768feecbf1e0e9f85d6c2341
|
| 3 |
+
size 19288
|
training_args.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:141b8c574fad469c69f1096f214b83f7c7364266f51555c4dd421e97e70dca21
|
| 3 |
+
size 4920
|
vocab.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"\uc644": 0, "\uc800": 1, "\ub797": 2, "\ucabd": 3, "\uc5c7": 4, "\ubbf8": 5, "\uc7a0": 6, "\ub17c": 7, "\uac9f": 8, "\ucef7": 9, "\uba4b": 10, "\ub611": 11, "\ub534": 12, "\ube60": 13, "\ub189": 14, "\uc090": 15, "\ub36e": 16, "\uafb9": 17, "\ub370": 18, "\ud138": 19, "\uaf3c": 20, "\ub5a0": 21, "\uce61": 22, "\ud480": 23, "\uc70c": 24, "\uac74": 25, "\uc2f6": 26, "\uad75": 27, "\ubb54": 28, "\ud604": 29, "\uc798": 30, "\ube68": 31, "\uae4d": 32, "\uac70": 33, "\ub139": 34, "\ub108": 35, "\ub128": 36, "\ucd0c": 37, "\ub9ce": 38, "\ub518": 39, "\ucda9": 40, "\uc794": 41, "\ub9db": 42, "\uba69": 43, "\ub538": 44, "\uc811": 45, "\ub9c8": 46, "\ud55c": 47, "\uc90d": 48, "\ub1a7": 49, "\ubab0": 50, "\uc601": 51, "\uc5f0": 52, "\ub2f9": 53, "\ub123": 54, "\uc12c": 55, "\uc27d": 56, "\ud615": 57, "\ub04c": 58, "\ucc38": 59, "\ud31c": 60, "\ucd95": 61, "\uc500": 62, "\uc694": 63, "\uadc0": 64, "\ucce5": 65, "\ud504": 66, "\uc5d4": 67, "\uacac": 68, "\uc80a": 69, "\uc78e": 70, "\ud0c1": 71, "\ub2ee": 72, "\ucf54": 73, "\uc5d0": 74, "\uca6c": 75, "\uae08": 76, "\uc5c8": 77, "\ub77d": 78, "\ud6fc": 79, "\ucc44": 80, "\ub798": 81, "\ub4f1": 82, "\uac16": 83, "\ubc84": 84, "\ub728": 85, "\ubd09": 86, "\uafb8": 87, "\ubd14": 88, "\uac90": 89, "\ud5e8": 90, "\uba64": 91, "\uacc4": 92, "\uc4f8": 93, "\uae40": 94, "\uadfc": 95, "\ubcbd": 96, "\ub9bd": 97, "\uc608": 98, "\ub978": 99, "\ub0ad": 100, "\ucc98": 101, "\uc369": 102, "\uc820": 103, "\uba4d": 104, "\uc82f": 105, "\ub461": 106, "\ud2c0": 107, "\ucbe4": 108, "\ubbfc": 109, "\ubc00": 110, "\uc92d": 111, "\uc058": 112, "\uc0c9": 113, "\uac1d": 114, "\uba3c": 115, "\ub0d0": 116, "\ub499": 117, "\uac10": 118, "\ud1a0": 119, "\uaf48": 120, "\uaca1": 121, "\uc9f8": 122, "\uace4": 123, "\ub298": 124, "\ub0b8": 125, "\uac1c": 126, "\uba55": 127, "\ub460": 128, "\uad7d": 129, "\ub35c": 130, "\ub2d0": 131, "\uc464": 132, "\uc88b": 133, "\ub9dd": 134, "\ubb58": 135, "\ud63c": 136, "\ud31f": 137, "\ud5d0": 138, "\uc57d": 139, "\ud0c7": 140, "\ub625": 141, "\uce68": 142, "\ub374": 143, "\uc368": 144, "\ud070": 145, "\ubc29": 146, "\ubc1f": 147, "\uacaa": 148, "\ud310": 149, "\uc78a": 150, "\ub07c": 151, "\uce59": 152, "\ub8e9": 153, "\ud45c": 154, "\ud15d": 155, "\ub0c9": 156, "\uc904": 157, "\uadf8": 158, "\uad70": 159, "\ud5f4": 160, "\uc2eb": 161, "\uc2ac": 162, "\uac13": 163, "\uc8fd": 164, "\uad6c": 165, "\uc815": 166, "\uace1": 167, "\ub364": 168, "\uae68": 169, "\ub8fd": 170, "\uae50": 171, "\ub9b0": 172, "\ubb63": 173, "\ud0c4": 174, "\ub0a0": 175, "\uc5c5": 176, "\ub118": 177, "\uce60": 178, "\uc6b8": 179, "\ub791": 180, "\ubc8b": 181, "\ubcbc": 182, "\ub3cc": 183, "\uac8c": 184, "\uc5d8": 185, "\uacf0": 186, "\ub141": 187, "\ud758": 188, "\ube44": 189, "\ubb38": 190, "\ubcc4": 191, "\uafe9": 192, "\uae5c": 193, "\uad00": 194, "\ud3b8": 195, "\ub0ab": 196, "\ud3c9": 197, "\uce69": 198, "\uc797": 199, "\ud1b5": 200, "\ub290": 201, "\ubb18": 202, "\ud751": 203, "\uad7c": 204, "\ud074": 205, "\uc557": 206, "\uc190": 207, "\uc84b": 208, "\uaf2d": 209, ",": 210, "\uc911": 211, "\ubcfc": 212, "\uccd0": 213, "\uaddc": 214, "\ud314": 215, "\uc0c1": 216, "\uc2ed": 217, "\uc13c": 218, "\uc5bc": 219, "\ubb3b": 220, "\uc2b7": 221, "\uba65": 222, "\uc370": 223, "\ub51c": 224, "\uac78": 225, "\ub8cc": 226, "\ub2f5": 227, "\ub837": 228, "\uc790": 229, "\ub4f8": 230, "\ud321": 231, "\ubc8c": 232, "\uacf5": 233, "\ud0dc": 234, "\uba40": 235, "\ubd10": 236, "\uba38": 237, "\ubcf5": 238, "\ub7ff": 239, "\ub985": 240, "\ubeff": 241, "\ubc18": 242, "\uce78": 243, "\uc2ec": 244, "\ud48d": 245, "\uc18c": 246, "\ub2ff": 247, "\ucfe0": 248, "\uba39": 249, "\ub9b4": 250, "\uc77c": 251, "\ub808": 252, "\ub9bc": 253, "\ucc28": 254, "\ud154": 255, "\ub531": 256, "\uc384": 257, "\ub801": 258, "\ub418": 259, "\uad6d": 260, "\ub5b5": 261, "\ud130": 262, "\ud6a8": 263, "\uae38": 264, "\uc591": 265, "\uc218": 266, "\uc554": 267, "\uc124": 268, "\uc2f8": 269, "\ub7ec": 270, "\ub458": 271, "\ubc88": 272, "\uad74": 273, "\ub110": 274, "\ucd94": 275, "\ubabb": 276, "\uc0bc": 277, "\ub828": 278, "\ub46c": 279, "\ud37c": 280, "\uae5f": 281, "\uc9d1": 282, "\uc0cc": 283, "\uc606": 284, "\ud788": 285, "\uaca0": 286, "\uadf9": 287, "\ub97c": 288, "\uc704": 289, "\uacfc": 290, "\uc6b4": 291, "\ubcc0": 292, "\ube48": 293, "\uce74": 294, "\uc2dc": 295, "\uac00": 296, "\uc584": 297, "\ub550": 298, "\ub4ec": 299, "\ud5d8": 300, "\ub488": 301, "\uaf2c": 302, "\uc788": 303, "\ubaa8": 304, "\ud638": 305, "\uc2f1": 306, "\ub7c9": 307, "\uba67": 308, "\uaed1": 309, "\ub148": 310, "\uc14b": 311, "\ucf00": 312, "\uaf34": 313, "\ud280": 314, "\uc624": 315, "\ube57": 316, "\ud478": 317, "\ubb47": 318, "\uc2f9": 319, "\ubc25": 320, "\uc2a8": 321, "\uba70": 322, "\uc90f": 323, "\uada4": 324, "\uac11": 325, "\ud6cd": 326, "\ub179": 327, "\uc0b4": 328, "\ub4e4": 329, "\uc831": 330, "\ud559": 331, "\ud230": 332, "\uc52c": 333, "\uc838": 334, "\uc796": 335, "\ub150": 336, "\uc2a4": 337, "\ub048": 338, "\uc824": 339, "\ud5f7": 340, "\ucee4": 341, "\ub974": 342, "\ucc45": 343, "\ub151": 344, "\uc4f0": 345, "\ub2cc": 346, "\uc73c": 347, "\uc37b": 348, "\uaef4": 349, "\uce5c": 350, "\ubf51": 351, "\uc131": 352, "\ub0e5": 353, "\ub214": 354, "\uc5b5": 355, "\uc22b": 356, "\uacbd": 357, "\ub4a4": 358, "\ucf20": 359, "\ubabd": 360, "\uad58": 361, "\ub2f4": 362, "\uc881": 363, "\uc81c": 364, "\uae14": 365, "\uc548": 366, "\uc880": 367, "\uc740": 368, "\uccd4": 369, "\ub2ec": 370, "\uc0db": 371, "\ub530": 372, "\uc1a1": 373, "\ub840": 374, "?": 375, "\ub780": 376, "\ub9e4": 377, "\ucd5c": 378, "\uc83c": 379, "\ub807": 380, "\ub465": 381, "\uc538": 382, "\ub3c4": 383, "\uae30": 384, "\uc220": 385, "\ub8e8": 386, "\ub860": 387, "\uccad": 388, "\uc9c0": 389, "\ub0b3": 390, "\ud574": 391, "\uc545": 392, "\uace8": 393, "\ub824": 394, "\ud569": 395, "\uc14d": 396, "\ubfdc": 397, "\ubcd1": 398, "\ub481": 399, "\uc2c0": 400, "\ubb50": 401, "\ub4fc": 402, "\ud30c": 403, "\ub198": 404, "\ub355": 405, "\ub057": 406, "\uc6a9": 407, "\ub354": 408, "\ub178": 409, "\uae34": 410, "\ub5a8": 411, "\ubca1": 412, "\uad7f": 413, "\uc530": 414, "\uc54a": 415, "\ubcb5": 416, "\uc6c3": 417, "\uc870": 418, "\ud22c": 419, "\uba87": 420, "\uad50": 421, "\ud2b8": 422, "\ub871": 423, "\ubd84": 424, "\ub144": 425, "\ub420": 426, "\ub834": 427, "\uba74": 428, "\uc49c": 429, "\ud5f9": 430, "\uaecf": 431, "\ubc97": 432, "\ud68c": 433, "\ub2e8": 434, "\uc758": 435, "\uac19": 436, "\ubfcc": 437, "\uc6d0": 438, "\ucf69": 439, "\ub825": 440, "\ucc29": 441, "\uc9c8": 442, "\uc2e4": 443, "\ub294": 444, "\uc998": 445, "\uc778": 446, "\uacb8": 447, "\uaf07": 448, "\uc934": 449, "\ud2bf": 450, "\ud734": 451, "\uacb0": 452, "\ub839": 453, "\uacbb": 454, "\uccb4": 455, "\ub958": 456, "\ub7b5": 457, "\ubb34": 458, "\ud06c": 459, "\uc094": 460, "\ucef4": 461, "\uce6d": 462, "\uc9e4": 463, "\ub7f0": 464, "\ud544": 465, "\ubaa9": 466, "\ub9c9": 467, "\ub2a0": 468, "\ub9bf": 469, "\ub85d": 470, "\ub358": 471, "\uc544": 472, "\uaca9": 473, "\uc5f4": 474, "\ud655": 475, "\ub2e4": 476, "\ubc15": 477, "\uc9c1": 478, "\uaf3d": 479, "\uc313": 480, "\ub044": 481, "\ud5d9": 482, "\ucb64": 483, "\ub77c": 484, "\ub9cc": 485, "\ub274": 486, "\uc8fc": 487, "\ucc9c": 488, "\uc120": 489, "\ubc16": 490, "\uc5fc": 491, "\uc2b9": 492, "\ud5e4": 493, "\ub098": 494, "\ucc2e": 495, "\ubd05": 496, "\uc0dd": 497, "\ud640": 498, "\ub4b7": 499, "\ucabc": 500, "\uac83": 501, "\ud1f4": 502, "\uccb8": 503, "\ub299": 504, "\uacf1": 505, "\uc37d": 506, "\ub300": 507, "\ubc11": 508, "\ud6c4": 509, "\uba58": 510, "\uaff0": 511, "\ud5c8": 512, "\uacf3": 513, "\uc653": 514, "\uc751": 515, "\uc5e3": 516, "\uc784": 517, "\uca99": 518, "\uc7a5": 519, "\uc871": 520, "\ub194": 521, "\uade4": 522, "\uaed8": 523, "\ub73b": 524, "\uc158": 525, "\ub054": 526, "\ub4e0": 527, "\uc640": 528, "\ubb35": 529, "\uc57c": 530, "\ubc14": 531, "\ubcf4": 532, "\uc11d": 533, "\uac77": 534, "\uc5c6": 535, "\ucc1d": 536, "\ucd10": 537, "\ub81b": 538, "\ud65c": 539, "\ud0a4": 540, "\ub878": 541, "\ub4dc": 542, "\ub2b4": 543, "\ub3c5": 544, "\uc90c": 545, "\ud654": 546, "\uc60c": 547, "\uc918": 548, "\ud2f0": 549, "\uc75c": 550, "\ub5a1": 551, "\uc2dd": 552, "\uc529": 553, "\ubd81": 554, "\ucd9c": 555, "\uc74d": 556, "\ud568": 557, "\ud07c": 558, "\ub9ac": 559, "\uc988": 560, "\uc5c9": 561, "\uac15": 562, "\ub78c": 563, "\ubd90": 564, "\uc628": 565, "\uc300": 566, "\uba54": 567, "\ub2a6": 568, "\ud78c": 569, "\ud56d": 570, "\ud134": 571, "\uc0ac": 572, "\ub124": 573, "\ub369": 574, "\ub193": 575, "\u1161": 576, "\uc2b5": 577, "\uc167": 578, "\ub80c": 579, "\ub1a9": 580, "\uc7ac": 581, "\ub153": 582, "\uac24": 583, "\ubcb3": 584, "\ud1a7": 585, "\ubc24": 586, "\ud6c8": 587, "\uc61b": 588, "\ubd88": 589, "\uc6b1": 590, "\uc874": 591, "\uc6d4": 592, "\ub7fc": 593, "\uc785": 594, "\uc54c": 595, "\ube4c": 596, "\ub490": 597, "\ubed0": 598, "\ucd98": 599, "\u11af": 600, "\uc720": 601, "\ub47f": 602, "\ub1a4": 603, "\uc804": 604, "\uc816": 605, "\uba85": 606, "\ud488": 607, "\ubc85": 608, "\uc11c": 609, "\ud798": 610, "\uc250": 611, "\uc9c7": 612, "\uc55a": 613, "\ud0d1": 614, "\uac85": 615, "\uc721": 616, "\ud611": 617, "\uc74c": 618, "\uc30d": 619, "\uc598": 620, "\uc744": 621, "\uc774": 622, "\ub5bb": 623, "\uc791": 624, ".": 625, "\ub220": 626, "\ubca8": 627, "\uc495": 628, "\uacc1": 629, "\ub9f7": 630, "\ud3ec": 631, "\uc55e": 632, "\uac08": 633, "\ubbff": 634, "\uc5b8": 635, "\uace0": 636, "\ub3d9": 637, "\uafc0": 638, "\ubd25": 639, "\ud2b9": 640, "\ub8d0": 641, "\ucc30": 642, "\uac2f": 643, "\ub9de": 644, "\ub3fc": 645, "\uc7c1": 646, "\ubca0": 647, "\ub09c": 648, "\uae4c": 649, "\ud639": 650, "\ubc27": 651, "\ub78f": 652, "\ubcf8": 653, "\ucb49": 654, "\uc9d3": 655, "\uad82": 656, "\uc9f1": 657, "\ubc94": 658, "\uc787": 659, "\ubc95": 660, "\ud2bc": 661, "\ub20c": 662, "\uc5ec": 663, "\uc62e": 664, "\uc5b4": 665, "\ub984": 666, "\ud53c": 667, "\uac04": 668, "\ubd99": 669, "\ud14c": 670, "\ub9d0": 671, "\ud560": 672, "\ucca0": 673, "\uc138": 674, "\uac07": 675, "\uc5bd": 676, "\u1101": 677, "\uc50c": 678, "\uc6c5": 679, "\uc18d": 680, "\uc528": 681, "\ubc30": 682, "\uc559": 683, "\uc6e8": 684, "\ub2c8": 685, "\ub155": 686, "\ub9e8": 687, "\uc6b0": 688, "\ud750": 689, "\uc560": 690, "\uc5ed": 691, "\ub140": 692, "\uae61": 693, "\uc5ff": 694, "\ub54c": 695, "\ub7fd": 696, "\uc65c": 697, "\ubc1c": 698, "\uc84d": 699, "\ub137": 700, "\uc808": 701, "\ub0a8": 702, "\uca4c": 703, "\uc9d0": 704, "\ub055": 705, "\uad11": 706, "\uc580": 707, "\uc637": 708, "\ub0ae": 709, "\uc775": 710, "\ucf1c": 711, "\uca0b": 712, "\uc5e5": 713, "\uc678": 714, "\ub2d9": 715, "\ubd80": 716, "\uc26c": 717, "\ud5cc": 718, "\ud76c": 719, "\ud0d3": 720, "\ubb3c": 721, "\ub835": 722, "\uae00": 723, "\ub10c": 724, "\ub0bc": 725, "\uac01": 726, "\uc99d": 727, "\ubfd4": 728, "\ub35f": 729, "\ubd04": 730, "\uc2e0": 731, "\uc813": 732, "\ub610": 733, "\ub85c": 734, "\ud754": 735, "\uce6b": 736, "\ud0c0": 737, "\uc12f": 738, "\uac20": 739, "\uc9dd": 740, "\uaca8": 741, "\uae54": 742, "\uc655": 743, "\ub204": 744, "\ub498": 745, "\ub545": 746, "\ubc31": 747, "\ub5a4": 748, "\ub0b4": 749, "\ub385": 750, "\uc9c4": 751, "\ub2d8": 752, "\ub188": 753, "\ub3c8": 754, "\ub48c": 755, "\ub180": 756, "\ub05d": 757, "\ucacc": 758, "\uc21c": 759, "\uc0c8": 760, "\uc6c0": 762, "\uac81": 763, "\uc989": 764, "\ub450": 765, "\ud034": 766, "\uc0b0": 767, "\ud391": 768, "\ucc3e": 769, "\ub4ef": 770, "\ub81d": 771, "\uad81": 772, "\ub72c": 773, "\ucc0d": 774, "\ucc2c": 775, "\ubfd0": 776, "\ub2e5": 777, "\uc900": 778, "\uc885": 779, "\ubbc4": 780, "\uc6c1": 781, "\uac80": 782, "\ubc1b": 783, "\ud150": 784, "\ub18d": 785, "\ud2c8": 786, "\uc5bb": 787, "\uaebc": 788, "\ud558": 789, "\uce58": 790, "\ub514": 791, "\uc62c": 792, "\ucd08": 793, "\ubd23": 794, "\uc801": 795, "\uba83": 796, "\ub7ed": 797, "|": 761, "[UNK]": 798, "[PAD]": 799}
|
wandb/debug-internal.log
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
wandb/debug.log
ADDED
|
@@ -0,0 +1,217 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
2024-05-24 20:27:37,417 INFO MainThread:34 [wandb_setup.py:_flush():76] Current SDK version is 0.16.6
|
| 2 |
+
2024-05-24 20:27:37,417 INFO MainThread:34 [wandb_setup.py:_flush():76] Configure stats pid to 34
|
| 3 |
+
2024-05-24 20:27:37,417 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
|
| 4 |
+
2024-05-24 20:27:37,417 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from /kaggle/working/wandb/settings
|
| 5 |
+
2024-05-24 20:27:37,417 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
|
| 6 |
+
2024-05-24 20:27:37,417 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
|
| 7 |
+
2024-05-24 20:27:37,417 INFO MainThread:34 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program': '<python with no main file>'}
|
| 8 |
+
2024-05-24 20:27:37,417 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying login settings: {}
|
| 9 |
+
2024-05-24 20:27:37,417 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying login settings: {'api_key': '***REDACTED***'}
|
| 10 |
+
2024-05-24 20:27:37,417 INFO MainThread:34 [wandb_init.py:_log_setup():521] Logging user logs to /kaggle/working/wandb/run-20240524_202737-n1w0kmmv/logs/debug.log
|
| 11 |
+
2024-05-24 20:27:37,417 INFO MainThread:34 [wandb_init.py:_log_setup():522] Logging internal logs to /kaggle/working/wandb/run-20240524_202737-n1w0kmmv/logs/debug-internal.log
|
| 12 |
+
2024-05-24 20:27:37,417 INFO MainThread:34 [wandb_init.py:_jupyter_setup():467] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x785a802b9120>
|
| 13 |
+
2024-05-24 20:27:37,418 INFO MainThread:34 [wandb_init.py:init():561] calling init triggers
|
| 14 |
+
2024-05-24 20:27:37,418 INFO MainThread:34 [wandb_init.py:init():568] wandb.init called with sweep_config: {}
|
| 15 |
+
config: {}
|
| 16 |
+
2024-05-24 20:27:37,418 INFO MainThread:34 [wandb_init.py:init():611] starting backend
|
| 17 |
+
2024-05-24 20:27:37,418 INFO MainThread:34 [wandb_init.py:init():615] setting up manager
|
| 18 |
+
2024-05-24 20:27:37,420 INFO MainThread:34 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
|
| 19 |
+
2024-05-24 20:27:37,423 INFO MainThread:34 [wandb_init.py:init():623] backend started and connected
|
| 20 |
+
2024-05-24 20:27:37,434 INFO MainThread:34 [wandb_run.py:_label_probe_notebook():1299] probe notebook
|
| 21 |
+
2024-05-24 20:27:38,191 INFO MainThread:34 [wandb_init.py:init():715] updated telemetry
|
| 22 |
+
2024-05-24 20:27:38,195 INFO MainThread:34 [wandb_init.py:init():748] communicating run to backend with 90.0 second timeout
|
| 23 |
+
2024-05-24 20:27:38,448 INFO MainThread:34 [wandb_run.py:_on_init():2357] communicating current version
|
| 24 |
+
2024-05-24 20:27:38,512 INFO MainThread:34 [wandb_run.py:_on_init():2366] got version response upgrade_message: "wandb version 0.17.0 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
|
| 25 |
+
|
| 26 |
+
2024-05-24 20:27:38,514 INFO MainThread:34 [wandb_init.py:init():799] starting run threads in backend
|
| 27 |
+
2024-05-24 20:27:54,627 INFO MainThread:34 [wandb_run.py:_console_start():2335] atexit reg
|
| 28 |
+
2024-05-24 20:27:54,628 INFO MainThread:34 [wandb_run.py:_redirect():2190] redirect: wrap_raw
|
| 29 |
+
2024-05-24 20:27:54,628 INFO MainThread:34 [wandb_run.py:_redirect():2255] Wrapping output streams.
|
| 30 |
+
2024-05-24 20:27:54,629 INFO MainThread:34 [wandb_run.py:_redirect():2280] Redirects installed.
|
| 31 |
+
2024-05-24 20:27:54,630 INFO MainThread:34 [wandb_init.py:init():842] run started, returning control to user process
|
| 32 |
+
2024-05-24 20:27:54,637 INFO MainThread:34 [wandb_run.py:_config_callback():1347] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['Wav2Vec2ForPreTraining'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'facebook/wav2vec2-base', 'transformers_version': '4.39.3', 'freeze_feat_extract_train': True, 'mask_channel_length': 10, 'mask_channel_min_space': 1, 'mask_channel_other': 0.0, 'mask_channel_prob': 0.0, 'mask_channel_selection': 'static', 'mask_time_min_space': 1, 'mask_time_other': 0.0, 'mask_time_selection': 'static', 'model_type': 'wav2vec2', 'no_mask_channel_overlap': False, 'no_mask_time_overlap': False, 'num_feat_extract_layers': 7, 'hidden_size': 768, 'feat_extract_norm': 'group', 'feat_extract_activation': 'gelu', 'conv_dim': [512, 512, 512, 512, 512, 512, 512], 'conv_stride': [5, 2, 2, 2, 2, 2, 2], 'conv_kernel': [10, 3, 3, 3, 3, 2, 2], 'conv_bias': False, 'num_conv_pos_embeddings': 128, 'num_conv_pos_embedding_groups': 16, 'num_hidden_layers': 12, 'intermediate_size': 3072, 'hidden_act': 'gelu', 'num_attention_heads': 12, 'hidden_dropout': 0.1, 'attention_dropout': 0.1, 'activation_dropout': 0.0, 'feat_proj_dropout': 0.1, 'final_dropout': 0.0, 'layerdrop': 0.0, 'layer_norm_eps': 1e-05, 'initializer_range': 0.02, 'vocab_size': 32, 'do_stable_layer_norm': False, 'use_weighted_layer_sum': False, 'apply_spec_augment': True, 'mask_time_prob': 0.05, 'mask_time_length': 10, 'mask_time_min_masks': 2, 'mask_feature_prob': 0.0, 'mask_feature_length': 10, 'mask_feature_min_masks': 0, 'num_codevectors_per_group': 320, 'num_codevector_groups': 2, 'contrastive_logits_temperature': 0.1, 'feat_quantizer_dropout': 0.0, 'num_negatives': 100, 'codevector_dim': 256, 'proj_codevector_dim': 256, 'diversity_loss_weight': 0.1, 'ctc_loss_reduction': 'sum', 'ctc_zero_infinity': False, 'add_adapter': False, 'adapter_kernel_size': 3, 'adapter_stride': 2, 'num_adapter_layers': 3, 'output_hidden_size': 768, 'adapter_attn_dim': None, 'classifier_proj_size': 256, 'tdnn_dim': [512, 512, 512, 512, 1500], 'tdnn_kernel': [5, 3, 3, 1, 1], 'tdnn_dilation': [1, 2, 3, 1, 1], 'xvector_output_dim': 512, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 32, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 0.0001, 'weight_decay': 0.005, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 30, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 1000, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/May24_20-24-49_890e8b3ca76b', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 2, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': 500, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': True, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': True, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None}
|
| 33 |
+
2024-05-24 20:28:04,489 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 34 |
+
2024-05-24 20:28:04,489 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 35 |
+
2024-05-24 20:28:58,368 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 36 |
+
2024-05-24 20:28:59,625 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 37 |
+
2024-05-24 20:28:59,625 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 38 |
+
2024-05-24 20:29:08,949 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 39 |
+
2024-05-24 20:29:08,951 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 40 |
+
2024-05-24 20:29:08,951 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 41 |
+
2024-05-24 20:29:12,141 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 42 |
+
2024-05-24 20:29:16,856 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 43 |
+
2024-05-24 20:29:16,857 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 44 |
+
2024-05-24 20:29:20,764 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 45 |
+
2024-05-24 20:29:24,478 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 46 |
+
2024-05-24 20:29:24,478 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 47 |
+
2024-05-24 20:29:30,221 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 48 |
+
2024-05-24 20:30:45,072 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 49 |
+
2024-05-24 20:30:45,072 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 50 |
+
2024-05-24 20:30:58,707 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 51 |
+
2024-05-24 20:30:58,709 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 52 |
+
2024-05-24 20:30:58,710 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 53 |
+
2024-05-24 20:31:02,188 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 54 |
+
2024-05-24 20:31:02,613 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 55 |
+
2024-05-24 20:31:02,613 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 56 |
+
2024-05-24 20:31:04,725 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 57 |
+
2024-05-24 20:31:04,727 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 58 |
+
2024-05-24 20:31:04,727 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 59 |
+
2024-05-24 20:31:07,065 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 60 |
+
2024-05-24 20:31:07,298 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 61 |
+
2024-05-24 20:31:07,298 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 62 |
+
2024-05-24 20:31:10,187 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 63 |
+
2024-05-24 20:31:10,193 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 64 |
+
2024-05-24 20:31:10,193 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 65 |
+
2024-05-24 20:31:12,404 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 66 |
+
2024-05-24 20:31:50,143 INFO MainThread:34 [wandb_run.py:_config_callback():1347] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['Wav2Vec2ForPreTraining'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'facebook/wav2vec2-base', 'transformers_version': '4.39.3', 'freeze_feat_extract_train': True, 'mask_channel_length': 10, 'mask_channel_min_space': 1, 'mask_channel_other': 0.0, 'mask_channel_prob': 0.0, 'mask_channel_selection': 'static', 'mask_time_min_space': 1, 'mask_time_other': 0.0, 'mask_time_selection': 'static', 'model_type': 'wav2vec2', 'no_mask_channel_overlap': False, 'no_mask_time_overlap': False, 'num_feat_extract_layers': 7, 'hidden_size': 768, 'feat_extract_norm': 'group', 'feat_extract_activation': 'gelu', 'conv_dim': [512, 512, 512, 512, 512, 512, 512], 'conv_stride': [5, 2, 2, 2, 2, 2, 2], 'conv_kernel': [10, 3, 3, 3, 3, 2, 2], 'conv_bias': False, 'num_conv_pos_embeddings': 128, 'num_conv_pos_embedding_groups': 16, 'num_hidden_layers': 12, 'intermediate_size': 3072, 'hidden_act': 'gelu', 'num_attention_heads': 12, 'hidden_dropout': 0.1, 'attention_dropout': 0.1, 'activation_dropout': 0.0, 'feat_proj_dropout': 0.1, 'final_dropout': 0.0, 'layerdrop': 0.0, 'layer_norm_eps': 1e-05, 'initializer_range': 0.02, 'vocab_size': 32, 'do_stable_layer_norm': False, 'use_weighted_layer_sum': False, 'apply_spec_augment': True, 'mask_time_prob': 0.05, 'mask_time_length': 10, 'mask_time_min_masks': 2, 'mask_feature_prob': 0.0, 'mask_feature_length': 10, 'mask_feature_min_masks': 0, 'num_codevectors_per_group': 320, 'num_codevector_groups': 2, 'contrastive_logits_temperature': 0.1, 'feat_quantizer_dropout': 0.0, 'num_negatives': 100, 'codevector_dim': 256, 'proj_codevector_dim': 256, 'diversity_loss_weight': 0.1, 'ctc_loss_reduction': 'sum', 'ctc_zero_infinity': False, 'add_adapter': False, 'adapter_kernel_size': 3, 'adapter_stride': 2, 'num_adapter_layers': 3, 'output_hidden_size': 768, 'adapter_attn_dim': None, 'classifier_proj_size': 256, 'tdnn_dim': [512, 512, 512, 512, 1500], 'tdnn_kernel': [5, 3, 3, 1, 1], 'tdnn_dilation': [1, 2, 3, 1, 1], 'xvector_output_dim': 512, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 32, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 0.0001, 'weight_decay': 0.005, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 30, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 1000, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/May24_20-31-07_890e8b3ca76b', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 2, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': 500, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': True, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': True, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None}
|
| 67 |
+
2024-05-24 20:31:58,626 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 68 |
+
2024-05-24 20:31:58,626 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 69 |
+
2024-05-24 20:34:06,571 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 70 |
+
2024-05-24 20:34:06,717 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 71 |
+
2024-05-24 20:34:06,717 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 72 |
+
2024-05-24 20:34:10,001 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 73 |
+
2024-05-24 20:34:46,464 INFO MainThread:34 [wandb_run.py:_config_callback():1347] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['Wav2Vec2ForPreTraining'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'facebook/wav2vec2-base', 'transformers_version': '4.39.3', 'freeze_feat_extract_train': True, 'mask_channel_length': 10, 'mask_channel_min_space': 1, 'mask_channel_other': 0.0, 'mask_channel_prob': 0.0, 'mask_channel_selection': 'static', 'mask_time_min_space': 1, 'mask_time_other': 0.0, 'mask_time_selection': 'static', 'model_type': 'wav2vec2', 'no_mask_channel_overlap': False, 'no_mask_time_overlap': False, 'num_feat_extract_layers': 7, 'hidden_size': 768, 'feat_extract_norm': 'group', 'feat_extract_activation': 'gelu', 'conv_dim': [512, 512, 512, 512, 512, 512, 512], 'conv_stride': [5, 2, 2, 2, 2, 2, 2], 'conv_kernel': [10, 3, 3, 3, 3, 2, 2], 'conv_bias': False, 'num_conv_pos_embeddings': 128, 'num_conv_pos_embedding_groups': 16, 'num_hidden_layers': 12, 'intermediate_size': 3072, 'hidden_act': 'gelu', 'num_attention_heads': 12, 'hidden_dropout': 0.1, 'attention_dropout': 0.1, 'activation_dropout': 0.0, 'feat_proj_dropout': 0.1, 'final_dropout': 0.0, 'layerdrop': 0.0, 'layer_norm_eps': 1e-05, 'initializer_range': 0.02, 'vocab_size': 32, 'do_stable_layer_norm': False, 'use_weighted_layer_sum': False, 'apply_spec_augment': True, 'mask_time_prob': 0.05, 'mask_time_length': 10, 'mask_time_min_masks': 2, 'mask_feature_prob': 0.0, 'mask_feature_length': 10, 'mask_feature_min_masks': 0, 'num_codevectors_per_group': 320, 'num_codevector_groups': 2, 'contrastive_logits_temperature': 0.1, 'feat_quantizer_dropout': 0.0, 'num_negatives': 100, 'codevector_dim': 256, 'proj_codevector_dim': 256, 'diversity_loss_weight': 0.1, 'ctc_loss_reduction': 'sum', 'ctc_zero_infinity': False, 'add_adapter': False, 'adapter_kernel_size': 3, 'adapter_stride': 2, 'num_adapter_layers': 3, 'output_hidden_size': 768, 'adapter_attn_dim': None, 'classifier_proj_size': 256, 'tdnn_dim': [512, 512, 512, 512, 1500], 'tdnn_kernel': [5, 3, 3, 1, 1], 'tdnn_dilation': [1, 2, 3, 1, 1], 'xvector_output_dim': 512, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 64, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 0.0001, 'weight_decay': 0.005, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 30, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 1000, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/May24_20-34-06_890e8b3ca76b', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 2, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': 500, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': True, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': True, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None}
|
| 74 |
+
2024-05-24 20:34:55,928 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 75 |
+
2024-05-24 20:34:55,929 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 76 |
+
2024-05-24 20:35:26,131 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 77 |
+
2024-05-24 20:35:26,139 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 78 |
+
2024-05-24 20:35:26,139 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 79 |
+
2024-05-24 20:35:32,729 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 80 |
+
2024-05-24 20:35:32,732 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 81 |
+
2024-05-24 20:35:32,732 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 82 |
+
2024-05-24 20:35:34,729 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 83 |
+
2024-05-24 20:35:34,732 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 84 |
+
2024-05-24 20:35:34,733 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 85 |
+
2024-05-24 20:35:47,483 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 86 |
+
2024-05-24 20:35:47,488 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 87 |
+
2024-05-24 20:35:47,488 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 88 |
+
2024-05-24 20:35:56,249 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 89 |
+
2024-05-24 20:35:56,253 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 90 |
+
2024-05-24 20:35:56,253 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 91 |
+
2024-05-24 20:37:20,366 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 92 |
+
2024-05-24 20:37:20,370 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 93 |
+
2024-05-24 20:37:20,370 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 94 |
+
2024-05-24 20:37:28,481 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 95 |
+
2024-05-24 20:37:28,510 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 96 |
+
2024-05-24 20:37:28,511 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 97 |
+
2024-05-24 20:38:00,946 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 98 |
+
2024-05-24 20:38:00,950 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 99 |
+
2024-05-24 20:38:00,950 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 100 |
+
2024-05-24 20:38:04,524 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 101 |
+
2024-05-24 20:38:04,550 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 102 |
+
2024-05-24 20:38:04,550 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 103 |
+
2024-05-24 20:38:12,369 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 104 |
+
2024-05-24 20:38:12,432 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 105 |
+
2024-05-24 20:38:12,432 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 106 |
+
2024-05-24 20:38:16,938 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 107 |
+
2024-05-24 20:38:16,998 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 108 |
+
2024-05-24 20:38:16,998 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 109 |
+
2024-05-24 20:38:21,248 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 110 |
+
2024-05-24 20:38:21,260 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 111 |
+
2024-05-24 20:38:21,260 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 112 |
+
2024-05-24 20:38:22,930 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 113 |
+
2024-05-24 20:38:22,940 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 114 |
+
2024-05-24 20:38:22,940 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 115 |
+
2024-05-24 20:38:37,851 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 116 |
+
2024-05-24 20:38:55,254 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 117 |
+
2024-05-24 20:38:55,254 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 118 |
+
2024-05-24 20:38:58,788 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 119 |
+
2024-05-24 20:38:58,827 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 120 |
+
2024-05-24 20:38:58,827 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 121 |
+
2024-05-24 20:39:00,762 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 122 |
+
2024-05-24 20:40:16,333 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 123 |
+
2024-05-24 20:40:16,334 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 124 |
+
2024-05-24 20:40:21,472 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 125 |
+
2024-05-24 20:40:22,682 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 126 |
+
2024-05-24 20:40:22,682 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 127 |
+
2024-05-24 20:40:28,979 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 128 |
+
2024-05-24 20:40:33,961 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 129 |
+
2024-05-24 20:40:33,961 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 130 |
+
2024-05-24 20:40:36,994 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 131 |
+
2024-05-24 20:40:36,996 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 132 |
+
2024-05-24 20:40:36,997 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 133 |
+
2024-05-24 20:40:42,203 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 134 |
+
2024-05-24 20:40:42,621 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 135 |
+
2024-05-24 20:40:42,621 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 136 |
+
2024-05-24 20:40:44,473 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 137 |
+
2024-05-24 20:40:44,475 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 138 |
+
2024-05-24 20:40:44,475 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 139 |
+
2024-05-24 20:40:48,431 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 140 |
+
2024-05-24 20:40:48,669 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 141 |
+
2024-05-24 20:40:48,669 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 142 |
+
2024-05-24 20:40:50,725 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 143 |
+
2024-05-24 20:40:50,730 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 144 |
+
2024-05-24 20:40:50,730 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 145 |
+
2024-05-24 20:40:52,537 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 146 |
+
2024-05-24 20:41:30,450 INFO MainThread:34 [wandb_run.py:_config_callback():1347] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['Wav2Vec2ForPreTraining'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'facebook/wav2vec2-base', 'transformers_version': '4.39.3', 'freeze_feat_extract_train': True, 'mask_channel_length': 10, 'mask_channel_min_space': 1, 'mask_channel_other': 0.0, 'mask_channel_prob': 0.0, 'mask_channel_selection': 'static', 'mask_time_min_space': 1, 'mask_time_other': 0.0, 'mask_time_selection': 'static', 'model_type': 'wav2vec2', 'no_mask_channel_overlap': False, 'no_mask_time_overlap': False, 'num_feat_extract_layers': 7, 'hidden_size': 768, 'feat_extract_norm': 'group', 'feat_extract_activation': 'gelu', 'conv_dim': [512, 512, 512, 512, 512, 512, 512], 'conv_stride': [5, 2, 2, 2, 2, 2, 2], 'conv_kernel': [10, 3, 3, 3, 3, 2, 2], 'conv_bias': False, 'num_conv_pos_embeddings': 128, 'num_conv_pos_embedding_groups': 16, 'num_hidden_layers': 12, 'intermediate_size': 3072, 'hidden_act': 'gelu', 'num_attention_heads': 12, 'hidden_dropout': 0.1, 'attention_dropout': 0.1, 'activation_dropout': 0.0, 'feat_proj_dropout': 0.1, 'final_dropout': 0.0, 'layerdrop': 0.0, 'layer_norm_eps': 1e-05, 'initializer_range': 0.02, 'vocab_size': 32, 'do_stable_layer_norm': False, 'use_weighted_layer_sum': False, 'apply_spec_augment': True, 'mask_time_prob': 0.05, 'mask_time_length': 10, 'mask_time_min_masks': 2, 'mask_feature_prob': 0.0, 'mask_feature_length': 10, 'mask_feature_min_masks': 0, 'num_codevectors_per_group': 320, 'num_codevector_groups': 2, 'contrastive_logits_temperature': 0.1, 'feat_quantizer_dropout': 0.0, 'num_negatives': 100, 'codevector_dim': 256, 'proj_codevector_dim': 256, 'diversity_loss_weight': 0.1, 'ctc_loss_reduction': 'sum', 'ctc_zero_infinity': False, 'add_adapter': False, 'adapter_kernel_size': 3, 'adapter_stride': 2, 'num_adapter_layers': 3, 'output_hidden_size': 768, 'adapter_attn_dim': None, 'classifier_proj_size': 256, 'tdnn_dim': [512, 512, 512, 512, 1500], 'tdnn_kernel': [5, 3, 3, 1, 1], 'tdnn_dilation': [1, 2, 3, 1, 1], 'xvector_output_dim': 512, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 64, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 0.0001, 'weight_decay': 0.005, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 30, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 1000, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/May24_20-40-48_890e8b3ca76b', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 2, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': 500, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': True, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': True, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None}
|
| 147 |
+
2024-05-24 20:41:39,043 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 148 |
+
2024-05-24 20:41:39,044 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 149 |
+
2024-05-24 20:42:26,950 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 150 |
+
2024-05-24 20:42:27,128 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 151 |
+
2024-05-24 20:42:27,128 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 152 |
+
2024-05-24 20:42:29,871 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 153 |
+
2024-05-24 20:43:07,276 INFO MainThread:34 [wandb_run.py:_config_callback():1347] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['Wav2Vec2ForPreTraining'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'facebook/wav2vec2-base', 'transformers_version': '4.39.3', 'freeze_feat_extract_train': True, 'mask_channel_length': 10, 'mask_channel_min_space': 1, 'mask_channel_other': 0.0, 'mask_channel_prob': 0.0, 'mask_channel_selection': 'static', 'mask_time_min_space': 1, 'mask_time_other': 0.0, 'mask_time_selection': 'static', 'model_type': 'wav2vec2', 'no_mask_channel_overlap': False, 'no_mask_time_overlap': False, 'num_feat_extract_layers': 7, 'hidden_size': 768, 'feat_extract_norm': 'group', 'feat_extract_activation': 'gelu', 'conv_dim': [512, 512, 512, 512, 512, 512, 512], 'conv_stride': [5, 2, 2, 2, 2, 2, 2], 'conv_kernel': [10, 3, 3, 3, 3, 2, 2], 'conv_bias': False, 'num_conv_pos_embeddings': 128, 'num_conv_pos_embedding_groups': 16, 'num_hidden_layers': 12, 'intermediate_size': 3072, 'hidden_act': 'gelu', 'num_attention_heads': 12, 'hidden_dropout': 0.1, 'attention_dropout': 0.1, 'activation_dropout': 0.0, 'feat_proj_dropout': 0.1, 'final_dropout': 0.0, 'layerdrop': 0.0, 'layer_norm_eps': 1e-05, 'initializer_range': 0.02, 'vocab_size': 32, 'do_stable_layer_norm': False, 'use_weighted_layer_sum': False, 'apply_spec_augment': True, 'mask_time_prob': 0.05, 'mask_time_length': 10, 'mask_time_min_masks': 2, 'mask_feature_prob': 0.0, 'mask_feature_length': 10, 'mask_feature_min_masks': 0, 'num_codevectors_per_group': 320, 'num_codevector_groups': 2, 'contrastive_logits_temperature': 0.1, 'feat_quantizer_dropout': 0.0, 'num_negatives': 100, 'codevector_dim': 256, 'proj_codevector_dim': 256, 'diversity_loss_weight': 0.1, 'ctc_loss_reduction': 'sum', 'ctc_zero_infinity': False, 'add_adapter': False, 'adapter_kernel_size': 3, 'adapter_stride': 2, 'num_adapter_layers': 3, 'output_hidden_size': 768, 'adapter_attn_dim': None, 'classifier_proj_size': 256, 'tdnn_dim': [512, 512, 512, 512, 1500], 'tdnn_kernel': [5, 3, 3, 1, 1], 'tdnn_dilation': [1, 2, 3, 1, 1], 'xvector_output_dim': 512, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 16, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 0.0001, 'weight_decay': 0.005, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 30, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 1000, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/May24_20-42-26_890e8b3ca76b', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 2, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': 500, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': True, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': True, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None}
|
| 154 |
+
2024-05-24 20:43:12,247 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 155 |
+
2024-05-24 20:43:12,247 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 156 |
+
2024-05-24 20:44:08,643 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 157 |
+
2024-05-24 20:44:14,294 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 158 |
+
2024-05-24 20:44:14,295 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 159 |
+
2024-05-24 20:44:39,719 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 160 |
+
2024-05-24 20:44:39,770 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 161 |
+
2024-05-24 20:44:39,770 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 162 |
+
2024-05-24 20:44:43,543 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 163 |
+
2024-05-24 20:46:31,143 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 164 |
+
2024-05-24 20:46:31,143 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 165 |
+
2024-05-24 20:46:49,361 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 166 |
+
2024-05-24 20:46:49,391 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 167 |
+
2024-05-24 20:46:49,391 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 168 |
+
2024-05-24 20:46:52,169 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 169 |
+
2024-05-24 20:46:54,391 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 170 |
+
2024-05-24 20:46:54,391 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 171 |
+
2024-05-24 20:46:58,317 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 172 |
+
2024-05-24 20:46:58,322 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 173 |
+
2024-05-24 20:46:58,322 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 174 |
+
2024-05-24 20:47:06,482 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 175 |
+
2024-05-24 20:47:07,830 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 176 |
+
2024-05-24 20:47:07,830 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 177 |
+
2024-05-24 20:47:10,833 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 178 |
+
2024-05-24 20:47:10,835 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 179 |
+
2024-05-24 20:47:10,839 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 180 |
+
2024-05-24 20:47:14,002 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 181 |
+
2024-05-24 20:48:02,460 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 182 |
+
2024-05-24 20:48:02,460 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 183 |
+
2024-05-24 20:48:05,731 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 184 |
+
2024-05-24 20:48:05,733 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 185 |
+
2024-05-24 20:48:05,734 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 186 |
+
2024-05-24 20:48:08,287 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 187 |
+
2024-05-24 20:48:08,714 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 188 |
+
2024-05-24 20:48:08,714 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 189 |
+
2024-05-24 20:48:10,290 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 190 |
+
2024-05-24 20:48:10,293 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 191 |
+
2024-05-24 20:48:10,294 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 192 |
+
2024-05-24 20:48:13,841 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 193 |
+
2024-05-24 20:48:14,091 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 194 |
+
2024-05-24 20:48:14,092 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 195 |
+
2024-05-24 20:48:16,125 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 196 |
+
2024-05-24 20:48:40,373 INFO MainThread:34 [wandb_run.py:_config_callback():1347] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['Wav2Vec2ForPreTraining'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'facebook/wav2vec2-base', 'transformers_version': '4.39.3', 'freeze_feat_extract_train': True, 'mask_channel_length': 10, 'mask_channel_min_space': 1, 'mask_channel_other': 0.0, 'mask_channel_prob': 0.0, 'mask_channel_selection': 'static', 'mask_time_min_space': 1, 'mask_time_other': 0.0, 'mask_time_selection': 'static', 'model_type': 'wav2vec2', 'no_mask_channel_overlap': False, 'no_mask_time_overlap': False, 'num_feat_extract_layers': 7, 'hidden_size': 768, 'feat_extract_norm': 'group', 'feat_extract_activation': 'gelu', 'conv_dim': [512, 512, 512, 512, 512, 512, 512], 'conv_stride': [5, 2, 2, 2, 2, 2, 2], 'conv_kernel': [10, 3, 3, 3, 3, 2, 2], 'conv_bias': False, 'num_conv_pos_embeddings': 128, 'num_conv_pos_embedding_groups': 16, 'num_hidden_layers': 12, 'intermediate_size': 3072, 'hidden_act': 'gelu', 'num_attention_heads': 12, 'hidden_dropout': 0.1, 'attention_dropout': 0.1, 'activation_dropout': 0.0, 'feat_proj_dropout': 0.1, 'final_dropout': 0.0, 'layerdrop': 0.0, 'layer_norm_eps': 1e-05, 'initializer_range': 0.02, 'vocab_size': 32, 'do_stable_layer_norm': False, 'use_weighted_layer_sum': False, 'apply_spec_augment': True, 'mask_time_prob': 0.05, 'mask_time_length': 10, 'mask_time_min_masks': 2, 'mask_feature_prob': 0.0, 'mask_feature_length': 10, 'mask_feature_min_masks': 0, 'num_codevectors_per_group': 320, 'num_codevector_groups': 2, 'contrastive_logits_temperature': 0.1, 'feat_quantizer_dropout': 0.0, 'num_negatives': 100, 'codevector_dim': 256, 'proj_codevector_dim': 256, 'diversity_loss_weight': 0.1, 'ctc_loss_reduction': 'sum', 'ctc_zero_infinity': False, 'add_adapter': False, 'adapter_kernel_size': 3, 'adapter_stride': 2, 'num_adapter_layers': 3, 'output_hidden_size': 768, 'adapter_attn_dim': None, 'classifier_proj_size': 256, 'tdnn_dim': [512, 512, 512, 512, 1500], 'tdnn_kernel': [5, 3, 3, 1, 1], 'tdnn_dilation': [1, 2, 3, 1, 1], 'xvector_output_dim': 512, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 16, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 0.0001, 'weight_decay': 0.005, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 30, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 1000, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/May24_20-48-13_890e8b3ca76b', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 2, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': 500, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': True, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': True, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None}
|
| 197 |
+
2024-05-24 20:48:42,631 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 198 |
+
2024-05-24 20:48:42,632 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 199 |
+
2024-05-24 20:49:00,381 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 200 |
+
2024-05-24 20:49:00,425 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 201 |
+
2024-05-24 20:49:00,425 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 202 |
+
2024-05-24 20:49:03,721 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 203 |
+
2024-05-24 20:49:03,725 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 204 |
+
2024-05-24 20:49:03,726 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 205 |
+
2024-05-24 20:49:08,089 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 206 |
+
2024-05-24 20:49:08,455 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 207 |
+
2024-05-24 20:49:08,456 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 208 |
+
2024-05-24 20:51:52,779 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 209 |
+
2024-05-24 20:51:52,781 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 210 |
+
2024-05-24 20:51:52,781 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 211 |
+
2024-05-24 20:51:56,593 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 212 |
+
2024-05-24 20:52:22,759 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 213 |
+
2024-05-24 20:52:22,759 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 214 |
+
2024-05-24 20:52:49,570 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 215 |
+
2024-05-24 20:52:49,572 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 216 |
+
2024-05-24 20:52:49,572 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 217 |
+
2024-05-24 20:52:52,107 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
wandb/run-20240524_202737-n1w0kmmv/files/conda-environment.yaml
ADDED
|
File without changes
|
wandb/run-20240524_202737-n1w0kmmv/files/config.yaml
ADDED
|
@@ -0,0 +1,866 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
wandb_version: 1
|
| 2 |
+
|
| 3 |
+
_wandb:
|
| 4 |
+
desc: null
|
| 5 |
+
value:
|
| 6 |
+
python_version: 3.10.13
|
| 7 |
+
cli_version: 0.16.6
|
| 8 |
+
framework: huggingface
|
| 9 |
+
huggingface_version: 4.39.3
|
| 10 |
+
is_jupyter_run: true
|
| 11 |
+
is_kaggle_kernel: true
|
| 12 |
+
start_time: 1716582457.0
|
| 13 |
+
t:
|
| 14 |
+
1:
|
| 15 |
+
- 1
|
| 16 |
+
- 2
|
| 17 |
+
- 3
|
| 18 |
+
- 5
|
| 19 |
+
- 11
|
| 20 |
+
- 12
|
| 21 |
+
- 49
|
| 22 |
+
- 51
|
| 23 |
+
- 53
|
| 24 |
+
- 55
|
| 25 |
+
- 71
|
| 26 |
+
- 105
|
| 27 |
+
2:
|
| 28 |
+
- 1
|
| 29 |
+
- 2
|
| 30 |
+
- 3
|
| 31 |
+
- 5
|
| 32 |
+
- 11
|
| 33 |
+
- 12
|
| 34 |
+
- 49
|
| 35 |
+
- 51
|
| 36 |
+
- 53
|
| 37 |
+
- 55
|
| 38 |
+
- 71
|
| 39 |
+
- 105
|
| 40 |
+
3:
|
| 41 |
+
- 7
|
| 42 |
+
- 23
|
| 43 |
+
4: 3.10.13
|
| 44 |
+
5: 0.16.6
|
| 45 |
+
6: 4.39.3
|
| 46 |
+
8:
|
| 47 |
+
- 1
|
| 48 |
+
- 2
|
| 49 |
+
- 5
|
| 50 |
+
9:
|
| 51 |
+
1: transformers_trainer
|
| 52 |
+
13: linux-x86_64
|
| 53 |
+
m:
|
| 54 |
+
- 1: train/global_step
|
| 55 |
+
6:
|
| 56 |
+
- 3
|
| 57 |
+
- 1: train/loss
|
| 58 |
+
5: 1
|
| 59 |
+
6:
|
| 60 |
+
- 1
|
| 61 |
+
- 1: train/grad_norm
|
| 62 |
+
5: 1
|
| 63 |
+
6:
|
| 64 |
+
- 1
|
| 65 |
+
- 1: train/learning_rate
|
| 66 |
+
5: 1
|
| 67 |
+
6:
|
| 68 |
+
- 1
|
| 69 |
+
- 1: train/epoch
|
| 70 |
+
5: 1
|
| 71 |
+
6:
|
| 72 |
+
- 1
|
| 73 |
+
- 1: eval/loss
|
| 74 |
+
5: 1
|
| 75 |
+
6:
|
| 76 |
+
- 1
|
| 77 |
+
- 1: eval/wer
|
| 78 |
+
5: 1
|
| 79 |
+
6:
|
| 80 |
+
- 1
|
| 81 |
+
- 1: eval/runtime
|
| 82 |
+
5: 1
|
| 83 |
+
6:
|
| 84 |
+
- 1
|
| 85 |
+
- 1: eval/samples_per_second
|
| 86 |
+
5: 1
|
| 87 |
+
6:
|
| 88 |
+
- 1
|
| 89 |
+
- 1: eval/steps_per_second
|
| 90 |
+
5: 1
|
| 91 |
+
6:
|
| 92 |
+
- 1
|
| 93 |
+
return_dict:
|
| 94 |
+
desc: null
|
| 95 |
+
value: true
|
| 96 |
+
output_hidden_states:
|
| 97 |
+
desc: null
|
| 98 |
+
value: false
|
| 99 |
+
output_attentions:
|
| 100 |
+
desc: null
|
| 101 |
+
value: false
|
| 102 |
+
torchscript:
|
| 103 |
+
desc: null
|
| 104 |
+
value: false
|
| 105 |
+
torch_dtype:
|
| 106 |
+
desc: null
|
| 107 |
+
value: null
|
| 108 |
+
use_bfloat16:
|
| 109 |
+
desc: null
|
| 110 |
+
value: false
|
| 111 |
+
tf_legacy_loss:
|
| 112 |
+
desc: null
|
| 113 |
+
value: false
|
| 114 |
+
pruned_heads:
|
| 115 |
+
desc: null
|
| 116 |
+
value: {}
|
| 117 |
+
tie_word_embeddings:
|
| 118 |
+
desc: null
|
| 119 |
+
value: true
|
| 120 |
+
chunk_size_feed_forward:
|
| 121 |
+
desc: null
|
| 122 |
+
value: 0
|
| 123 |
+
is_encoder_decoder:
|
| 124 |
+
desc: null
|
| 125 |
+
value: false
|
| 126 |
+
is_decoder:
|
| 127 |
+
desc: null
|
| 128 |
+
value: false
|
| 129 |
+
cross_attention_hidden_size:
|
| 130 |
+
desc: null
|
| 131 |
+
value: null
|
| 132 |
+
add_cross_attention:
|
| 133 |
+
desc: null
|
| 134 |
+
value: false
|
| 135 |
+
tie_encoder_decoder:
|
| 136 |
+
desc: null
|
| 137 |
+
value: false
|
| 138 |
+
max_length:
|
| 139 |
+
desc: null
|
| 140 |
+
value: 20
|
| 141 |
+
min_length:
|
| 142 |
+
desc: null
|
| 143 |
+
value: 0
|
| 144 |
+
do_sample:
|
| 145 |
+
desc: null
|
| 146 |
+
value: false
|
| 147 |
+
early_stopping:
|
| 148 |
+
desc: null
|
| 149 |
+
value: false
|
| 150 |
+
num_beams:
|
| 151 |
+
desc: null
|
| 152 |
+
value: 1
|
| 153 |
+
num_beam_groups:
|
| 154 |
+
desc: null
|
| 155 |
+
value: 1
|
| 156 |
+
diversity_penalty:
|
| 157 |
+
desc: null
|
| 158 |
+
value: 0.0
|
| 159 |
+
temperature:
|
| 160 |
+
desc: null
|
| 161 |
+
value: 1.0
|
| 162 |
+
top_k:
|
| 163 |
+
desc: null
|
| 164 |
+
value: 50
|
| 165 |
+
top_p:
|
| 166 |
+
desc: null
|
| 167 |
+
value: 1.0
|
| 168 |
+
typical_p:
|
| 169 |
+
desc: null
|
| 170 |
+
value: 1.0
|
| 171 |
+
repetition_penalty:
|
| 172 |
+
desc: null
|
| 173 |
+
value: 1.0
|
| 174 |
+
length_penalty:
|
| 175 |
+
desc: null
|
| 176 |
+
value: 1.0
|
| 177 |
+
no_repeat_ngram_size:
|
| 178 |
+
desc: null
|
| 179 |
+
value: 0
|
| 180 |
+
encoder_no_repeat_ngram_size:
|
| 181 |
+
desc: null
|
| 182 |
+
value: 0
|
| 183 |
+
bad_words_ids:
|
| 184 |
+
desc: null
|
| 185 |
+
value: null
|
| 186 |
+
num_return_sequences:
|
| 187 |
+
desc: null
|
| 188 |
+
value: 1
|
| 189 |
+
output_scores:
|
| 190 |
+
desc: null
|
| 191 |
+
value: false
|
| 192 |
+
return_dict_in_generate:
|
| 193 |
+
desc: null
|
| 194 |
+
value: false
|
| 195 |
+
forced_bos_token_id:
|
| 196 |
+
desc: null
|
| 197 |
+
value: null
|
| 198 |
+
forced_eos_token_id:
|
| 199 |
+
desc: null
|
| 200 |
+
value: null
|
| 201 |
+
remove_invalid_values:
|
| 202 |
+
desc: null
|
| 203 |
+
value: false
|
| 204 |
+
exponential_decay_length_penalty:
|
| 205 |
+
desc: null
|
| 206 |
+
value: null
|
| 207 |
+
suppress_tokens:
|
| 208 |
+
desc: null
|
| 209 |
+
value: null
|
| 210 |
+
begin_suppress_tokens:
|
| 211 |
+
desc: null
|
| 212 |
+
value: null
|
| 213 |
+
architectures:
|
| 214 |
+
desc: null
|
| 215 |
+
value:
|
| 216 |
+
- Wav2Vec2ForPreTraining
|
| 217 |
+
finetuning_task:
|
| 218 |
+
desc: null
|
| 219 |
+
value: null
|
| 220 |
+
id2label:
|
| 221 |
+
desc: null
|
| 222 |
+
value:
|
| 223 |
+
'0': LABEL_0
|
| 224 |
+
'1': LABEL_1
|
| 225 |
+
label2id:
|
| 226 |
+
desc: null
|
| 227 |
+
value:
|
| 228 |
+
LABEL_0: 0
|
| 229 |
+
LABEL_1: 1
|
| 230 |
+
tokenizer_class:
|
| 231 |
+
desc: null
|
| 232 |
+
value: null
|
| 233 |
+
prefix:
|
| 234 |
+
desc: null
|
| 235 |
+
value: null
|
| 236 |
+
bos_token_id:
|
| 237 |
+
desc: null
|
| 238 |
+
value: 1
|
| 239 |
+
pad_token_id:
|
| 240 |
+
desc: null
|
| 241 |
+
value: 0
|
| 242 |
+
eos_token_id:
|
| 243 |
+
desc: null
|
| 244 |
+
value: 2
|
| 245 |
+
sep_token_id:
|
| 246 |
+
desc: null
|
| 247 |
+
value: null
|
| 248 |
+
decoder_start_token_id:
|
| 249 |
+
desc: null
|
| 250 |
+
value: null
|
| 251 |
+
task_specific_params:
|
| 252 |
+
desc: null
|
| 253 |
+
value: null
|
| 254 |
+
problem_type:
|
| 255 |
+
desc: null
|
| 256 |
+
value: null
|
| 257 |
+
_name_or_path:
|
| 258 |
+
desc: null
|
| 259 |
+
value: facebook/wav2vec2-base
|
| 260 |
+
transformers_version:
|
| 261 |
+
desc: null
|
| 262 |
+
value: 4.39.3
|
| 263 |
+
freeze_feat_extract_train:
|
| 264 |
+
desc: null
|
| 265 |
+
value: true
|
| 266 |
+
mask_channel_length:
|
| 267 |
+
desc: null
|
| 268 |
+
value: 10
|
| 269 |
+
mask_channel_min_space:
|
| 270 |
+
desc: null
|
| 271 |
+
value: 1
|
| 272 |
+
mask_channel_other:
|
| 273 |
+
desc: null
|
| 274 |
+
value: 0.0
|
| 275 |
+
mask_channel_prob:
|
| 276 |
+
desc: null
|
| 277 |
+
value: 0.0
|
| 278 |
+
mask_channel_selection:
|
| 279 |
+
desc: null
|
| 280 |
+
value: static
|
| 281 |
+
mask_time_min_space:
|
| 282 |
+
desc: null
|
| 283 |
+
value: 1
|
| 284 |
+
mask_time_other:
|
| 285 |
+
desc: null
|
| 286 |
+
value: 0.0
|
| 287 |
+
mask_time_selection:
|
| 288 |
+
desc: null
|
| 289 |
+
value: static
|
| 290 |
+
model_type:
|
| 291 |
+
desc: null
|
| 292 |
+
value: wav2vec2
|
| 293 |
+
no_mask_channel_overlap:
|
| 294 |
+
desc: null
|
| 295 |
+
value: false
|
| 296 |
+
no_mask_time_overlap:
|
| 297 |
+
desc: null
|
| 298 |
+
value: false
|
| 299 |
+
num_feat_extract_layers:
|
| 300 |
+
desc: null
|
| 301 |
+
value: 7
|
| 302 |
+
hidden_size:
|
| 303 |
+
desc: null
|
| 304 |
+
value: 768
|
| 305 |
+
feat_extract_norm:
|
| 306 |
+
desc: null
|
| 307 |
+
value: group
|
| 308 |
+
feat_extract_activation:
|
| 309 |
+
desc: null
|
| 310 |
+
value: gelu
|
| 311 |
+
conv_dim:
|
| 312 |
+
desc: null
|
| 313 |
+
value:
|
| 314 |
+
- 512
|
| 315 |
+
- 512
|
| 316 |
+
- 512
|
| 317 |
+
- 512
|
| 318 |
+
- 512
|
| 319 |
+
- 512
|
| 320 |
+
- 512
|
| 321 |
+
conv_stride:
|
| 322 |
+
desc: null
|
| 323 |
+
value:
|
| 324 |
+
- 5
|
| 325 |
+
- 2
|
| 326 |
+
- 2
|
| 327 |
+
- 2
|
| 328 |
+
- 2
|
| 329 |
+
- 2
|
| 330 |
+
- 2
|
| 331 |
+
conv_kernel:
|
| 332 |
+
desc: null
|
| 333 |
+
value:
|
| 334 |
+
- 10
|
| 335 |
+
- 3
|
| 336 |
+
- 3
|
| 337 |
+
- 3
|
| 338 |
+
- 3
|
| 339 |
+
- 2
|
| 340 |
+
- 2
|
| 341 |
+
conv_bias:
|
| 342 |
+
desc: null
|
| 343 |
+
value: false
|
| 344 |
+
num_conv_pos_embeddings:
|
| 345 |
+
desc: null
|
| 346 |
+
value: 128
|
| 347 |
+
num_conv_pos_embedding_groups:
|
| 348 |
+
desc: null
|
| 349 |
+
value: 16
|
| 350 |
+
num_hidden_layers:
|
| 351 |
+
desc: null
|
| 352 |
+
value: 12
|
| 353 |
+
intermediate_size:
|
| 354 |
+
desc: null
|
| 355 |
+
value: 3072
|
| 356 |
+
hidden_act:
|
| 357 |
+
desc: null
|
| 358 |
+
value: gelu
|
| 359 |
+
num_attention_heads:
|
| 360 |
+
desc: null
|
| 361 |
+
value: 12
|
| 362 |
+
hidden_dropout:
|
| 363 |
+
desc: null
|
| 364 |
+
value: 0.1
|
| 365 |
+
attention_dropout:
|
| 366 |
+
desc: null
|
| 367 |
+
value: 0.1
|
| 368 |
+
activation_dropout:
|
| 369 |
+
desc: null
|
| 370 |
+
value: 0.0
|
| 371 |
+
feat_proj_dropout:
|
| 372 |
+
desc: null
|
| 373 |
+
value: 0.1
|
| 374 |
+
final_dropout:
|
| 375 |
+
desc: null
|
| 376 |
+
value: 0.0
|
| 377 |
+
layerdrop:
|
| 378 |
+
desc: null
|
| 379 |
+
value: 0.0
|
| 380 |
+
layer_norm_eps:
|
| 381 |
+
desc: null
|
| 382 |
+
value: 1.0e-05
|
| 383 |
+
initializer_range:
|
| 384 |
+
desc: null
|
| 385 |
+
value: 0.02
|
| 386 |
+
vocab_size:
|
| 387 |
+
desc: null
|
| 388 |
+
value: 32
|
| 389 |
+
do_stable_layer_norm:
|
| 390 |
+
desc: null
|
| 391 |
+
value: false
|
| 392 |
+
use_weighted_layer_sum:
|
| 393 |
+
desc: null
|
| 394 |
+
value: false
|
| 395 |
+
apply_spec_augment:
|
| 396 |
+
desc: null
|
| 397 |
+
value: true
|
| 398 |
+
mask_time_prob:
|
| 399 |
+
desc: null
|
| 400 |
+
value: 0.05
|
| 401 |
+
mask_time_length:
|
| 402 |
+
desc: null
|
| 403 |
+
value: 10
|
| 404 |
+
mask_time_min_masks:
|
| 405 |
+
desc: null
|
| 406 |
+
value: 2
|
| 407 |
+
mask_feature_prob:
|
| 408 |
+
desc: null
|
| 409 |
+
value: 0.0
|
| 410 |
+
mask_feature_length:
|
| 411 |
+
desc: null
|
| 412 |
+
value: 10
|
| 413 |
+
mask_feature_min_masks:
|
| 414 |
+
desc: null
|
| 415 |
+
value: 0
|
| 416 |
+
num_codevectors_per_group:
|
| 417 |
+
desc: null
|
| 418 |
+
value: 320
|
| 419 |
+
num_codevector_groups:
|
| 420 |
+
desc: null
|
| 421 |
+
value: 2
|
| 422 |
+
contrastive_logits_temperature:
|
| 423 |
+
desc: null
|
| 424 |
+
value: 0.1
|
| 425 |
+
feat_quantizer_dropout:
|
| 426 |
+
desc: null
|
| 427 |
+
value: 0.0
|
| 428 |
+
num_negatives:
|
| 429 |
+
desc: null
|
| 430 |
+
value: 100
|
| 431 |
+
codevector_dim:
|
| 432 |
+
desc: null
|
| 433 |
+
value: 256
|
| 434 |
+
proj_codevector_dim:
|
| 435 |
+
desc: null
|
| 436 |
+
value: 256
|
| 437 |
+
diversity_loss_weight:
|
| 438 |
+
desc: null
|
| 439 |
+
value: 0.1
|
| 440 |
+
ctc_loss_reduction:
|
| 441 |
+
desc: null
|
| 442 |
+
value: sum
|
| 443 |
+
ctc_zero_infinity:
|
| 444 |
+
desc: null
|
| 445 |
+
value: false
|
| 446 |
+
add_adapter:
|
| 447 |
+
desc: null
|
| 448 |
+
value: false
|
| 449 |
+
adapter_kernel_size:
|
| 450 |
+
desc: null
|
| 451 |
+
value: 3
|
| 452 |
+
adapter_stride:
|
| 453 |
+
desc: null
|
| 454 |
+
value: 2
|
| 455 |
+
num_adapter_layers:
|
| 456 |
+
desc: null
|
| 457 |
+
value: 3
|
| 458 |
+
output_hidden_size:
|
| 459 |
+
desc: null
|
| 460 |
+
value: 768
|
| 461 |
+
adapter_attn_dim:
|
| 462 |
+
desc: null
|
| 463 |
+
value: null
|
| 464 |
+
classifier_proj_size:
|
| 465 |
+
desc: null
|
| 466 |
+
value: 256
|
| 467 |
+
tdnn_dim:
|
| 468 |
+
desc: null
|
| 469 |
+
value:
|
| 470 |
+
- 512
|
| 471 |
+
- 512
|
| 472 |
+
- 512
|
| 473 |
+
- 512
|
| 474 |
+
- 1500
|
| 475 |
+
tdnn_kernel:
|
| 476 |
+
desc: null
|
| 477 |
+
value:
|
| 478 |
+
- 5
|
| 479 |
+
- 3
|
| 480 |
+
- 3
|
| 481 |
+
- 1
|
| 482 |
+
- 1
|
| 483 |
+
tdnn_dilation:
|
| 484 |
+
desc: null
|
| 485 |
+
value:
|
| 486 |
+
- 1
|
| 487 |
+
- 2
|
| 488 |
+
- 3
|
| 489 |
+
- 1
|
| 490 |
+
- 1
|
| 491 |
+
xvector_output_dim:
|
| 492 |
+
desc: null
|
| 493 |
+
value: 512
|
| 494 |
+
output_dir:
|
| 495 |
+
desc: null
|
| 496 |
+
value: /kaggle/working/
|
| 497 |
+
overwrite_output_dir:
|
| 498 |
+
desc: null
|
| 499 |
+
value: false
|
| 500 |
+
do_train:
|
| 501 |
+
desc: null
|
| 502 |
+
value: false
|
| 503 |
+
do_eval:
|
| 504 |
+
desc: null
|
| 505 |
+
value: true
|
| 506 |
+
do_predict:
|
| 507 |
+
desc: null
|
| 508 |
+
value: false
|
| 509 |
+
evaluation_strategy:
|
| 510 |
+
desc: null
|
| 511 |
+
value: steps
|
| 512 |
+
prediction_loss_only:
|
| 513 |
+
desc: null
|
| 514 |
+
value: false
|
| 515 |
+
per_device_train_batch_size:
|
| 516 |
+
desc: null
|
| 517 |
+
value: 16
|
| 518 |
+
per_device_eval_batch_size:
|
| 519 |
+
desc: null
|
| 520 |
+
value: 8
|
| 521 |
+
per_gpu_train_batch_size:
|
| 522 |
+
desc: null
|
| 523 |
+
value: null
|
| 524 |
+
per_gpu_eval_batch_size:
|
| 525 |
+
desc: null
|
| 526 |
+
value: null
|
| 527 |
+
gradient_accumulation_steps:
|
| 528 |
+
desc: null
|
| 529 |
+
value: 1
|
| 530 |
+
eval_accumulation_steps:
|
| 531 |
+
desc: null
|
| 532 |
+
value: null
|
| 533 |
+
eval_delay:
|
| 534 |
+
desc: null
|
| 535 |
+
value: 0
|
| 536 |
+
learning_rate:
|
| 537 |
+
desc: null
|
| 538 |
+
value: 0.0001
|
| 539 |
+
weight_decay:
|
| 540 |
+
desc: null
|
| 541 |
+
value: 0.005
|
| 542 |
+
adam_beta1:
|
| 543 |
+
desc: null
|
| 544 |
+
value: 0.9
|
| 545 |
+
adam_beta2:
|
| 546 |
+
desc: null
|
| 547 |
+
value: 0.999
|
| 548 |
+
adam_epsilon:
|
| 549 |
+
desc: null
|
| 550 |
+
value: 1.0e-08
|
| 551 |
+
max_grad_norm:
|
| 552 |
+
desc: null
|
| 553 |
+
value: 1.0
|
| 554 |
+
num_train_epochs:
|
| 555 |
+
desc: null
|
| 556 |
+
value: 30
|
| 557 |
+
max_steps:
|
| 558 |
+
desc: null
|
| 559 |
+
value: -1
|
| 560 |
+
lr_scheduler_type:
|
| 561 |
+
desc: null
|
| 562 |
+
value: linear
|
| 563 |
+
lr_scheduler_kwargs:
|
| 564 |
+
desc: null
|
| 565 |
+
value: {}
|
| 566 |
+
warmup_ratio:
|
| 567 |
+
desc: null
|
| 568 |
+
value: 0.0
|
| 569 |
+
warmup_steps:
|
| 570 |
+
desc: null
|
| 571 |
+
value: 1000
|
| 572 |
+
log_level:
|
| 573 |
+
desc: null
|
| 574 |
+
value: passive
|
| 575 |
+
log_level_replica:
|
| 576 |
+
desc: null
|
| 577 |
+
value: warning
|
| 578 |
+
log_on_each_node:
|
| 579 |
+
desc: null
|
| 580 |
+
value: true
|
| 581 |
+
logging_dir:
|
| 582 |
+
desc: null
|
| 583 |
+
value: /kaggle/working/runs/May24_20-48-13_890e8b3ca76b
|
| 584 |
+
logging_strategy:
|
| 585 |
+
desc: null
|
| 586 |
+
value: steps
|
| 587 |
+
logging_first_step:
|
| 588 |
+
desc: null
|
| 589 |
+
value: false
|
| 590 |
+
logging_steps:
|
| 591 |
+
desc: null
|
| 592 |
+
value: 500
|
| 593 |
+
logging_nan_inf_filter:
|
| 594 |
+
desc: null
|
| 595 |
+
value: true
|
| 596 |
+
save_strategy:
|
| 597 |
+
desc: null
|
| 598 |
+
value: steps
|
| 599 |
+
save_steps:
|
| 600 |
+
desc: null
|
| 601 |
+
value: 500
|
| 602 |
+
save_total_limit:
|
| 603 |
+
desc: null
|
| 604 |
+
value: 2
|
| 605 |
+
save_safetensors:
|
| 606 |
+
desc: null
|
| 607 |
+
value: true
|
| 608 |
+
save_on_each_node:
|
| 609 |
+
desc: null
|
| 610 |
+
value: false
|
| 611 |
+
save_only_model:
|
| 612 |
+
desc: null
|
| 613 |
+
value: false
|
| 614 |
+
no_cuda:
|
| 615 |
+
desc: null
|
| 616 |
+
value: false
|
| 617 |
+
use_cpu:
|
| 618 |
+
desc: null
|
| 619 |
+
value: false
|
| 620 |
+
use_mps_device:
|
| 621 |
+
desc: null
|
| 622 |
+
value: false
|
| 623 |
+
seed:
|
| 624 |
+
desc: null
|
| 625 |
+
value: 42
|
| 626 |
+
data_seed:
|
| 627 |
+
desc: null
|
| 628 |
+
value: null
|
| 629 |
+
jit_mode_eval:
|
| 630 |
+
desc: null
|
| 631 |
+
value: false
|
| 632 |
+
use_ipex:
|
| 633 |
+
desc: null
|
| 634 |
+
value: false
|
| 635 |
+
bf16:
|
| 636 |
+
desc: null
|
| 637 |
+
value: false
|
| 638 |
+
fp16:
|
| 639 |
+
desc: null
|
| 640 |
+
value: true
|
| 641 |
+
fp16_opt_level:
|
| 642 |
+
desc: null
|
| 643 |
+
value: O1
|
| 644 |
+
half_precision_backend:
|
| 645 |
+
desc: null
|
| 646 |
+
value: auto
|
| 647 |
+
bf16_full_eval:
|
| 648 |
+
desc: null
|
| 649 |
+
value: false
|
| 650 |
+
fp16_full_eval:
|
| 651 |
+
desc: null
|
| 652 |
+
value: false
|
| 653 |
+
tf32:
|
| 654 |
+
desc: null
|
| 655 |
+
value: null
|
| 656 |
+
local_rank:
|
| 657 |
+
desc: null
|
| 658 |
+
value: 0
|
| 659 |
+
ddp_backend:
|
| 660 |
+
desc: null
|
| 661 |
+
value: null
|
| 662 |
+
tpu_num_cores:
|
| 663 |
+
desc: null
|
| 664 |
+
value: null
|
| 665 |
+
tpu_metrics_debug:
|
| 666 |
+
desc: null
|
| 667 |
+
value: false
|
| 668 |
+
debug:
|
| 669 |
+
desc: null
|
| 670 |
+
value: []
|
| 671 |
+
dataloader_drop_last:
|
| 672 |
+
desc: null
|
| 673 |
+
value: false
|
| 674 |
+
eval_steps:
|
| 675 |
+
desc: null
|
| 676 |
+
value: 500
|
| 677 |
+
dataloader_num_workers:
|
| 678 |
+
desc: null
|
| 679 |
+
value: 0
|
| 680 |
+
dataloader_prefetch_factor:
|
| 681 |
+
desc: null
|
| 682 |
+
value: null
|
| 683 |
+
past_index:
|
| 684 |
+
desc: null
|
| 685 |
+
value: -1
|
| 686 |
+
run_name:
|
| 687 |
+
desc: null
|
| 688 |
+
value: /kaggle/working/
|
| 689 |
+
disable_tqdm:
|
| 690 |
+
desc: null
|
| 691 |
+
value: false
|
| 692 |
+
remove_unused_columns:
|
| 693 |
+
desc: null
|
| 694 |
+
value: true
|
| 695 |
+
label_names:
|
| 696 |
+
desc: null
|
| 697 |
+
value: null
|
| 698 |
+
load_best_model_at_end:
|
| 699 |
+
desc: null
|
| 700 |
+
value: false
|
| 701 |
+
metric_for_best_model:
|
| 702 |
+
desc: null
|
| 703 |
+
value: null
|
| 704 |
+
greater_is_better:
|
| 705 |
+
desc: null
|
| 706 |
+
value: null
|
| 707 |
+
ignore_data_skip:
|
| 708 |
+
desc: null
|
| 709 |
+
value: false
|
| 710 |
+
fsdp:
|
| 711 |
+
desc: null
|
| 712 |
+
value: []
|
| 713 |
+
fsdp_min_num_params:
|
| 714 |
+
desc: null
|
| 715 |
+
value: 0
|
| 716 |
+
fsdp_config:
|
| 717 |
+
desc: null
|
| 718 |
+
value:
|
| 719 |
+
min_num_params: 0
|
| 720 |
+
xla: false
|
| 721 |
+
xla_fsdp_v2: false
|
| 722 |
+
xla_fsdp_grad_ckpt: false
|
| 723 |
+
fsdp_transformer_layer_cls_to_wrap:
|
| 724 |
+
desc: null
|
| 725 |
+
value: null
|
| 726 |
+
accelerator_config:
|
| 727 |
+
desc: null
|
| 728 |
+
value:
|
| 729 |
+
split_batches: false
|
| 730 |
+
dispatch_batches: null
|
| 731 |
+
even_batches: true
|
| 732 |
+
use_seedable_sampler: true
|
| 733 |
+
deepspeed:
|
| 734 |
+
desc: null
|
| 735 |
+
value: null
|
| 736 |
+
label_smoothing_factor:
|
| 737 |
+
desc: null
|
| 738 |
+
value: 0.0
|
| 739 |
+
optim:
|
| 740 |
+
desc: null
|
| 741 |
+
value: adamw_torch
|
| 742 |
+
optim_args:
|
| 743 |
+
desc: null
|
| 744 |
+
value: null
|
| 745 |
+
adafactor:
|
| 746 |
+
desc: null
|
| 747 |
+
value: false
|
| 748 |
+
group_by_length:
|
| 749 |
+
desc: null
|
| 750 |
+
value: true
|
| 751 |
+
length_column_name:
|
| 752 |
+
desc: null
|
| 753 |
+
value: length
|
| 754 |
+
report_to:
|
| 755 |
+
desc: null
|
| 756 |
+
value:
|
| 757 |
+
- tensorboard
|
| 758 |
+
- wandb
|
| 759 |
+
ddp_find_unused_parameters:
|
| 760 |
+
desc: null
|
| 761 |
+
value: null
|
| 762 |
+
ddp_bucket_cap_mb:
|
| 763 |
+
desc: null
|
| 764 |
+
value: null
|
| 765 |
+
ddp_broadcast_buffers:
|
| 766 |
+
desc: null
|
| 767 |
+
value: null
|
| 768 |
+
dataloader_pin_memory:
|
| 769 |
+
desc: null
|
| 770 |
+
value: true
|
| 771 |
+
dataloader_persistent_workers:
|
| 772 |
+
desc: null
|
| 773 |
+
value: false
|
| 774 |
+
skip_memory_metrics:
|
| 775 |
+
desc: null
|
| 776 |
+
value: true
|
| 777 |
+
use_legacy_prediction_loop:
|
| 778 |
+
desc: null
|
| 779 |
+
value: false
|
| 780 |
+
push_to_hub:
|
| 781 |
+
desc: null
|
| 782 |
+
value: true
|
| 783 |
+
resume_from_checkpoint:
|
| 784 |
+
desc: null
|
| 785 |
+
value: null
|
| 786 |
+
hub_model_id:
|
| 787 |
+
desc: null
|
| 788 |
+
value: null
|
| 789 |
+
hub_strategy:
|
| 790 |
+
desc: null
|
| 791 |
+
value: every_save
|
| 792 |
+
hub_token:
|
| 793 |
+
desc: null
|
| 794 |
+
value: <HUB_TOKEN>
|
| 795 |
+
hub_private_repo:
|
| 796 |
+
desc: null
|
| 797 |
+
value: false
|
| 798 |
+
hub_always_push:
|
| 799 |
+
desc: null
|
| 800 |
+
value: false
|
| 801 |
+
gradient_checkpointing:
|
| 802 |
+
desc: null
|
| 803 |
+
value: true
|
| 804 |
+
gradient_checkpointing_kwargs:
|
| 805 |
+
desc: null
|
| 806 |
+
value: null
|
| 807 |
+
include_inputs_for_metrics:
|
| 808 |
+
desc: null
|
| 809 |
+
value: false
|
| 810 |
+
fp16_backend:
|
| 811 |
+
desc: null
|
| 812 |
+
value: auto
|
| 813 |
+
push_to_hub_model_id:
|
| 814 |
+
desc: null
|
| 815 |
+
value: null
|
| 816 |
+
push_to_hub_organization:
|
| 817 |
+
desc: null
|
| 818 |
+
value: null
|
| 819 |
+
push_to_hub_token:
|
| 820 |
+
desc: null
|
| 821 |
+
value: <PUSH_TO_HUB_TOKEN>
|
| 822 |
+
mp_parameters:
|
| 823 |
+
desc: null
|
| 824 |
+
value: ''
|
| 825 |
+
auto_find_batch_size:
|
| 826 |
+
desc: null
|
| 827 |
+
value: false
|
| 828 |
+
full_determinism:
|
| 829 |
+
desc: null
|
| 830 |
+
value: false
|
| 831 |
+
torchdynamo:
|
| 832 |
+
desc: null
|
| 833 |
+
value: null
|
| 834 |
+
ray_scope:
|
| 835 |
+
desc: null
|
| 836 |
+
value: last
|
| 837 |
+
ddp_timeout:
|
| 838 |
+
desc: null
|
| 839 |
+
value: 1800
|
| 840 |
+
torch_compile:
|
| 841 |
+
desc: null
|
| 842 |
+
value: false
|
| 843 |
+
torch_compile_backend:
|
| 844 |
+
desc: null
|
| 845 |
+
value: null
|
| 846 |
+
torch_compile_mode:
|
| 847 |
+
desc: null
|
| 848 |
+
value: null
|
| 849 |
+
dispatch_batches:
|
| 850 |
+
desc: null
|
| 851 |
+
value: null
|
| 852 |
+
split_batches:
|
| 853 |
+
desc: null
|
| 854 |
+
value: null
|
| 855 |
+
include_tokens_per_second:
|
| 856 |
+
desc: null
|
| 857 |
+
value: false
|
| 858 |
+
include_num_input_tokens_seen:
|
| 859 |
+
desc: null
|
| 860 |
+
value: false
|
| 861 |
+
neftune_noise_alpha:
|
| 862 |
+
desc: null
|
| 863 |
+
value: null
|
| 864 |
+
optim_target_modules:
|
| 865 |
+
desc: null
|
| 866 |
+
value: null
|
wandb/run-20240524_202737-n1w0kmmv/files/output.log
ADDED
|
@@ -0,0 +1,171 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
/opt/conda/lib/python3.10/site-packages/transformers/models/wav2vec2/processing_wav2vec2.py:156: UserWarning: `as_target_processor` is deprecated and will be removed in v5 of Transformers. You can process your labels by using the argument `text` of the regular `__call__` method (either in the same call as your audio inputs, or in a separate call.
|
| 2 |
+
warnings.warn(
|
| 3 |
+
/opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
|
| 4 |
+
warnings.warn(
|
| 5 |
+
Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.
|
| 6 |
+
/opt/conda/lib/python3.10/site-packages/transformers/configuration_utils.py:363: UserWarning: Passing `gradient_checkpointing` to a config initialization is deprecated and will be removed in v5 Transformers. Using `model.gradient_checkpointing_enable()` instead, or if you are using the `Trainer` API, pass `gradient_checkpointing=True` in your `TrainingArguments`.
|
| 7 |
+
warnings.warn(
|
| 8 |
+
Some weights of Wav2Vec2ForCTC were not initialized from the model checkpoint at facebook/wav2vec2-base and are newly initialized: ['lm_head.bias', 'lm_head.weight', 'wav2vec2.encoder.pos_conv_embed.conv.parametrizations.weight.original0', 'wav2vec2.encoder.pos_conv_embed.conv.parametrizations.weight.original1']
|
| 9 |
+
You should probably TRAIN this model on a down-stream task to be able to use it for predictions and inference.
|
| 10 |
+
/opt/conda/lib/python3.10/site-packages/datasets/load.py:756: FutureWarning: The repository for wer contains custom code which must be executed to correctly load the metric. You can inspect the repository content at https://raw.githubusercontent.com/huggingface/datasets/2.18.0/metrics/wer/wer.py
|
| 11 |
+
You can avoid this message in future by passing the argument `trust_remote_code=True`.
|
| 12 |
+
Passing `trust_remote_code=True` will be mandatory to load this metric from the next major release of `datasets`.
|
| 13 |
+
warnings.warn(
|
| 14 |
+
/opt/conda/lib/python3.10/site-packages/accelerate/accelerator.py:436: FutureWarning: Passing the following arguments to `Accelerator` is deprecated and will be removed in version 1.0 of Accelerate: dict_keys(['dispatch_batches', 'split_batches', 'even_batches', 'use_seedable_sampler']). Please pass an `accelerate.DataLoaderConfiguration` instead:
|
| 15 |
+
dataloader_config = DataLoaderConfiguration(dispatch_batches=None, split_batches=False, even_batches=True, use_seedable_sampler=True)
|
| 16 |
+
warnings.warn(
|
| 17 |
+
[34m[1mwandb[39m[22m: [33mWARNING[39m Calling wandb.login() after wandb.init() has no effect.
|
| 18 |
+
/opt/conda/lib/python3.10/site-packages/transformers/models/wav2vec2/processing_wav2vec2.py:156: UserWarning: `as_target_processor` is deprecated and will be removed in v5 of Transformers. You can process your labels by using the argument `text` of the regular `__call__` method (either in the same call as your audio inputs, or in a separate call.
|
| 19 |
+
warnings.warn(
|
| 20 |
+
/opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
|
| 21 |
+
warnings.warn(
|
| 22 |
+
/opt/conda/lib/python3.10/site-packages/accelerate/accelerator.py:436: FutureWarning: Passing the following arguments to `Accelerator` is deprecated and will be removed in version 1.0 of Accelerate: dict_keys(['dispatch_batches', 'split_batches', 'even_batches', 'use_seedable_sampler']). Please pass an `accelerate.DataLoaderConfiguration` instead:
|
| 23 |
+
dataloader_config = DataLoaderConfiguration(dispatch_batches=None, split_batches=False, even_batches=True, use_seedable_sampler=True)
|
| 24 |
+
warnings.warn(
|
| 25 |
+
/opt/conda/lib/python3.10/site-packages/transformers/models/wav2vec2/processing_wav2vec2.py:156: UserWarning: `as_target_processor` is deprecated and will be removed in v5 of Transformers. You can process your labels by using the argument `text` of the regular `__call__` method (either in the same call as your audio inputs, or in a separate call.
|
| 26 |
+
warnings.warn(
|
| 27 |
+
/opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
|
| 28 |
+
warnings.warn(
|
| 29 |
+
0 지금부터 그러면 본격적으로 우리가 도련동�� 대해서 조사를 할 거라 예? 이 마을이 ...
|
| 30 |
+
1 예, 그건 한 칠백년 전에 이제 그 설촌이 시작이 되었다고 헙니다.
|
| 31 |
+
2 예. 칠백년 전에 설촌이 뒈엇는데 이제 그루후에 이제 성씨들이 여러 성씨들이 많이 ...
|
| 32 |
+
3 예. 그러면은 칠백년부터 허는데 설촌할 때 어떤 성씨들이 헷덴 말도 이신가마씨?
|
| 33 |
+
4 그 다음 양씨. 고씨. 마 대략적으로 요런 순서가 뒘서양.
|
| 34 |
+
...
|
| 35 |
+
995 이제 뭘 부린다고 허까. 이제 사람이면 그 살짝살짝 빼여지곡 뭐허곡 헹근에 일을 아...
|
| 36 |
+
996 그런 그 뭐가 잇는데. 에 짐 실런 오단에 그 구루마 찬, 참 그 마차 찬 차 그냥...
|
| 37 |
+
997 것도 딴 거 보민 그 소가 그 좀 소머리도 그런 거 보민 영리허긴 영 리헌 거라. ...
|
| 38 |
+
998 아무 거 그런 거 저런 거 안 보면은 묵묵히 일을 잘허는데, 그런 거 보면은 이제 ...
|
| 39 |
+
999 이제 힘이 좀 약헌 사름안티 가면은 그냥 제멋대로 임잘 막 놀려먹 주.
|
| 40 |
+
Name: text, Length: 1000, dtype: object
|
| 41 |
+
1000
|
| 42 |
+
file \
|
| 43 |
+
0 /kaggle/input/prepared-dataset/jss-dataset/wav...
|
| 44 |
+
1 /kaggle/input/prepared-dataset/jss-dataset/wav...
|
| 45 |
+
2 /kaggle/input/prepared-dataset/jss-dataset/wav...
|
| 46 |
+
3 /kaggle/input/prepared-dataset/jss-dataset/wav...
|
| 47 |
+
4 /kaggle/input/prepared-dataset/jss-dataset/wav...
|
| 48 |
+
... ...
|
| 49 |
+
9995 /kaggle/input/prepared-dataset/jss-dataset/wav...
|
| 50 |
+
9996 /kaggle/input/prepared-dataset/jss-dataset/wav...
|
| 51 |
+
9997 /kaggle/input/prepared-dataset/jss-dataset/wav...
|
| 52 |
+
9998 /kaggle/input/prepared-dataset/jss-dataset/wav...
|
| 53 |
+
9999 /kaggle/input/prepared-dataset/jss-dataset/wav...
|
| 54 |
+
text
|
| 55 |
+
0 지금부터 그러면 본격적으로 우리가 도련동에 대해서 조사를 할 거라 예? 이 마을이 ...
|
| 56 |
+
1 예, 그건 한 칠백년 전에 이제 그 설촌이 시작이 되었다고 헙니다.
|
| 57 |
+
2 예. 칠백년 전에 설촌이 뒈엇는데 이제 그루후에 이제 성씨들이 여러 성씨들이 많이 ...
|
| 58 |
+
3 예. 그러면은 칠백년부터 허는데 설촌할 때 어떤 성씨들이 헷덴 말도 이신가마씨?
|
| 59 |
+
4 그 다음 양씨. 고씨. 마 대략적으로 요런 순서가 뒘서양.
|
| 60 |
+
... ...
|
| 61 |
+
9995 메족메족 날 때 뜨셔가지고 서리 안 내리겟다 허당 자당 보민 뒷날 아첵이 서리 내리...
|
| 62 |
+
9996 서리 내릴 땐 어떵헤야 뒈마씨?
|
| 63 |
+
9997 게난 서리 내릴 때는 흑을 덮어줭은에. 흑을 덮어줘야 돼.
|
| 64 |
+
9998 실시를 많이 허주게. 일찍 놧당은.
|
| 65 |
+
9999 게난 요즘은 비니루가 이시니까. 비니루에도 고망 아이 뚜러지민 워낙 열 받으민 것도...
|
| 66 |
+
[10000 rows x 2 columns]
|
| 67 |
+
Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.
|
| 68 |
+
/opt/conda/lib/python3.10/site-packages/transformers/configuration_utils.py:363: UserWarning: Passing `gradient_checkpointing` to a config initialization is deprecated and will be removed in v5 Transformers. Using `model.gradient_checkpointing_enable()` instead, or if you are using the `Trainer` API, pass `gradient_checkpointing=True` in your `TrainingArguments`.
|
| 69 |
+
warnings.warn(
|
| 70 |
+
Some weights of Wav2Vec2ForCTC were not initialized from the model checkpoint at facebook/wav2vec2-base and are newly initialized: ['lm_head.bias', 'lm_head.weight', 'wav2vec2.encoder.pos_conv_embed.conv.parametrizations.weight.original0', 'wav2vec2.encoder.pos_conv_embed.conv.parametrizations.weight.original1']
|
| 71 |
+
You should probably TRAIN this model on a down-stream task to be able to use it for predictions and inference.
|
| 72 |
+
/opt/conda/lib/python3.10/site-packages/datasets/load.py:756: FutureWarning: The repository for wer contains custom code which must be executed to correctly load the metric. You can inspect the repository content at https://raw.githubusercontent.com/huggingface/datasets/2.18.0/metrics/wer/wer.py
|
| 73 |
+
You can avoid this message in future by passing the argument `trust_remote_code=True`.
|
| 74 |
+
Passing `trust_remote_code=True` will be mandatory to load this metric from the next major release of `datasets`.
|
| 75 |
+
warnings.warn(
|
| 76 |
+
/opt/conda/lib/python3.10/site-packages/accelerate/accelerator.py:436: FutureWarning: Passing the following arguments to `Accelerator` is deprecated and will be removed in version 1.0 of Accelerate: dict_keys(['dispatch_batches', 'split_batches', 'even_batches', 'use_seedable_sampler']). Please pass an `accelerate.DataLoaderConfiguration` instead:
|
| 77 |
+
dataloader_config = DataLoaderConfiguration(dispatch_batches=None, split_batches=False, even_batches=True, use_seedable_sampler=True)
|
| 78 |
+
warnings.warn(
|
| 79 |
+
[34m[1mwandb[39m[22m: [33mWARNING[39m Calling wandb.login() after wandb.init() has no effect.
|
| 80 |
+
/opt/conda/lib/python3.10/site-packages/transformers/models/wav2vec2/processing_wav2vec2.py:156: UserWarning: `as_target_processor` is deprecated and will be removed in v5 of Transformers. You can process your labels by using the argument `text` of the regular `__call__` method (either in the same call as your audio inputs, or in a separate call.
|
| 81 |
+
warnings.warn(
|
| 82 |
+
/opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
|
| 83 |
+
warnings.warn(
|
| 84 |
+
/opt/conda/lib/python3.10/site-packages/accelerate/accelerator.py:436: FutureWarning: Passing the following arguments to `Accelerator` is deprecated and will be removed in version 1.0 of Accelerate: dict_keys(['dispatch_batches', 'split_batches', 'even_batches', 'use_seedable_sampler']). Please pass an `accelerate.DataLoaderConfiguration` instead:
|
| 85 |
+
dataloader_config = DataLoaderConfiguration(dispatch_batches=None, split_batches=False, even_batches=True, use_seedable_sampler=True)
|
| 86 |
+
warnings.warn(
|
| 87 |
+
/opt/conda/lib/python3.10/site-packages/transformers/models/wav2vec2/processing_wav2vec2.py:156: UserWarning: `as_target_processor` is deprecated and will be removed in v5 of Transformers. You can process your labels by using the argument `text` of the regular `__call__` method (either in the same call as your audio inputs, or in a separate call.
|
| 88 |
+
warnings.warn(
|
| 89 |
+
/opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
|
| 90 |
+
warnings.warn(
|
| 91 |
+
file \
|
| 92 |
+
0 /kaggle/input/prepared-dataset/jss-dataset/wav...
|
| 93 |
+
1 /kaggle/input/prepared-dataset/jss-dataset/wav...
|
| 94 |
+
2 /kaggle/input/prepared-dataset/jss-dataset/wav...
|
| 95 |
+
3 /kaggle/input/prepared-dataset/jss-dataset/wav...
|
| 96 |
+
4 /kaggle/input/prepared-dataset/jss-dataset/wav...
|
| 97 |
+
... ...
|
| 98 |
+
9995 /kaggle/input/prepared-dataset/jss-dataset/wav...
|
| 99 |
+
9996 /kaggle/input/prepared-dataset/jss-dataset/wav...
|
| 100 |
+
9997 /kaggle/input/prepared-dataset/jss-dataset/wav...
|
| 101 |
+
9998 /kaggle/input/prepared-dataset/jss-dataset/wav...
|
| 102 |
+
9999 /kaggle/input/prepared-dataset/jss-dataset/wav...
|
| 103 |
+
text
|
| 104 |
+
0 지금부터 그러면 본격적으로 우리가 도련동에 대해서 조사를 할 거라 예? 이 마을이 ...
|
| 105 |
+
1 예, 그건 한 칠백년 전에 이제 그 설촌이 시작이 되었다고 헙니다.
|
| 106 |
+
2 예. 칠백년 전에 설촌이 뒈엇는데 이제 그루후에 이제 성씨들이 여러 성씨들이 많이 ...
|
| 107 |
+
3 예. 그러면은 칠백년부터 허는데 설촌할 때 어떤 성씨들이 헷덴 말도 이신가마씨?
|
| 108 |
+
4 그 다음 양씨. 고씨. 마 대략적으로 요런 순서가 뒘서양.
|
| 109 |
+
... ...
|
| 110 |
+
9995 메족메족 날 때 뜨셔가지고 서리 안 내리겟다 허당 자당 보민 뒷날 아첵이 서리 내리...
|
| 111 |
+
9996 서리 내릴 땐 어떵헤야 뒈마씨?
|
| 112 |
+
9997 게난 서리 내릴 때는 흑을 덮어줭은에. 흑을 덮어줘야 돼.
|
| 113 |
+
9998 실시를 많이 허주게. 일찍 놧당은.
|
| 114 |
+
9999 게난 요즘은 비니루가 이시니까. 비니루에도 고망 아이 뚜러지민 워낙 열 받으민 것도...
|
| 115 |
+
[10000 rows x 2 columns]
|
| 116 |
+
file \
|
| 117 |
+
0 /kaggle/input/prepared-dataset/jss-dataset/wav...
|
| 118 |
+
1 /kaggle/input/prepared-dataset/jss-dataset/wav...
|
| 119 |
+
2 /kaggle/input/prepared-dataset/jss-dataset/wav...
|
| 120 |
+
3 /kaggle/input/prepared-dataset/jss-dataset/wav...
|
| 121 |
+
4 /kaggle/input/prepared-dataset/jss-dataset/wav...
|
| 122 |
+
... ...
|
| 123 |
+
9995 /kaggle/input/prepared-dataset/jss-dataset/wav...
|
| 124 |
+
9996 /kaggle/input/prepared-dataset/jss-dataset/wav...
|
| 125 |
+
9997 /kaggle/input/prepared-dataset/jss-dataset/wav...
|
| 126 |
+
9998 /kaggle/input/prepared-dataset/jss-dataset/wav...
|
| 127 |
+
9999 /kaggle/input/prepared-dataset/jss-dataset/wav...
|
| 128 |
+
text \
|
| 129 |
+
0 지금부터 그러면 본격적으로 우리가 도련동에 대해서 조사를 할 거라 예? 이 마을이 ...
|
| 130 |
+
1 예, 그건 한 칠백년 전에 이제 그 설촌이 시작이 되었다고 헙니다.
|
| 131 |
+
2 예. 칠백년 전에 설촌이 뒈엇는데 ���제 그루후에 이제 성씨들이 여러 성씨들이 많이 ...
|
| 132 |
+
3 예. 그러면은 칠백년부터 허는데 설촌할 때 어떤 성씨들이 헷덴 말도 이신가마씨?
|
| 133 |
+
4 그 다음 양씨. 고씨. 마 대략적으로 요런 순서가 뒘서양.
|
| 134 |
+
... ...
|
| 135 |
+
9995 메족메족 날 때 뜨셔가지고 서리 안 내리겟다 허당 자당 보민 뒷날 아첵이 서리 내리...
|
| 136 |
+
9996 서리 내릴 땐 어떵헤야 뒈마씨?
|
| 137 |
+
9997 게난 서리 내릴 때는 흑을 덮어줭은에. 흑을 덮어줘야 돼.
|
| 138 |
+
9998 실시를 많이 허주게. 일찍 놧당은.
|
| 139 |
+
9999 게난 요즘은 비니루가 이시니까. 비니루에도 고망 아이 뚜러지민 워낙 열 받으민 것도...
|
| 140 |
+
audio
|
| 141 |
+
0 {'array': [8.564606e-05, -4.7375994e-05, 2.878...
|
| 142 |
+
1 {'array': [-0.000804591, -0.00090140395, -0.00...
|
| 143 |
+
2 {'array': [0.0007377937, 0.00092779216, 0.0005...
|
| 144 |
+
3 {'array': [0.00077229546, 0.0008207406, 0.0001...
|
| 145 |
+
4 {'array': [0.00037264611, 0.00073660095, 0.000...
|
| 146 |
+
... ...
|
| 147 |
+
9995 {'array': [-0.00028759346, -0.00042953866, -0....
|
| 148 |
+
9996 {'array': [0.00026829867, 0.0003072208, -0.000...
|
| 149 |
+
9997 {'array': [0.00016293344, 0.00022468295, -5.42...
|
| 150 |
+
9998 {'array': [0.00092347193, 0.0012520454, 0.0012...
|
| 151 |
+
9999 {'array': [0.00054650987, 0.0008459267, 0.0008...
|
| 152 |
+
[10000 rows x 3 columns]
|
| 153 |
+
Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.
|
| 154 |
+
/opt/conda/lib/python3.10/site-packages/transformers/configuration_utils.py:363: UserWarning: Passing `gradient_checkpointing` to a config initialization is deprecated and will be removed in v5 Transformers. Using `model.gradient_checkpointing_enable()` instead, or if you are using the `Trainer` API, pass `gradient_checkpointing=True` in your `TrainingArguments`.
|
| 155 |
+
warnings.warn(
|
| 156 |
+
Some weights of Wav2Vec2ForCTC were not initialized from the model checkpoint at facebook/wav2vec2-base and are newly initialized: ['lm_head.bias', 'lm_head.weight', 'wav2vec2.encoder.pos_conv_embed.conv.parametrizations.weight.original0', 'wav2vec2.encoder.pos_conv_embed.conv.parametrizations.weight.original1']
|
| 157 |
+
You should probably TRAIN this model on a down-stream task to be able to use it for predictions and inference.
|
| 158 |
+
/opt/conda/lib/python3.10/site-packages/datasets/load.py:756: FutureWarning: The repository for wer contains custom code which must be executed to correctly load the metric. You can inspect the repository content at https://raw.githubusercontent.com/huggingface/datasets/2.18.0/metrics/wer/wer.py
|
| 159 |
+
You can avoid this message in future by passing the argument `trust_remote_code=True`.
|
| 160 |
+
Passing `trust_remote_code=True` will be mandatory to load this metric from the next major release of `datasets`.
|
| 161 |
+
warnings.warn(
|
| 162 |
+
/opt/conda/lib/python3.10/site-packages/accelerate/accelerator.py:436: FutureWarning: Passing the following arguments to `Accelerator` is deprecated and will be removed in version 1.0 of Accelerate: dict_keys(['dispatch_batches', 'split_batches', 'even_batches', 'use_seedable_sampler']). Please pass an `accelerate.DataLoaderConfiguration` instead:
|
| 163 |
+
dataloader_config = DataLoaderConfiguration(dispatch_batches=None, split_batches=False, even_batches=True, use_seedable_sampler=True)
|
| 164 |
+
warnings.warn(
|
| 165 |
+
/opt/conda/lib/python3.10/site-packages/transformers/models/wav2vec2/processing_wav2vec2.py:156: UserWarning: `as_target_processor` is deprecated and will be removed in v5 of Transformers. You can process your labels by using the argument `text` of the regular `__call__` method (either in the same call as your audio inputs, or in a separate call.
|
| 166 |
+
warnings.warn(
|
| 167 |
+
/opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
|
| 168 |
+
warnings.warn(
|
| 169 |
+
/opt/conda/lib/python3.10/site-packages/transformers/models/wav2vec2/processing_wav2vec2.py:156: UserWarning: `as_target_processor` is deprecated and will be removed in v5 of Transformers. You can process your labels by using the argument `text` of the regular `__call__` method (either in the same call as your audio inputs, or in a separate call.
|
| 170 |
+
warnings.warn(
|
| 171 |
+
/opt/conda/lib/python3.10/site-packages/torch/utils/checkpoint.py:429: UserWarning: torch.utils.checkpoint: please pass in use_reentrant=True or use_reentrant=False explicitly. The default value of use_reentrant will be updated to be False in the future. To maintain current behavior, pass use_reentrant=True. It is recommended that you use use_reentrant=False. Refer to docs for more details on the differences between the two variants.
|
wandb/run-20240524_202737-n1w0kmmv/files/requirements.txt
ADDED
|
@@ -0,0 +1,862 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
Babel==2.14.0
|
| 2 |
+
Boruta==0.3
|
| 3 |
+
Brotli==1.0.9
|
| 4 |
+
CVXcanon==0.1.2
|
| 5 |
+
Cartopy==0.23.0
|
| 6 |
+
Cython==3.0.8
|
| 7 |
+
Deprecated==1.2.14
|
| 8 |
+
Farama-Notifications==0.0.4
|
| 9 |
+
Flask==3.0.3
|
| 10 |
+
Geohash==1.0
|
| 11 |
+
GitPython==3.1.41
|
| 12 |
+
ImageHash==4.3.1
|
| 13 |
+
Janome==0.5.0
|
| 14 |
+
Jinja2==3.1.2
|
| 15 |
+
LunarCalendar==0.0.9
|
| 16 |
+
Mako==1.3.3
|
| 17 |
+
Markdown==3.5.2
|
| 18 |
+
MarkupSafe==2.1.3
|
| 19 |
+
MarkupSafe==2.1.5
|
| 20 |
+
Pillow==9.5.0
|
| 21 |
+
PuLP==2.8.0
|
| 22 |
+
PyArabic==0.6.15
|
| 23 |
+
PyJWT==2.8.0
|
| 24 |
+
PyMeeus==0.5.12
|
| 25 |
+
PySocks==1.7.1
|
| 26 |
+
PyUpSet==0.1.1.post7
|
| 27 |
+
PyWavelets==1.5.0
|
| 28 |
+
PyYAML==6.0.1
|
| 29 |
+
Pygments==2.17.2
|
| 30 |
+
Pympler==1.0.1
|
| 31 |
+
QtPy==2.4.1
|
| 32 |
+
Rtree==1.2.0
|
| 33 |
+
SQLAlchemy==2.0.25
|
| 34 |
+
SecretStorage==3.3.3
|
| 35 |
+
Send2Trash==1.8.2
|
| 36 |
+
Shapely==1.8.5.post1
|
| 37 |
+
Shimmy==1.3.0
|
| 38 |
+
SimpleITK==2.3.1
|
| 39 |
+
TPOT==0.12.1
|
| 40 |
+
Theano-PyMC==1.1.2
|
| 41 |
+
Theano==1.0.5
|
| 42 |
+
Wand==0.6.13
|
| 43 |
+
Werkzeug==3.0.2
|
| 44 |
+
absl-py==1.4.0
|
| 45 |
+
accelerate==0.29.3
|
| 46 |
+
access==1.1.9
|
| 47 |
+
affine==2.4.0
|
| 48 |
+
aiobotocore==2.12.3
|
| 49 |
+
aiofiles==22.1.0
|
| 50 |
+
aiohttp-cors==0.7.0
|
| 51 |
+
aiohttp==3.9.1
|
| 52 |
+
aioitertools==0.11.0
|
| 53 |
+
aiorwlock==1.3.0
|
| 54 |
+
aiosignal==1.3.1
|
| 55 |
+
aiosqlite==0.19.0
|
| 56 |
+
albumentations==1.4.0
|
| 57 |
+
alembic==1.13.1
|
| 58 |
+
altair==5.3.0
|
| 59 |
+
annotated-types==0.6.0
|
| 60 |
+
annoy==1.17.3
|
| 61 |
+
anyio==4.2.0
|
| 62 |
+
apache-beam==2.46.0
|
| 63 |
+
aplus==0.11.0
|
| 64 |
+
appdirs==1.4.4
|
| 65 |
+
archspec==0.2.3
|
| 66 |
+
argon2-cffi-bindings==21.2.0
|
| 67 |
+
argon2-cffi==23.1.0
|
| 68 |
+
array-record==0.5.0
|
| 69 |
+
arrow==1.3.0
|
| 70 |
+
arviz==0.18.0
|
| 71 |
+
astroid==3.1.0
|
| 72 |
+
astropy-iers-data==0.2024.4.15.2.45.49
|
| 73 |
+
astropy==6.0.1
|
| 74 |
+
asttokens==2.4.1
|
| 75 |
+
astunparse==1.6.3
|
| 76 |
+
async-lru==2.0.4
|
| 77 |
+
async-timeout==4.0.3
|
| 78 |
+
attrs==23.2.0
|
| 79 |
+
audioread==3.0.1
|
| 80 |
+
autopep8==2.0.4
|
| 81 |
+
backoff==2.2.1
|
| 82 |
+
bayesian-optimization==1.4.3
|
| 83 |
+
beatrix_jupyterlab==2023.128.151533
|
| 84 |
+
beautifulsoup4==4.12.2
|
| 85 |
+
blake3==0.2.1
|
| 86 |
+
bleach==6.1.0
|
| 87 |
+
blessed==1.20.0
|
| 88 |
+
blinker==1.7.0
|
| 89 |
+
blis==0.7.10
|
| 90 |
+
blosc2==2.6.2
|
| 91 |
+
bokeh==3.4.1
|
| 92 |
+
boltons==23.1.1
|
| 93 |
+
boto3==1.26.100
|
| 94 |
+
botocore==1.34.69
|
| 95 |
+
bq_helper==0.4.1
|
| 96 |
+
bqplot==0.12.43
|
| 97 |
+
branca==0.7.1
|
| 98 |
+
brewer2mpl==1.4.1
|
| 99 |
+
brotlipy==0.7.0
|
| 100 |
+
cached-property==1.5.2
|
| 101 |
+
cachetools==4.2.4
|
| 102 |
+
cachetools==5.3.2
|
| 103 |
+
catalogue==2.0.10
|
| 104 |
+
catalyst==22.4
|
| 105 |
+
catboost==1.2.3
|
| 106 |
+
category-encoders==2.6.3
|
| 107 |
+
certifi==2024.2.2
|
| 108 |
+
cesium==0.12.1
|
| 109 |
+
cffi==1.16.0
|
| 110 |
+
charset-normalizer==3.3.2
|
| 111 |
+
chex==0.1.86
|
| 112 |
+
cleverhans==4.0.0
|
| 113 |
+
click-plugins==1.1.1
|
| 114 |
+
click==8.1.7
|
| 115 |
+
cligj==0.7.2
|
| 116 |
+
cloud-tpu-client==0.10
|
| 117 |
+
cloud-tpu-profiler==2.4.0
|
| 118 |
+
cloudpathlib==0.16.0
|
| 119 |
+
cloudpickle==2.2.1
|
| 120 |
+
cloudpickle==3.0.0
|
| 121 |
+
cmdstanpy==1.2.2
|
| 122 |
+
colorama==0.4.6
|
| 123 |
+
colorcet==3.1.0
|
| 124 |
+
colorful==0.5.6
|
| 125 |
+
colorlog==6.8.2
|
| 126 |
+
colorlover==0.3.0
|
| 127 |
+
comm==0.2.1
|
| 128 |
+
conda-libmamba-solver==23.7.0
|
| 129 |
+
conda-package-handling==2.2.0
|
| 130 |
+
conda==23.7.4
|
| 131 |
+
conda_package_streaming==0.9.0
|
| 132 |
+
confection==0.1.4
|
| 133 |
+
contextily==1.6.0
|
| 134 |
+
contourpy==1.2.0
|
| 135 |
+
contourpy==1.2.1
|
| 136 |
+
convertdate==2.4.0
|
| 137 |
+
crcmod==1.7
|
| 138 |
+
cryptography==41.0.7
|
| 139 |
+
cuda-python==12.4.0
|
| 140 |
+
cudf==23.8.0
|
| 141 |
+
cufflinks==0.17.3
|
| 142 |
+
cuml==23.8.0
|
| 143 |
+
cupy==13.0.0
|
| 144 |
+
cycler==0.12.1
|
| 145 |
+
cymem==2.0.8
|
| 146 |
+
cytoolz==0.12.3
|
| 147 |
+
daal4py==2024.3.0
|
| 148 |
+
daal==2024.3.0
|
| 149 |
+
dacite==1.8.1
|
| 150 |
+
dask-cuda==23.8.0
|
| 151 |
+
dask-cudf==23.8.0
|
| 152 |
+
dask-expr==1.0.11
|
| 153 |
+
dask==2024.4.1
|
| 154 |
+
dataclasses-json==0.6.4
|
| 155 |
+
dataproc_jupyter_plugin==0.1.66
|
| 156 |
+
datasets==2.18.0
|
| 157 |
+
datashader==0.16.0
|
| 158 |
+
datatile==1.0.3
|
| 159 |
+
db-dtypes==1.2.0
|
| 160 |
+
deap==1.4.1
|
| 161 |
+
debugpy==1.8.0
|
| 162 |
+
decorator==5.1.1
|
| 163 |
+
deepdiff==7.0.1
|
| 164 |
+
defusedxml==0.7.1
|
| 165 |
+
deprecation==2.1.0
|
| 166 |
+
descartes==1.1.0
|
| 167 |
+
dill==0.3.8
|
| 168 |
+
dipy==1.9.0
|
| 169 |
+
distlib==0.3.8
|
| 170 |
+
distributed==2023.7.1
|
| 171 |
+
distro==1.9.0
|
| 172 |
+
dm-tree==0.1.8
|
| 173 |
+
docker-pycreds==0.4.0
|
| 174 |
+
docker==7.0.0
|
| 175 |
+
docopt==0.6.2
|
| 176 |
+
docstring-parser==0.15
|
| 177 |
+
docstring-to-markdown==0.15
|
| 178 |
+
docutils==0.21.1
|
| 179 |
+
earthengine-api==0.1.399
|
| 180 |
+
easydict==1.13
|
| 181 |
+
easyocr==1.7.1
|
| 182 |
+
ecos==2.0.13
|
| 183 |
+
eli5==0.13.0
|
| 184 |
+
emoji==2.11.0
|
| 185 |
+
en-core-web-lg==3.7.1
|
| 186 |
+
en-core-web-sm==3.7.1
|
| 187 |
+
entrypoints==0.4
|
| 188 |
+
ephem==4.1.5
|
| 189 |
+
esda==2.5.1
|
| 190 |
+
essentia==2.1b6.dev1110
|
| 191 |
+
et-xmlfile==1.1.0
|
| 192 |
+
etils==1.6.0
|
| 193 |
+
exceptiongroup==1.2.0
|
| 194 |
+
executing==2.0.1
|
| 195 |
+
explainable-ai-sdk==1.3.3
|
| 196 |
+
fastai==2.7.14
|
| 197 |
+
fastapi==0.108.0
|
| 198 |
+
fastavro==1.9.3
|
| 199 |
+
fastcore==1.5.29
|
| 200 |
+
fastdownload==0.0.7
|
| 201 |
+
fasteners==0.19
|
| 202 |
+
fastjsonschema==2.19.1
|
| 203 |
+
fastprogress==1.0.3
|
| 204 |
+
fastrlock==0.8.2
|
| 205 |
+
fasttext==0.9.2
|
| 206 |
+
feather-format==0.4.1
|
| 207 |
+
featuretools==1.30.0
|
| 208 |
+
filelock==3.13.1
|
| 209 |
+
fiona==1.9.6
|
| 210 |
+
fitter==1.7.0
|
| 211 |
+
flake8==7.0.0
|
| 212 |
+
flashtext==2.7
|
| 213 |
+
flatbuffers==23.5.26
|
| 214 |
+
flax==0.8.2
|
| 215 |
+
folium==0.16.0
|
| 216 |
+
fonttools==4.47.0
|
| 217 |
+
fonttools==4.51.0
|
| 218 |
+
fqdn==1.5.1
|
| 219 |
+
frozendict==2.4.2
|
| 220 |
+
frozenlist==1.4.1
|
| 221 |
+
fsspec==2024.2.0
|
| 222 |
+
fsspec==2024.3.1
|
| 223 |
+
funcy==2.0
|
| 224 |
+
fury==0.10.0
|
| 225 |
+
future==1.0.0
|
| 226 |
+
fuzzywuzzy==0.18.0
|
| 227 |
+
gast==0.5.4
|
| 228 |
+
gatspy==0.3
|
| 229 |
+
gcsfs==2024.2.0
|
| 230 |
+
gensim==4.3.2
|
| 231 |
+
geographiclib==2.0
|
| 232 |
+
geojson==3.1.0
|
| 233 |
+
geopandas==0.14.3
|
| 234 |
+
geoplot==0.5.1
|
| 235 |
+
geopy==2.4.1
|
| 236 |
+
geoviews==1.12.0
|
| 237 |
+
ggplot==0.11.5
|
| 238 |
+
giddy==2.3.5
|
| 239 |
+
gitdb==4.0.11
|
| 240 |
+
google-ai-generativelanguage==0.6.2
|
| 241 |
+
google-api-core==2.11.1
|
| 242 |
+
google-api-core==2.18.0
|
| 243 |
+
google-api-python-client==2.126.0
|
| 244 |
+
google-apitools==0.5.31
|
| 245 |
+
google-auth-httplib2==0.2.0
|
| 246 |
+
google-auth-oauthlib==1.2.0
|
| 247 |
+
google-auth==2.26.1
|
| 248 |
+
google-cloud-aiplatform==0.6.0a1
|
| 249 |
+
google-cloud-artifact-registry==1.10.0
|
| 250 |
+
google-cloud-automl==1.0.1
|
| 251 |
+
google-cloud-bigquery==2.34.4
|
| 252 |
+
google-cloud-bigtable==1.7.3
|
| 253 |
+
google-cloud-core==2.4.1
|
| 254 |
+
google-cloud-datastore==2.19.0
|
| 255 |
+
google-cloud-dlp==3.14.0
|
| 256 |
+
google-cloud-jupyter-config==0.0.5
|
| 257 |
+
google-cloud-language==2.13.3
|
| 258 |
+
google-cloud-monitoring==2.18.0
|
| 259 |
+
google-cloud-pubsub==2.19.0
|
| 260 |
+
google-cloud-pubsublite==1.9.0
|
| 261 |
+
google-cloud-recommendations-ai==0.7.1
|
| 262 |
+
google-cloud-resource-manager==1.11.0
|
| 263 |
+
google-cloud-spanner==3.40.1
|
| 264 |
+
google-cloud-storage==1.44.0
|
| 265 |
+
google-cloud-translate==3.12.1
|
| 266 |
+
google-cloud-videointelligence==2.13.3
|
| 267 |
+
google-cloud-vision==2.8.0
|
| 268 |
+
google-crc32c==1.5.0
|
| 269 |
+
google-generativeai==0.5.1
|
| 270 |
+
google-pasta==0.2.0
|
| 271 |
+
google-resumable-media==2.7.0
|
| 272 |
+
googleapis-common-protos==1.62.0
|
| 273 |
+
gplearn==0.4.2
|
| 274 |
+
gpustat==1.0.0
|
| 275 |
+
gpxpy==1.6.2
|
| 276 |
+
graphviz==0.20.3
|
| 277 |
+
greenlet==3.0.3
|
| 278 |
+
grpc-google-iam-v1==0.12.7
|
| 279 |
+
grpcio-status==1.48.1
|
| 280 |
+
grpcio-status==1.48.2
|
| 281 |
+
grpcio==1.51.1
|
| 282 |
+
grpcio==1.60.0
|
| 283 |
+
gviz-api==1.10.0
|
| 284 |
+
gym-notices==0.0.8
|
| 285 |
+
gym==0.26.2
|
| 286 |
+
gymnasium==0.29.0
|
| 287 |
+
h11==0.14.0
|
| 288 |
+
h2o==3.46.0.1
|
| 289 |
+
h5netcdf==1.3.0
|
| 290 |
+
h5py==3.10.0
|
| 291 |
+
haversine==2.8.1
|
| 292 |
+
hdfs==2.7.3
|
| 293 |
+
hep-ml==0.7.2
|
| 294 |
+
hijri-converter==2.3.1
|
| 295 |
+
hmmlearn==0.3.2
|
| 296 |
+
holidays==0.24
|
| 297 |
+
holoviews==1.18.3
|
| 298 |
+
hpsklearn==0.1.0
|
| 299 |
+
html5lib==1.1
|
| 300 |
+
htmlmin==0.1.12
|
| 301 |
+
httpcore==1.0.5
|
| 302 |
+
httplib2==0.21.0
|
| 303 |
+
httptools==0.6.1
|
| 304 |
+
httpx==0.27.0
|
| 305 |
+
huggingface-hub==0.22.2
|
| 306 |
+
hunspell==0.5.5
|
| 307 |
+
hydra-slayer==0.5.0
|
| 308 |
+
hyperopt==0.2.7
|
| 309 |
+
hypertools==0.8.0
|
| 310 |
+
idna==3.6
|
| 311 |
+
igraph==0.11.4
|
| 312 |
+
imagecodecs==2024.1.1
|
| 313 |
+
imageio==2.33.1
|
| 314 |
+
imbalanced-learn==0.12.2
|
| 315 |
+
imgaug==0.4.0
|
| 316 |
+
importlib-metadata==6.11.0
|
| 317 |
+
importlib-metadata==7.0.1
|
| 318 |
+
importlib-resources==6.1.1
|
| 319 |
+
inequality==1.0.1
|
| 320 |
+
iniconfig==2.0.0
|
| 321 |
+
ipydatawidgets==4.3.5
|
| 322 |
+
ipykernel==6.28.0
|
| 323 |
+
ipyleaflet==0.18.2
|
| 324 |
+
ipympl==0.7.0
|
| 325 |
+
ipython-genutils==0.2.0
|
| 326 |
+
ipython-genutils==0.2.0
|
| 327 |
+
ipython-sql==0.5.0
|
| 328 |
+
ipython==8.20.0
|
| 329 |
+
ipyvolume==0.6.3
|
| 330 |
+
ipyvue==1.11.0
|
| 331 |
+
ipyvuetify==1.9.4
|
| 332 |
+
ipywebrtc==0.6.0
|
| 333 |
+
ipywidgets==7.7.1
|
| 334 |
+
isoduration==20.11.0
|
| 335 |
+
isort==5.13.2
|
| 336 |
+
isoweek==1.3.3
|
| 337 |
+
itsdangerous==2.2.0
|
| 338 |
+
jaraco.classes==3.3.0
|
| 339 |
+
jax-jumpy==1.0.0
|
| 340 |
+
jax==0.4.23
|
| 341 |
+
jaxlib==0.4.23.dev20240116
|
| 342 |
+
jedi==0.19.1
|
| 343 |
+
jeepney==0.8.0
|
| 344 |
+
jieba==0.42.1
|
| 345 |
+
jiwer==3.0.4
|
| 346 |
+
jmespath==1.0.1
|
| 347 |
+
joblib==1.4.0
|
| 348 |
+
json5==0.9.14
|
| 349 |
+
jsonpatch==1.33
|
| 350 |
+
jsonpointer==2.4
|
| 351 |
+
jsonschema-specifications==2023.12.1
|
| 352 |
+
jsonschema==4.20.0
|
| 353 |
+
jupyter-console==6.6.3
|
| 354 |
+
jupyter-events==0.9.0
|
| 355 |
+
jupyter-http-over-ws==0.0.8
|
| 356 |
+
jupyter-lsp==1.5.1
|
| 357 |
+
jupyter-server-mathjax==0.2.6
|
| 358 |
+
jupyter-ydoc==0.2.5
|
| 359 |
+
jupyter_client==7.4.9
|
| 360 |
+
jupyter_client==8.6.0
|
| 361 |
+
jupyter_core==5.7.1
|
| 362 |
+
jupyter_server==2.12.5
|
| 363 |
+
jupyter_server_fileid==0.9.1
|
| 364 |
+
jupyter_server_proxy==4.1.0
|
| 365 |
+
jupyter_server_terminals==0.5.1
|
| 366 |
+
jupyter_server_ydoc==0.8.0
|
| 367 |
+
jupyterlab-lsp==5.1.0
|
| 368 |
+
jupyterlab-widgets==3.0.9
|
| 369 |
+
jupyterlab==4.1.6
|
| 370 |
+
jupyterlab_git==0.44.0
|
| 371 |
+
jupyterlab_pygments==0.3.0
|
| 372 |
+
jupyterlab_server==2.25.2
|
| 373 |
+
jupytext==1.16.0
|
| 374 |
+
kaggle-environments==1.14.3
|
| 375 |
+
kaggle==1.6.12
|
| 376 |
+
kagglehub==0.2.3
|
| 377 |
+
keras-cv==0.8.2
|
| 378 |
+
keras-nlp==0.9.3
|
| 379 |
+
keras-tuner==1.4.6
|
| 380 |
+
keras==3.2.1
|
| 381 |
+
kernels-mixer==0.0.7
|
| 382 |
+
keyring==24.3.0
|
| 383 |
+
keyrings.google-artifactregistry-auth==1.1.2
|
| 384 |
+
kfp-pipeline-spec==0.2.2
|
| 385 |
+
kfp-server-api==2.0.5
|
| 386 |
+
kfp==2.5.0
|
| 387 |
+
kiwisolver==1.4.5
|
| 388 |
+
kmapper==2.0.1
|
| 389 |
+
kmodes==0.12.2
|
| 390 |
+
korean-lunar-calendar==0.3.1
|
| 391 |
+
kornia==0.7.2
|
| 392 |
+
kornia_rs==0.1.3
|
| 393 |
+
kt-legacy==1.0.5
|
| 394 |
+
kubernetes==26.1.0
|
| 395 |
+
langcodes==3.3.0
|
| 396 |
+
langid==1.1.6
|
| 397 |
+
lazy_loader==0.3
|
| 398 |
+
learntools==0.3.4
|
| 399 |
+
leven==1.0.4
|
| 400 |
+
libclang==16.0.6
|
| 401 |
+
libmambapy==1.5.0
|
| 402 |
+
libpysal==4.9.2
|
| 403 |
+
librosa==0.10.1
|
| 404 |
+
lightgbm==4.2.0
|
| 405 |
+
lightning-utilities==0.11.2
|
| 406 |
+
lime==0.2.0.1
|
| 407 |
+
line-profiler==4.1.2
|
| 408 |
+
linkify-it-py==2.0.3
|
| 409 |
+
llvmlite==0.41.1
|
| 410 |
+
llvmlite==0.42.0
|
| 411 |
+
lml==0.1.0
|
| 412 |
+
locket==1.0.0
|
| 413 |
+
loguru==0.7.2
|
| 414 |
+
lxml==5.2.1
|
| 415 |
+
lz4==4.3.3
|
| 416 |
+
mamba==1.5.0
|
| 417 |
+
mapclassify==2.6.1
|
| 418 |
+
markdown-it-py==3.0.0
|
| 419 |
+
marshmallow==3.21.1
|
| 420 |
+
matplotlib-inline==0.1.6
|
| 421 |
+
matplotlib-venn==0.11.10
|
| 422 |
+
matplotlib==3.7.5
|
| 423 |
+
matplotlib==3.8.4
|
| 424 |
+
mccabe==0.7.0
|
| 425 |
+
mdit-py-plugins==0.4.0
|
| 426 |
+
mdurl==0.1.2
|
| 427 |
+
memory-profiler==0.61.0
|
| 428 |
+
menuinst==2.0.1
|
| 429 |
+
mercantile==1.2.1
|
| 430 |
+
mgwr==2.2.1
|
| 431 |
+
missingno==0.5.2
|
| 432 |
+
mistune==0.8.4
|
| 433 |
+
mizani==0.11.1
|
| 434 |
+
ml-dtypes==0.2.0
|
| 435 |
+
mlcrate==0.2.0
|
| 436 |
+
mlens==0.2.3
|
| 437 |
+
mlxtend==0.23.1
|
| 438 |
+
mne==1.6.1
|
| 439 |
+
mnist==0.2.2
|
| 440 |
+
momepy==0.7.0
|
| 441 |
+
more-itertools==10.2.0
|
| 442 |
+
mpld3==0.5.10
|
| 443 |
+
mpmath==1.3.0
|
| 444 |
+
msgpack==1.0.7
|
| 445 |
+
multidict==6.0.4
|
| 446 |
+
multimethod==1.10
|
| 447 |
+
multipledispatch==1.0.0
|
| 448 |
+
multiprocess==0.70.16
|
| 449 |
+
munkres==1.1.4
|
| 450 |
+
murmurhash==1.0.10
|
| 451 |
+
mypy-extensions==1.0.0
|
| 452 |
+
namex==0.0.8
|
| 453 |
+
nb-conda-kernels==2.3.1
|
| 454 |
+
nb_conda==2.2.1
|
| 455 |
+
nbclassic==1.0.0
|
| 456 |
+
nbclient==0.5.13
|
| 457 |
+
nbconvert==6.4.5
|
| 458 |
+
nbdime==3.2.0
|
| 459 |
+
nbformat==5.9.2
|
| 460 |
+
ndindex==1.8
|
| 461 |
+
nest-asyncio==1.5.8
|
| 462 |
+
networkx==3.2.1
|
| 463 |
+
nibabel==5.2.1
|
| 464 |
+
nilearn==0.10.4
|
| 465 |
+
ninja==1.11.1.1
|
| 466 |
+
nltk==3.2.4
|
| 467 |
+
nose==1.3.7
|
| 468 |
+
notebook==6.5.4
|
| 469 |
+
notebook==6.5.6
|
| 470 |
+
notebook_executor==0.2
|
| 471 |
+
notebook_shim==0.2.3
|
| 472 |
+
numba==0.58.1
|
| 473 |
+
numba==0.59.1
|
| 474 |
+
numexpr==2.10.0
|
| 475 |
+
numpy==1.26.4
|
| 476 |
+
nvidia-ml-py==11.495.46
|
| 477 |
+
nvtx==0.2.10
|
| 478 |
+
oauth2client==4.1.3
|
| 479 |
+
oauthlib==3.2.2
|
| 480 |
+
objsize==0.6.1
|
| 481 |
+
odfpy==1.4.1
|
| 482 |
+
olefile==0.47
|
| 483 |
+
onnx==1.16.0
|
| 484 |
+
opencensus-context==0.1.3
|
| 485 |
+
opencensus==0.11.4
|
| 486 |
+
opencv-contrib-python==4.9.0.80
|
| 487 |
+
opencv-python-headless==4.9.0.80
|
| 488 |
+
opencv-python==4.9.0.80
|
| 489 |
+
openpyxl==3.1.2
|
| 490 |
+
openslide-python==1.3.1
|
| 491 |
+
opentelemetry-api==1.22.0
|
| 492 |
+
opentelemetry-exporter-otlp-proto-common==1.22.0
|
| 493 |
+
opentelemetry-exporter-otlp-proto-grpc==1.22.0
|
| 494 |
+
opentelemetry-exporter-otlp-proto-http==1.22.0
|
| 495 |
+
opentelemetry-exporter-otlp==1.22.0
|
| 496 |
+
opentelemetry-proto==1.22.0
|
| 497 |
+
opentelemetry-sdk==1.22.0
|
| 498 |
+
opentelemetry-semantic-conventions==0.43b0
|
| 499 |
+
opt-einsum==3.3.0
|
| 500 |
+
optax==0.2.2
|
| 501 |
+
optree==0.11.0
|
| 502 |
+
optuna==3.6.1
|
| 503 |
+
orbax-checkpoint==0.5.9
|
| 504 |
+
ordered-set==4.1.0
|
| 505 |
+
orjson==3.9.10
|
| 506 |
+
ortools==9.4.1874
|
| 507 |
+
osmnx==1.9.2
|
| 508 |
+
overrides==7.4.0
|
| 509 |
+
packaging==21.3
|
| 510 |
+
pandas-datareader==0.10.0
|
| 511 |
+
pandas-profiling==3.6.6
|
| 512 |
+
pandas-summary==0.2.0
|
| 513 |
+
pandas==2.1.4
|
| 514 |
+
pandas==2.2.2
|
| 515 |
+
pandasql==0.7.3
|
| 516 |
+
pandocfilters==1.5.0
|
| 517 |
+
panel==1.4.1
|
| 518 |
+
papermill==2.5.0
|
| 519 |
+
param==2.1.0
|
| 520 |
+
parso==0.8.3
|
| 521 |
+
partd==1.4.1
|
| 522 |
+
path.py==12.5.0
|
| 523 |
+
path==16.14.0
|
| 524 |
+
pathos==0.3.2
|
| 525 |
+
pathy==0.10.3
|
| 526 |
+
patsy==0.5.6
|
| 527 |
+
pdf2image==1.17.0
|
| 528 |
+
pettingzoo==1.24.0
|
| 529 |
+
pexpect==4.8.0
|
| 530 |
+
pexpect==4.9.0
|
| 531 |
+
phik==0.12.4
|
| 532 |
+
pickleshare==0.7.5
|
| 533 |
+
pillow==10.3.0
|
| 534 |
+
pip==23.3.2
|
| 535 |
+
pkgutil_resolve_name==1.3.10
|
| 536 |
+
platformdirs==4.2.0
|
| 537 |
+
plotly-express==0.4.1
|
| 538 |
+
plotly==5.18.0
|
| 539 |
+
plotnine==0.13.4
|
| 540 |
+
pluggy==1.4.0
|
| 541 |
+
pointpats==2.4.0
|
| 542 |
+
polars==0.20.21
|
| 543 |
+
polyglot==16.7.4
|
| 544 |
+
pooch==1.8.1
|
| 545 |
+
pox==0.3.4
|
| 546 |
+
ppca==0.0.4
|
| 547 |
+
ppft==1.7.6.8
|
| 548 |
+
preprocessing==0.1.13
|
| 549 |
+
preshed==3.0.9
|
| 550 |
+
prettytable==3.9.0
|
| 551 |
+
progressbar2==4.4.2
|
| 552 |
+
prometheus-client==0.19.0
|
| 553 |
+
promise==2.3
|
| 554 |
+
prompt-toolkit==3.0.42
|
| 555 |
+
prompt-toolkit==3.0.43
|
| 556 |
+
prophet==1.1.1
|
| 557 |
+
proto-plus==1.23.0
|
| 558 |
+
protobuf==3.20.3
|
| 559 |
+
protobuf==4.21.12
|
| 560 |
+
psutil==5.9.3
|
| 561 |
+
psutil==5.9.7
|
| 562 |
+
ptyprocess==0.7.0
|
| 563 |
+
pudb==2024.1
|
| 564 |
+
pure-eval==0.2.2
|
| 565 |
+
py-cpuinfo==9.0.0
|
| 566 |
+
py-spy==0.3.14
|
| 567 |
+
py4j==0.10.9.7
|
| 568 |
+
pyLDAvis==3.4.1
|
| 569 |
+
pyOpenSSL==23.3.0
|
| 570 |
+
pyaml==23.12.0
|
| 571 |
+
pyarrow-hotfix==0.6
|
| 572 |
+
pyarrow==15.0.2
|
| 573 |
+
pyasn1-modules==0.3.0
|
| 574 |
+
pyasn1==0.5.1
|
| 575 |
+
pybind11==2.12.0
|
| 576 |
+
pyclipper==1.3.0.post5
|
| 577 |
+
pycodestyle==2.11.1
|
| 578 |
+
pycosat==0.6.6
|
| 579 |
+
pycparser==2.21
|
| 580 |
+
pycryptodome==3.20.0
|
| 581 |
+
pyct==0.5.0
|
| 582 |
+
pycuda==2024.1
|
| 583 |
+
pydantic==2.5.3
|
| 584 |
+
pydantic==2.7.0
|
| 585 |
+
pydantic_core==2.14.6
|
| 586 |
+
pydantic_core==2.18.1
|
| 587 |
+
pydegensac==0.1.2
|
| 588 |
+
pydicom==2.4.4
|
| 589 |
+
pydocstyle==6.3.0
|
| 590 |
+
pydot==1.4.2
|
| 591 |
+
pydub==0.25.1
|
| 592 |
+
pyemd==1.0.0
|
| 593 |
+
pyerfa==2.0.1.4
|
| 594 |
+
pyexcel-io==0.6.6
|
| 595 |
+
pyexcel-ods==0.6.0
|
| 596 |
+
pyflakes==3.2.0
|
| 597 |
+
pygltflib==1.16.2
|
| 598 |
+
pykalman==0.9.7
|
| 599 |
+
pylibraft==23.8.0
|
| 600 |
+
pylint==3.1.0
|
| 601 |
+
pymc3==3.11.4
|
| 602 |
+
pymongo==3.13.0
|
| 603 |
+
pynndescent==0.5.12
|
| 604 |
+
pynvml==11.4.1
|
| 605 |
+
pynvrtc==9.2
|
| 606 |
+
pyparsing==3.1.1
|
| 607 |
+
pyparsing==3.1.2
|
| 608 |
+
pypdf==4.2.0
|
| 609 |
+
pyproj==3.6.1
|
| 610 |
+
pysal==24.1
|
| 611 |
+
pyshp==2.3.1
|
| 612 |
+
pytesseract==0.3.10
|
| 613 |
+
pytest==8.1.1
|
| 614 |
+
python-bidi==0.4.2
|
| 615 |
+
python-dateutil==2.9.0.post0
|
| 616 |
+
python-dotenv==1.0.0
|
| 617 |
+
python-json-logger==2.0.7
|
| 618 |
+
python-louvain==0.16
|
| 619 |
+
python-lsp-jsonrpc==1.1.2
|
| 620 |
+
python-lsp-server==1.11.0
|
| 621 |
+
python-slugify==8.0.4
|
| 622 |
+
python-utils==3.8.2
|
| 623 |
+
pythreejs==2.4.2
|
| 624 |
+
pytoolconfig==1.3.1
|
| 625 |
+
pytools==2024.1.1
|
| 626 |
+
pytorch-ignite==0.5.0.post2
|
| 627 |
+
pytorch-lightning==2.2.2
|
| 628 |
+
pytz==2023.3.post1
|
| 629 |
+
pytz==2024.1
|
| 630 |
+
pyu2f==0.1.5
|
| 631 |
+
pyviz_comms==3.0.2
|
| 632 |
+
pyzmq==24.0.1
|
| 633 |
+
pyzmq==25.1.2
|
| 634 |
+
qgrid==1.3.1
|
| 635 |
+
qtconsole==5.5.1
|
| 636 |
+
quantecon==0.7.2
|
| 637 |
+
qudida==0.0.4
|
| 638 |
+
raft-dask==23.8.0
|
| 639 |
+
rapidfuzz==3.9.1
|
| 640 |
+
rasterio==1.3.10
|
| 641 |
+
rasterstats==0.19.0
|
| 642 |
+
ray-cpp==2.9.0
|
| 643 |
+
ray==2.9.0
|
| 644 |
+
referencing==0.32.1
|
| 645 |
+
regex==2023.12.25
|
| 646 |
+
requests-oauthlib==1.3.1
|
| 647 |
+
requests-toolbelt==0.10.1
|
| 648 |
+
requests==2.31.0
|
| 649 |
+
retrying==1.3.3
|
| 650 |
+
retrying==1.3.4
|
| 651 |
+
rfc3339-validator==0.1.4
|
| 652 |
+
rfc3986-validator==0.1.1
|
| 653 |
+
rgf-python==3.12.0
|
| 654 |
+
rich-click==1.7.4
|
| 655 |
+
rich==13.7.0
|
| 656 |
+
rich==13.7.1
|
| 657 |
+
rmm==23.8.0
|
| 658 |
+
rope==1.13.0
|
| 659 |
+
rpds-py==0.16.2
|
| 660 |
+
rsa==4.9
|
| 661 |
+
ruamel-yaml-conda==0.15.100
|
| 662 |
+
ruamel.yaml.clib==0.2.7
|
| 663 |
+
ruamel.yaml==0.17.40
|
| 664 |
+
s2sphere==0.2.5
|
| 665 |
+
s3fs==2024.2.0
|
| 666 |
+
s3transfer==0.6.2
|
| 667 |
+
safetensors==0.4.3
|
| 668 |
+
scattertext==0.1.19
|
| 669 |
+
scikit-image==0.22.0
|
| 670 |
+
scikit-learn-intelex==2024.3.0
|
| 671 |
+
scikit-learn==1.2.2
|
| 672 |
+
scikit-multilearn==0.2.0
|
| 673 |
+
scikit-optimize==0.10.1
|
| 674 |
+
scikit-plot==0.3.7
|
| 675 |
+
scikit-surprise==1.1.3
|
| 676 |
+
scipy==1.11.4
|
| 677 |
+
scipy==1.13.0
|
| 678 |
+
seaborn==0.12.2
|
| 679 |
+
segment_anything==1.0
|
| 680 |
+
segregation==2.5
|
| 681 |
+
semver==3.0.2
|
| 682 |
+
sentencepiece==0.2.0
|
| 683 |
+
sentry-sdk==1.45.0
|
| 684 |
+
setproctitle==1.3.3
|
| 685 |
+
setuptools-git==1.2
|
| 686 |
+
setuptools-scm==8.0.4
|
| 687 |
+
setuptools==69.0.3
|
| 688 |
+
shap==0.44.1
|
| 689 |
+
shapely==2.0.4
|
| 690 |
+
shellingham==1.5.4
|
| 691 |
+
simpervisor==1.0.0
|
| 692 |
+
simplejson==3.19.2
|
| 693 |
+
six==1.16.0
|
| 694 |
+
sklearn-pandas==2.2.0
|
| 695 |
+
slicer==0.0.7
|
| 696 |
+
smart-open==6.4.0
|
| 697 |
+
smmap==5.0.1
|
| 698 |
+
sniffio==1.3.0
|
| 699 |
+
snowballstemmer==2.2.0
|
| 700 |
+
snuggs==1.4.7
|
| 701 |
+
sortedcontainers==2.4.0
|
| 702 |
+
soundfile==0.12.1
|
| 703 |
+
soupsieve==2.5
|
| 704 |
+
soxr==0.3.7
|
| 705 |
+
spacy-legacy==3.0.12
|
| 706 |
+
spacy-loggers==1.0.5
|
| 707 |
+
spacy==3.7.3
|
| 708 |
+
spaghetti==1.7.5.post1
|
| 709 |
+
spectral==0.23.1
|
| 710 |
+
spglm==1.1.0
|
| 711 |
+
sphinx-rtd-theme==0.2.4
|
| 712 |
+
spint==1.0.7
|
| 713 |
+
splot==1.1.5.post1
|
| 714 |
+
spopt==0.6.0
|
| 715 |
+
spreg==1.4.2
|
| 716 |
+
spvcm==0.3.0
|
| 717 |
+
sqlparse==0.4.4
|
| 718 |
+
squarify==0.4.3
|
| 719 |
+
srsly==2.4.8
|
| 720 |
+
stable-baselines3==2.1.0
|
| 721 |
+
stack-data==0.6.2
|
| 722 |
+
stack-data==0.6.3
|
| 723 |
+
stanio==0.5.0
|
| 724 |
+
starlette==0.32.0.post1
|
| 725 |
+
statsmodels==0.14.1
|
| 726 |
+
stemming==1.0.1
|
| 727 |
+
stop-words==2018.7.23
|
| 728 |
+
stopit==1.1.2
|
| 729 |
+
stumpy==1.12.0
|
| 730 |
+
sympy==1.12
|
| 731 |
+
tables==3.9.2
|
| 732 |
+
tabulate==0.9.0
|
| 733 |
+
tangled-up-in-unicode==0.2.0
|
| 734 |
+
tbb==2021.12.0
|
| 735 |
+
tblib==3.0.0
|
| 736 |
+
tenacity==8.2.3
|
| 737 |
+
tensorboard-data-server==0.7.2
|
| 738 |
+
tensorboard-plugin-profile==2.15.0
|
| 739 |
+
tensorboard==2.15.1
|
| 740 |
+
tensorboardX==2.6.2.2
|
| 741 |
+
tensorflow-cloud==0.1.16
|
| 742 |
+
tensorflow-datasets==4.9.4
|
| 743 |
+
tensorflow-decision-forests==1.8.1
|
| 744 |
+
tensorflow-estimator==2.15.0
|
| 745 |
+
tensorflow-hub==0.16.1
|
| 746 |
+
tensorflow-io-gcs-filesystem==0.35.0
|
| 747 |
+
tensorflow-io==0.35.0
|
| 748 |
+
tensorflow-metadata==0.14.0
|
| 749 |
+
tensorflow-probability==0.23.0
|
| 750 |
+
tensorflow-serving-api==2.14.1
|
| 751 |
+
tensorflow-text==2.15.0
|
| 752 |
+
tensorflow-transform==0.14.0
|
| 753 |
+
tensorflow==2.15.0
|
| 754 |
+
tensorstore==0.1.56
|
| 755 |
+
termcolor==2.4.0
|
| 756 |
+
terminado==0.18.0
|
| 757 |
+
testpath==0.6.0
|
| 758 |
+
text-unidecode==1.3
|
| 759 |
+
textblob==0.18.0.post0
|
| 760 |
+
texttable==1.7.0
|
| 761 |
+
tf_keras==2.15.1
|
| 762 |
+
tfp-nightly==0.24.0.dev0
|
| 763 |
+
thinc==8.2.2
|
| 764 |
+
threadpoolctl==3.2.0
|
| 765 |
+
tifffile==2023.12.9
|
| 766 |
+
timm==0.9.16
|
| 767 |
+
tinycss2==1.2.1
|
| 768 |
+
tobler==0.11.2
|
| 769 |
+
tokenizers==0.15.2
|
| 770 |
+
toml==0.10.2
|
| 771 |
+
tomli==2.0.1
|
| 772 |
+
tomlkit==0.12.4
|
| 773 |
+
toolz==0.12.1
|
| 774 |
+
torch==2.1.2
|
| 775 |
+
torchaudio==2.1.2
|
| 776 |
+
torchdata==0.7.1
|
| 777 |
+
torchinfo==1.8.0
|
| 778 |
+
torchmetrics==1.3.2
|
| 779 |
+
torchtext==0.16.2
|
| 780 |
+
torchvision==0.16.2
|
| 781 |
+
tornado==6.3.3
|
| 782 |
+
tqdm==4.66.1
|
| 783 |
+
traceml==1.0.8
|
| 784 |
+
traitlets==5.9.0
|
| 785 |
+
traittypes==0.2.1
|
| 786 |
+
transformers==4.39.3
|
| 787 |
+
treelite-runtime==3.2.0
|
| 788 |
+
treelite==3.2.0
|
| 789 |
+
truststore==0.8.0
|
| 790 |
+
trx-python==0.2.9
|
| 791 |
+
tsfresh==0.20.2
|
| 792 |
+
typeguard==4.1.5
|
| 793 |
+
typer==0.9.0
|
| 794 |
+
typer==0.9.4
|
| 795 |
+
types-python-dateutil==2.8.19.20240106
|
| 796 |
+
typing-inspect==0.9.0
|
| 797 |
+
typing-utils==0.1.0
|
| 798 |
+
typing_extensions==4.9.0
|
| 799 |
+
tzdata==2023.4
|
| 800 |
+
uc-micro-py==1.0.3
|
| 801 |
+
ucx-py==0.33.0
|
| 802 |
+
ujson==5.9.0
|
| 803 |
+
umap-learn==0.5.6
|
| 804 |
+
unicodedata2==15.1.0
|
| 805 |
+
update-checker==0.18.0
|
| 806 |
+
uri-template==1.3.0
|
| 807 |
+
uritemplate==3.0.1
|
| 808 |
+
urllib3==1.26.18
|
| 809 |
+
urllib3==2.1.0
|
| 810 |
+
urwid==2.6.10
|
| 811 |
+
urwid_readline==0.14
|
| 812 |
+
uvicorn==0.25.0
|
| 813 |
+
uvloop==0.19.0
|
| 814 |
+
vaex-astro==0.9.3
|
| 815 |
+
vaex-core==4.17.1
|
| 816 |
+
vaex-hdf5==0.14.1
|
| 817 |
+
vaex-jupyter==0.8.2
|
| 818 |
+
vaex-ml==0.18.3
|
| 819 |
+
vaex-server==0.9.0
|
| 820 |
+
vaex-viz==0.5.4
|
| 821 |
+
vaex==4.17.0
|
| 822 |
+
vec_noise==1.1.4
|
| 823 |
+
vecstack==0.4.0
|
| 824 |
+
virtualenv==20.21.0
|
| 825 |
+
visions==0.7.5
|
| 826 |
+
vowpalwabbit==9.9.0
|
| 827 |
+
vtk==9.3.0
|
| 828 |
+
wandb==0.16.6
|
| 829 |
+
wasabi==1.1.2
|
| 830 |
+
watchfiles==0.21.0
|
| 831 |
+
wavio==0.0.8
|
| 832 |
+
wcwidth==0.2.13
|
| 833 |
+
weasel==0.3.4
|
| 834 |
+
webcolors==1.13
|
| 835 |
+
webencodings==0.5.1
|
| 836 |
+
websocket-client==1.7.0
|
| 837 |
+
websockets==12.0
|
| 838 |
+
wfdb==4.1.2
|
| 839 |
+
whatthepatch==1.0.5
|
| 840 |
+
wheel==0.42.0
|
| 841 |
+
widgetsnbextension==3.6.6
|
| 842 |
+
witwidget==1.8.1
|
| 843 |
+
woodwork==0.30.0
|
| 844 |
+
wordcloud==1.9.3
|
| 845 |
+
wordsegment==1.3.1
|
| 846 |
+
wrapt==1.14.1
|
| 847 |
+
xarray-einstats==0.7.0
|
| 848 |
+
xarray==2024.3.0
|
| 849 |
+
xgboost==2.0.3
|
| 850 |
+
xvfbwrapper==0.2.9
|
| 851 |
+
xxhash==3.4.1
|
| 852 |
+
xyzservices==2024.4.0
|
| 853 |
+
y-py==0.6.2
|
| 854 |
+
yapf==0.40.2
|
| 855 |
+
yarl==1.9.3
|
| 856 |
+
yarl==1.9.4
|
| 857 |
+
ydata-profiling==4.6.4
|
| 858 |
+
yellowbrick==1.5
|
| 859 |
+
ypy-websocket==0.8.4
|
| 860 |
+
zict==3.0.0
|
| 861 |
+
zipp==3.17.0
|
| 862 |
+
zstandard==0.22.0
|
wandb/run-20240524_202737-n1w0kmmv/files/wandb-metadata.json
ADDED
|
@@ -0,0 +1,62 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"os": "Linux-5.15.133+-x86_64-with-glibc2.31",
|
| 3 |
+
"python": "3.10.13",
|
| 4 |
+
"heartbeatAt": "2024-05-24T20:27:38.540904",
|
| 5 |
+
"startedAt": "2024-05-24T20:27:37.415317",
|
| 6 |
+
"docker": null,
|
| 7 |
+
"cuda": null,
|
| 8 |
+
"args": [],
|
| 9 |
+
"state": "running",
|
| 10 |
+
"program": "kaggle.ipynb",
|
| 11 |
+
"codePathLocal": null,
|
| 12 |
+
"root": "/kaggle/working",
|
| 13 |
+
"host": "890e8b3ca76b",
|
| 14 |
+
"username": "root",
|
| 15 |
+
"executable": "/opt/conda/bin/python3.10",
|
| 16 |
+
"cpu_count": 2,
|
| 17 |
+
"cpu_count_logical": 4,
|
| 18 |
+
"cpu_freq": {
|
| 19 |
+
"current": 2000.17,
|
| 20 |
+
"min": 0.0,
|
| 21 |
+
"max": 0.0
|
| 22 |
+
},
|
| 23 |
+
"cpu_freq_per_core": [
|
| 24 |
+
{
|
| 25 |
+
"current": 2000.17,
|
| 26 |
+
"min": 0.0,
|
| 27 |
+
"max": 0.0
|
| 28 |
+
},
|
| 29 |
+
{
|
| 30 |
+
"current": 2000.17,
|
| 31 |
+
"min": 0.0,
|
| 32 |
+
"max": 0.0
|
| 33 |
+
},
|
| 34 |
+
{
|
| 35 |
+
"current": 2000.17,
|
| 36 |
+
"min": 0.0,
|
| 37 |
+
"max": 0.0
|
| 38 |
+
},
|
| 39 |
+
{
|
| 40 |
+
"current": 2000.17,
|
| 41 |
+
"min": 0.0,
|
| 42 |
+
"max": 0.0
|
| 43 |
+
}
|
| 44 |
+
],
|
| 45 |
+
"disk": {
|
| 46 |
+
"/": {
|
| 47 |
+
"total": 8062.387607574463,
|
| 48 |
+
"used": 5599.177722930908
|
| 49 |
+
}
|
| 50 |
+
},
|
| 51 |
+
"gpu": "Tesla P100-PCIE-16GB",
|
| 52 |
+
"gpu_count": 1,
|
| 53 |
+
"gpu_devices": [
|
| 54 |
+
{
|
| 55 |
+
"name": "Tesla P100-PCIE-16GB",
|
| 56 |
+
"memory_total": 17179869184
|
| 57 |
+
}
|
| 58 |
+
],
|
| 59 |
+
"memory": {
|
| 60 |
+
"total": 31.357559204101562
|
| 61 |
+
}
|
| 62 |
+
}
|
wandb/run-20240524_202737-n1w0kmmv/files/wandb-summary.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"train/loss": 8079.875, "train/grad_norm": NaN, "train/learning_rate": 3.6e-06, "train/epoch": 11.36, "train/global_step": 500, "_timestamp": 1716584600.8230615, "_runtime": 2143.3996374607086, "_step": 1, "eval/loss": NaN, "eval/wer": 0.9984338292873923, "eval/runtime": 10.1243, "eval/samples_per_second": 19.754, "eval/steps_per_second": 2.469}
|
wandb/run-20240524_202737-n1w0kmmv/logs/debug-internal.log
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
wandb/run-20240524_202737-n1w0kmmv/logs/debug.log
ADDED
|
@@ -0,0 +1,217 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
2024-05-24 20:27:37,417 INFO MainThread:34 [wandb_setup.py:_flush():76] Current SDK version is 0.16.6
|
| 2 |
+
2024-05-24 20:27:37,417 INFO MainThread:34 [wandb_setup.py:_flush():76] Configure stats pid to 34
|
| 3 |
+
2024-05-24 20:27:37,417 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings
|
| 4 |
+
2024-05-24 20:27:37,417 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from /kaggle/working/wandb/settings
|
| 5 |
+
2024-05-24 20:27:37,417 INFO MainThread:34 [wandb_setup.py:_flush():76] Loading settings from environment variables: {}
|
| 6 |
+
2024-05-24 20:27:37,417 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False}
|
| 7 |
+
2024-05-24 20:27:37,417 INFO MainThread:34 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program': '<python with no main file>'}
|
| 8 |
+
2024-05-24 20:27:37,417 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying login settings: {}
|
| 9 |
+
2024-05-24 20:27:37,417 INFO MainThread:34 [wandb_setup.py:_flush():76] Applying login settings: {'api_key': '***REDACTED***'}
|
| 10 |
+
2024-05-24 20:27:37,417 INFO MainThread:34 [wandb_init.py:_log_setup():521] Logging user logs to /kaggle/working/wandb/run-20240524_202737-n1w0kmmv/logs/debug.log
|
| 11 |
+
2024-05-24 20:27:37,417 INFO MainThread:34 [wandb_init.py:_log_setup():522] Logging internal logs to /kaggle/working/wandb/run-20240524_202737-n1w0kmmv/logs/debug-internal.log
|
| 12 |
+
2024-05-24 20:27:37,417 INFO MainThread:34 [wandb_init.py:_jupyter_setup():467] configuring jupyter hooks <wandb.sdk.wandb_init._WandbInit object at 0x785a802b9120>
|
| 13 |
+
2024-05-24 20:27:37,418 INFO MainThread:34 [wandb_init.py:init():561] calling init triggers
|
| 14 |
+
2024-05-24 20:27:37,418 INFO MainThread:34 [wandb_init.py:init():568] wandb.init called with sweep_config: {}
|
| 15 |
+
config: {}
|
| 16 |
+
2024-05-24 20:27:37,418 INFO MainThread:34 [wandb_init.py:init():611] starting backend
|
| 17 |
+
2024-05-24 20:27:37,418 INFO MainThread:34 [wandb_init.py:init():615] setting up manager
|
| 18 |
+
2024-05-24 20:27:37,420 INFO MainThread:34 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
|
| 19 |
+
2024-05-24 20:27:37,423 INFO MainThread:34 [wandb_init.py:init():623] backend started and connected
|
| 20 |
+
2024-05-24 20:27:37,434 INFO MainThread:34 [wandb_run.py:_label_probe_notebook():1299] probe notebook
|
| 21 |
+
2024-05-24 20:27:38,191 INFO MainThread:34 [wandb_init.py:init():715] updated telemetry
|
| 22 |
+
2024-05-24 20:27:38,195 INFO MainThread:34 [wandb_init.py:init():748] communicating run to backend with 90.0 second timeout
|
| 23 |
+
2024-05-24 20:27:38,448 INFO MainThread:34 [wandb_run.py:_on_init():2357] communicating current version
|
| 24 |
+
2024-05-24 20:27:38,512 INFO MainThread:34 [wandb_run.py:_on_init():2366] got version response upgrade_message: "wandb version 0.17.0 is available! To upgrade, please run:\n $ pip install wandb --upgrade"
|
| 25 |
+
|
| 26 |
+
2024-05-24 20:27:38,514 INFO MainThread:34 [wandb_init.py:init():799] starting run threads in backend
|
| 27 |
+
2024-05-24 20:27:54,627 INFO MainThread:34 [wandb_run.py:_console_start():2335] atexit reg
|
| 28 |
+
2024-05-24 20:27:54,628 INFO MainThread:34 [wandb_run.py:_redirect():2190] redirect: wrap_raw
|
| 29 |
+
2024-05-24 20:27:54,628 INFO MainThread:34 [wandb_run.py:_redirect():2255] Wrapping output streams.
|
| 30 |
+
2024-05-24 20:27:54,629 INFO MainThread:34 [wandb_run.py:_redirect():2280] Redirects installed.
|
| 31 |
+
2024-05-24 20:27:54,630 INFO MainThread:34 [wandb_init.py:init():842] run started, returning control to user process
|
| 32 |
+
2024-05-24 20:27:54,637 INFO MainThread:34 [wandb_run.py:_config_callback():1347] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['Wav2Vec2ForPreTraining'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'facebook/wav2vec2-base', 'transformers_version': '4.39.3', 'freeze_feat_extract_train': True, 'mask_channel_length': 10, 'mask_channel_min_space': 1, 'mask_channel_other': 0.0, 'mask_channel_prob': 0.0, 'mask_channel_selection': 'static', 'mask_time_min_space': 1, 'mask_time_other': 0.0, 'mask_time_selection': 'static', 'model_type': 'wav2vec2', 'no_mask_channel_overlap': False, 'no_mask_time_overlap': False, 'num_feat_extract_layers': 7, 'hidden_size': 768, 'feat_extract_norm': 'group', 'feat_extract_activation': 'gelu', 'conv_dim': [512, 512, 512, 512, 512, 512, 512], 'conv_stride': [5, 2, 2, 2, 2, 2, 2], 'conv_kernel': [10, 3, 3, 3, 3, 2, 2], 'conv_bias': False, 'num_conv_pos_embeddings': 128, 'num_conv_pos_embedding_groups': 16, 'num_hidden_layers': 12, 'intermediate_size': 3072, 'hidden_act': 'gelu', 'num_attention_heads': 12, 'hidden_dropout': 0.1, 'attention_dropout': 0.1, 'activation_dropout': 0.0, 'feat_proj_dropout': 0.1, 'final_dropout': 0.0, 'layerdrop': 0.0, 'layer_norm_eps': 1e-05, 'initializer_range': 0.02, 'vocab_size': 32, 'do_stable_layer_norm': False, 'use_weighted_layer_sum': False, 'apply_spec_augment': True, 'mask_time_prob': 0.05, 'mask_time_length': 10, 'mask_time_min_masks': 2, 'mask_feature_prob': 0.0, 'mask_feature_length': 10, 'mask_feature_min_masks': 0, 'num_codevectors_per_group': 320, 'num_codevector_groups': 2, 'contrastive_logits_temperature': 0.1, 'feat_quantizer_dropout': 0.0, 'num_negatives': 100, 'codevector_dim': 256, 'proj_codevector_dim': 256, 'diversity_loss_weight': 0.1, 'ctc_loss_reduction': 'sum', 'ctc_zero_infinity': False, 'add_adapter': False, 'adapter_kernel_size': 3, 'adapter_stride': 2, 'num_adapter_layers': 3, 'output_hidden_size': 768, 'adapter_attn_dim': None, 'classifier_proj_size': 256, 'tdnn_dim': [512, 512, 512, 512, 1500], 'tdnn_kernel': [5, 3, 3, 1, 1], 'tdnn_dilation': [1, 2, 3, 1, 1], 'xvector_output_dim': 512, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 32, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 0.0001, 'weight_decay': 0.005, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 30, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 1000, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/May24_20-24-49_890e8b3ca76b', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 2, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': 500, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': True, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': True, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None}
|
| 33 |
+
2024-05-24 20:28:04,489 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 34 |
+
2024-05-24 20:28:04,489 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 35 |
+
2024-05-24 20:28:58,368 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 36 |
+
2024-05-24 20:28:59,625 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 37 |
+
2024-05-24 20:28:59,625 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 38 |
+
2024-05-24 20:29:08,949 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 39 |
+
2024-05-24 20:29:08,951 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 40 |
+
2024-05-24 20:29:08,951 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 41 |
+
2024-05-24 20:29:12,141 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 42 |
+
2024-05-24 20:29:16,856 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 43 |
+
2024-05-24 20:29:16,857 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 44 |
+
2024-05-24 20:29:20,764 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 45 |
+
2024-05-24 20:29:24,478 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 46 |
+
2024-05-24 20:29:24,478 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 47 |
+
2024-05-24 20:29:30,221 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 48 |
+
2024-05-24 20:30:45,072 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 49 |
+
2024-05-24 20:30:45,072 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 50 |
+
2024-05-24 20:30:58,707 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 51 |
+
2024-05-24 20:30:58,709 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 52 |
+
2024-05-24 20:30:58,710 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 53 |
+
2024-05-24 20:31:02,188 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 54 |
+
2024-05-24 20:31:02,613 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 55 |
+
2024-05-24 20:31:02,613 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 56 |
+
2024-05-24 20:31:04,725 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 57 |
+
2024-05-24 20:31:04,727 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 58 |
+
2024-05-24 20:31:04,727 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 59 |
+
2024-05-24 20:31:07,065 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 60 |
+
2024-05-24 20:31:07,298 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 61 |
+
2024-05-24 20:31:07,298 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 62 |
+
2024-05-24 20:31:10,187 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 63 |
+
2024-05-24 20:31:10,193 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 64 |
+
2024-05-24 20:31:10,193 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 65 |
+
2024-05-24 20:31:12,404 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 66 |
+
2024-05-24 20:31:50,143 INFO MainThread:34 [wandb_run.py:_config_callback():1347] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['Wav2Vec2ForPreTraining'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'facebook/wav2vec2-base', 'transformers_version': '4.39.3', 'freeze_feat_extract_train': True, 'mask_channel_length': 10, 'mask_channel_min_space': 1, 'mask_channel_other': 0.0, 'mask_channel_prob': 0.0, 'mask_channel_selection': 'static', 'mask_time_min_space': 1, 'mask_time_other': 0.0, 'mask_time_selection': 'static', 'model_type': 'wav2vec2', 'no_mask_channel_overlap': False, 'no_mask_time_overlap': False, 'num_feat_extract_layers': 7, 'hidden_size': 768, 'feat_extract_norm': 'group', 'feat_extract_activation': 'gelu', 'conv_dim': [512, 512, 512, 512, 512, 512, 512], 'conv_stride': [5, 2, 2, 2, 2, 2, 2], 'conv_kernel': [10, 3, 3, 3, 3, 2, 2], 'conv_bias': False, 'num_conv_pos_embeddings': 128, 'num_conv_pos_embedding_groups': 16, 'num_hidden_layers': 12, 'intermediate_size': 3072, 'hidden_act': 'gelu', 'num_attention_heads': 12, 'hidden_dropout': 0.1, 'attention_dropout': 0.1, 'activation_dropout': 0.0, 'feat_proj_dropout': 0.1, 'final_dropout': 0.0, 'layerdrop': 0.0, 'layer_norm_eps': 1e-05, 'initializer_range': 0.02, 'vocab_size': 32, 'do_stable_layer_norm': False, 'use_weighted_layer_sum': False, 'apply_spec_augment': True, 'mask_time_prob': 0.05, 'mask_time_length': 10, 'mask_time_min_masks': 2, 'mask_feature_prob': 0.0, 'mask_feature_length': 10, 'mask_feature_min_masks': 0, 'num_codevectors_per_group': 320, 'num_codevector_groups': 2, 'contrastive_logits_temperature': 0.1, 'feat_quantizer_dropout': 0.0, 'num_negatives': 100, 'codevector_dim': 256, 'proj_codevector_dim': 256, 'diversity_loss_weight': 0.1, 'ctc_loss_reduction': 'sum', 'ctc_zero_infinity': False, 'add_adapter': False, 'adapter_kernel_size': 3, 'adapter_stride': 2, 'num_adapter_layers': 3, 'output_hidden_size': 768, 'adapter_attn_dim': None, 'classifier_proj_size': 256, 'tdnn_dim': [512, 512, 512, 512, 1500], 'tdnn_kernel': [5, 3, 3, 1, 1], 'tdnn_dilation': [1, 2, 3, 1, 1], 'xvector_output_dim': 512, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 32, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 0.0001, 'weight_decay': 0.005, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 30, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 1000, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/May24_20-31-07_890e8b3ca76b', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 2, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': 500, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': True, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': True, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None}
|
| 67 |
+
2024-05-24 20:31:58,626 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 68 |
+
2024-05-24 20:31:58,626 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 69 |
+
2024-05-24 20:34:06,571 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 70 |
+
2024-05-24 20:34:06,717 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 71 |
+
2024-05-24 20:34:06,717 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 72 |
+
2024-05-24 20:34:10,001 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 73 |
+
2024-05-24 20:34:46,464 INFO MainThread:34 [wandb_run.py:_config_callback():1347] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['Wav2Vec2ForPreTraining'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'facebook/wav2vec2-base', 'transformers_version': '4.39.3', 'freeze_feat_extract_train': True, 'mask_channel_length': 10, 'mask_channel_min_space': 1, 'mask_channel_other': 0.0, 'mask_channel_prob': 0.0, 'mask_channel_selection': 'static', 'mask_time_min_space': 1, 'mask_time_other': 0.0, 'mask_time_selection': 'static', 'model_type': 'wav2vec2', 'no_mask_channel_overlap': False, 'no_mask_time_overlap': False, 'num_feat_extract_layers': 7, 'hidden_size': 768, 'feat_extract_norm': 'group', 'feat_extract_activation': 'gelu', 'conv_dim': [512, 512, 512, 512, 512, 512, 512], 'conv_stride': [5, 2, 2, 2, 2, 2, 2], 'conv_kernel': [10, 3, 3, 3, 3, 2, 2], 'conv_bias': False, 'num_conv_pos_embeddings': 128, 'num_conv_pos_embedding_groups': 16, 'num_hidden_layers': 12, 'intermediate_size': 3072, 'hidden_act': 'gelu', 'num_attention_heads': 12, 'hidden_dropout': 0.1, 'attention_dropout': 0.1, 'activation_dropout': 0.0, 'feat_proj_dropout': 0.1, 'final_dropout': 0.0, 'layerdrop': 0.0, 'layer_norm_eps': 1e-05, 'initializer_range': 0.02, 'vocab_size': 32, 'do_stable_layer_norm': False, 'use_weighted_layer_sum': False, 'apply_spec_augment': True, 'mask_time_prob': 0.05, 'mask_time_length': 10, 'mask_time_min_masks': 2, 'mask_feature_prob': 0.0, 'mask_feature_length': 10, 'mask_feature_min_masks': 0, 'num_codevectors_per_group': 320, 'num_codevector_groups': 2, 'contrastive_logits_temperature': 0.1, 'feat_quantizer_dropout': 0.0, 'num_negatives': 100, 'codevector_dim': 256, 'proj_codevector_dim': 256, 'diversity_loss_weight': 0.1, 'ctc_loss_reduction': 'sum', 'ctc_zero_infinity': False, 'add_adapter': False, 'adapter_kernel_size': 3, 'adapter_stride': 2, 'num_adapter_layers': 3, 'output_hidden_size': 768, 'adapter_attn_dim': None, 'classifier_proj_size': 256, 'tdnn_dim': [512, 512, 512, 512, 1500], 'tdnn_kernel': [5, 3, 3, 1, 1], 'tdnn_dilation': [1, 2, 3, 1, 1], 'xvector_output_dim': 512, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 64, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 0.0001, 'weight_decay': 0.005, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 30, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 1000, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/May24_20-34-06_890e8b3ca76b', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 2, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': 500, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': True, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': True, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None}
|
| 74 |
+
2024-05-24 20:34:55,928 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 75 |
+
2024-05-24 20:34:55,929 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 76 |
+
2024-05-24 20:35:26,131 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 77 |
+
2024-05-24 20:35:26,139 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 78 |
+
2024-05-24 20:35:26,139 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 79 |
+
2024-05-24 20:35:32,729 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 80 |
+
2024-05-24 20:35:32,732 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 81 |
+
2024-05-24 20:35:32,732 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 82 |
+
2024-05-24 20:35:34,729 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 83 |
+
2024-05-24 20:35:34,732 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 84 |
+
2024-05-24 20:35:34,733 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 85 |
+
2024-05-24 20:35:47,483 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 86 |
+
2024-05-24 20:35:47,488 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 87 |
+
2024-05-24 20:35:47,488 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 88 |
+
2024-05-24 20:35:56,249 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 89 |
+
2024-05-24 20:35:56,253 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 90 |
+
2024-05-24 20:35:56,253 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 91 |
+
2024-05-24 20:37:20,366 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 92 |
+
2024-05-24 20:37:20,370 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 93 |
+
2024-05-24 20:37:20,370 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 94 |
+
2024-05-24 20:37:28,481 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 95 |
+
2024-05-24 20:37:28,510 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 96 |
+
2024-05-24 20:37:28,511 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 97 |
+
2024-05-24 20:38:00,946 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 98 |
+
2024-05-24 20:38:00,950 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 99 |
+
2024-05-24 20:38:00,950 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 100 |
+
2024-05-24 20:38:04,524 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 101 |
+
2024-05-24 20:38:04,550 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 102 |
+
2024-05-24 20:38:04,550 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 103 |
+
2024-05-24 20:38:12,369 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 104 |
+
2024-05-24 20:38:12,432 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 105 |
+
2024-05-24 20:38:12,432 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 106 |
+
2024-05-24 20:38:16,938 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 107 |
+
2024-05-24 20:38:16,998 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 108 |
+
2024-05-24 20:38:16,998 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 109 |
+
2024-05-24 20:38:21,248 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 110 |
+
2024-05-24 20:38:21,260 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 111 |
+
2024-05-24 20:38:21,260 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 112 |
+
2024-05-24 20:38:22,930 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 113 |
+
2024-05-24 20:38:22,940 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 114 |
+
2024-05-24 20:38:22,940 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 115 |
+
2024-05-24 20:38:37,851 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 116 |
+
2024-05-24 20:38:55,254 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 117 |
+
2024-05-24 20:38:55,254 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 118 |
+
2024-05-24 20:38:58,788 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 119 |
+
2024-05-24 20:38:58,827 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 120 |
+
2024-05-24 20:38:58,827 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 121 |
+
2024-05-24 20:39:00,762 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 122 |
+
2024-05-24 20:40:16,333 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 123 |
+
2024-05-24 20:40:16,334 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 124 |
+
2024-05-24 20:40:21,472 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 125 |
+
2024-05-24 20:40:22,682 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 126 |
+
2024-05-24 20:40:22,682 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 127 |
+
2024-05-24 20:40:28,979 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 128 |
+
2024-05-24 20:40:33,961 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 129 |
+
2024-05-24 20:40:33,961 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 130 |
+
2024-05-24 20:40:36,994 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 131 |
+
2024-05-24 20:40:36,996 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 132 |
+
2024-05-24 20:40:36,997 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 133 |
+
2024-05-24 20:40:42,203 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 134 |
+
2024-05-24 20:40:42,621 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 135 |
+
2024-05-24 20:40:42,621 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 136 |
+
2024-05-24 20:40:44,473 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 137 |
+
2024-05-24 20:40:44,475 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 138 |
+
2024-05-24 20:40:44,475 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 139 |
+
2024-05-24 20:40:48,431 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 140 |
+
2024-05-24 20:40:48,669 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 141 |
+
2024-05-24 20:40:48,669 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 142 |
+
2024-05-24 20:40:50,725 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 143 |
+
2024-05-24 20:40:50,730 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 144 |
+
2024-05-24 20:40:50,730 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 145 |
+
2024-05-24 20:40:52,537 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 146 |
+
2024-05-24 20:41:30,450 INFO MainThread:34 [wandb_run.py:_config_callback():1347] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['Wav2Vec2ForPreTraining'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'facebook/wav2vec2-base', 'transformers_version': '4.39.3', 'freeze_feat_extract_train': True, 'mask_channel_length': 10, 'mask_channel_min_space': 1, 'mask_channel_other': 0.0, 'mask_channel_prob': 0.0, 'mask_channel_selection': 'static', 'mask_time_min_space': 1, 'mask_time_other': 0.0, 'mask_time_selection': 'static', 'model_type': 'wav2vec2', 'no_mask_channel_overlap': False, 'no_mask_time_overlap': False, 'num_feat_extract_layers': 7, 'hidden_size': 768, 'feat_extract_norm': 'group', 'feat_extract_activation': 'gelu', 'conv_dim': [512, 512, 512, 512, 512, 512, 512], 'conv_stride': [5, 2, 2, 2, 2, 2, 2], 'conv_kernel': [10, 3, 3, 3, 3, 2, 2], 'conv_bias': False, 'num_conv_pos_embeddings': 128, 'num_conv_pos_embedding_groups': 16, 'num_hidden_layers': 12, 'intermediate_size': 3072, 'hidden_act': 'gelu', 'num_attention_heads': 12, 'hidden_dropout': 0.1, 'attention_dropout': 0.1, 'activation_dropout': 0.0, 'feat_proj_dropout': 0.1, 'final_dropout': 0.0, 'layerdrop': 0.0, 'layer_norm_eps': 1e-05, 'initializer_range': 0.02, 'vocab_size': 32, 'do_stable_layer_norm': False, 'use_weighted_layer_sum': False, 'apply_spec_augment': True, 'mask_time_prob': 0.05, 'mask_time_length': 10, 'mask_time_min_masks': 2, 'mask_feature_prob': 0.0, 'mask_feature_length': 10, 'mask_feature_min_masks': 0, 'num_codevectors_per_group': 320, 'num_codevector_groups': 2, 'contrastive_logits_temperature': 0.1, 'feat_quantizer_dropout': 0.0, 'num_negatives': 100, 'codevector_dim': 256, 'proj_codevector_dim': 256, 'diversity_loss_weight': 0.1, 'ctc_loss_reduction': 'sum', 'ctc_zero_infinity': False, 'add_adapter': False, 'adapter_kernel_size': 3, 'adapter_stride': 2, 'num_adapter_layers': 3, 'output_hidden_size': 768, 'adapter_attn_dim': None, 'classifier_proj_size': 256, 'tdnn_dim': [512, 512, 512, 512, 1500], 'tdnn_kernel': [5, 3, 3, 1, 1], 'tdnn_dilation': [1, 2, 3, 1, 1], 'xvector_output_dim': 512, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 64, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 0.0001, 'weight_decay': 0.005, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 30, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 1000, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/May24_20-40-48_890e8b3ca76b', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 2, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': 500, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': True, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': True, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None}
|
| 147 |
+
2024-05-24 20:41:39,043 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 148 |
+
2024-05-24 20:41:39,044 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 149 |
+
2024-05-24 20:42:26,950 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 150 |
+
2024-05-24 20:42:27,128 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 151 |
+
2024-05-24 20:42:27,128 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 152 |
+
2024-05-24 20:42:29,871 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 153 |
+
2024-05-24 20:43:07,276 INFO MainThread:34 [wandb_run.py:_config_callback():1347] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['Wav2Vec2ForPreTraining'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'facebook/wav2vec2-base', 'transformers_version': '4.39.3', 'freeze_feat_extract_train': True, 'mask_channel_length': 10, 'mask_channel_min_space': 1, 'mask_channel_other': 0.0, 'mask_channel_prob': 0.0, 'mask_channel_selection': 'static', 'mask_time_min_space': 1, 'mask_time_other': 0.0, 'mask_time_selection': 'static', 'model_type': 'wav2vec2', 'no_mask_channel_overlap': False, 'no_mask_time_overlap': False, 'num_feat_extract_layers': 7, 'hidden_size': 768, 'feat_extract_norm': 'group', 'feat_extract_activation': 'gelu', 'conv_dim': [512, 512, 512, 512, 512, 512, 512], 'conv_stride': [5, 2, 2, 2, 2, 2, 2], 'conv_kernel': [10, 3, 3, 3, 3, 2, 2], 'conv_bias': False, 'num_conv_pos_embeddings': 128, 'num_conv_pos_embedding_groups': 16, 'num_hidden_layers': 12, 'intermediate_size': 3072, 'hidden_act': 'gelu', 'num_attention_heads': 12, 'hidden_dropout': 0.1, 'attention_dropout': 0.1, 'activation_dropout': 0.0, 'feat_proj_dropout': 0.1, 'final_dropout': 0.0, 'layerdrop': 0.0, 'layer_norm_eps': 1e-05, 'initializer_range': 0.02, 'vocab_size': 32, 'do_stable_layer_norm': False, 'use_weighted_layer_sum': False, 'apply_spec_augment': True, 'mask_time_prob': 0.05, 'mask_time_length': 10, 'mask_time_min_masks': 2, 'mask_feature_prob': 0.0, 'mask_feature_length': 10, 'mask_feature_min_masks': 0, 'num_codevectors_per_group': 320, 'num_codevector_groups': 2, 'contrastive_logits_temperature': 0.1, 'feat_quantizer_dropout': 0.0, 'num_negatives': 100, 'codevector_dim': 256, 'proj_codevector_dim': 256, 'diversity_loss_weight': 0.1, 'ctc_loss_reduction': 'sum', 'ctc_zero_infinity': False, 'add_adapter': False, 'adapter_kernel_size': 3, 'adapter_stride': 2, 'num_adapter_layers': 3, 'output_hidden_size': 768, 'adapter_attn_dim': None, 'classifier_proj_size': 256, 'tdnn_dim': [512, 512, 512, 512, 1500], 'tdnn_kernel': [5, 3, 3, 1, 1], 'tdnn_dilation': [1, 2, 3, 1, 1], 'xvector_output_dim': 512, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 16, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 0.0001, 'weight_decay': 0.005, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 30, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 1000, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/May24_20-42-26_890e8b3ca76b', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 2, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': 500, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': True, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': True, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None}
|
| 154 |
+
2024-05-24 20:43:12,247 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 155 |
+
2024-05-24 20:43:12,247 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 156 |
+
2024-05-24 20:44:08,643 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 157 |
+
2024-05-24 20:44:14,294 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 158 |
+
2024-05-24 20:44:14,295 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 159 |
+
2024-05-24 20:44:39,719 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 160 |
+
2024-05-24 20:44:39,770 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 161 |
+
2024-05-24 20:44:39,770 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 162 |
+
2024-05-24 20:44:43,543 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 163 |
+
2024-05-24 20:46:31,143 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 164 |
+
2024-05-24 20:46:31,143 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 165 |
+
2024-05-24 20:46:49,361 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 166 |
+
2024-05-24 20:46:49,391 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 167 |
+
2024-05-24 20:46:49,391 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 168 |
+
2024-05-24 20:46:52,169 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 169 |
+
2024-05-24 20:46:54,391 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 170 |
+
2024-05-24 20:46:54,391 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 171 |
+
2024-05-24 20:46:58,317 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 172 |
+
2024-05-24 20:46:58,322 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 173 |
+
2024-05-24 20:46:58,322 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 174 |
+
2024-05-24 20:47:06,482 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 175 |
+
2024-05-24 20:47:07,830 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 176 |
+
2024-05-24 20:47:07,830 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 177 |
+
2024-05-24 20:47:10,833 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 178 |
+
2024-05-24 20:47:10,835 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 179 |
+
2024-05-24 20:47:10,839 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 180 |
+
2024-05-24 20:47:14,002 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 181 |
+
2024-05-24 20:48:02,460 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 182 |
+
2024-05-24 20:48:02,460 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 183 |
+
2024-05-24 20:48:05,731 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 184 |
+
2024-05-24 20:48:05,733 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 185 |
+
2024-05-24 20:48:05,734 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 186 |
+
2024-05-24 20:48:08,287 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 187 |
+
2024-05-24 20:48:08,714 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 188 |
+
2024-05-24 20:48:08,714 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 189 |
+
2024-05-24 20:48:10,290 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 190 |
+
2024-05-24 20:48:10,293 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 191 |
+
2024-05-24 20:48:10,294 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 192 |
+
2024-05-24 20:48:13,841 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 193 |
+
2024-05-24 20:48:14,091 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 194 |
+
2024-05-24 20:48:14,092 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 195 |
+
2024-05-24 20:48:16,125 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 196 |
+
2024-05-24 20:48:40,373 INFO MainThread:34 [wandb_run.py:_config_callback():1347] config_cb None None {'return_dict': True, 'output_hidden_states': False, 'output_attentions': False, 'torchscript': False, 'torch_dtype': None, 'use_bfloat16': False, 'tf_legacy_loss': False, 'pruned_heads': {}, 'tie_word_embeddings': True, 'chunk_size_feed_forward': 0, 'is_encoder_decoder': False, 'is_decoder': False, 'cross_attention_hidden_size': None, 'add_cross_attention': False, 'tie_encoder_decoder': False, 'max_length': 20, 'min_length': 0, 'do_sample': False, 'early_stopping': False, 'num_beams': 1, 'num_beam_groups': 1, 'diversity_penalty': 0.0, 'temperature': 1.0, 'top_k': 50, 'top_p': 1.0, 'typical_p': 1.0, 'repetition_penalty': 1.0, 'length_penalty': 1.0, 'no_repeat_ngram_size': 0, 'encoder_no_repeat_ngram_size': 0, 'bad_words_ids': None, 'num_return_sequences': 1, 'output_scores': False, 'return_dict_in_generate': False, 'forced_bos_token_id': None, 'forced_eos_token_id': None, 'remove_invalid_values': False, 'exponential_decay_length_penalty': None, 'suppress_tokens': None, 'begin_suppress_tokens': None, 'architectures': ['Wav2Vec2ForPreTraining'], 'finetuning_task': None, 'id2label': {0: 'LABEL_0', 1: 'LABEL_1'}, 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'tokenizer_class': None, 'prefix': None, 'bos_token_id': 1, 'pad_token_id': 0, 'eos_token_id': 2, 'sep_token_id': None, 'decoder_start_token_id': None, 'task_specific_params': None, 'problem_type': None, '_name_or_path': 'facebook/wav2vec2-base', 'transformers_version': '4.39.3', 'freeze_feat_extract_train': True, 'mask_channel_length': 10, 'mask_channel_min_space': 1, 'mask_channel_other': 0.0, 'mask_channel_prob': 0.0, 'mask_channel_selection': 'static', 'mask_time_min_space': 1, 'mask_time_other': 0.0, 'mask_time_selection': 'static', 'model_type': 'wav2vec2', 'no_mask_channel_overlap': False, 'no_mask_time_overlap': False, 'num_feat_extract_layers': 7, 'hidden_size': 768, 'feat_extract_norm': 'group', 'feat_extract_activation': 'gelu', 'conv_dim': [512, 512, 512, 512, 512, 512, 512], 'conv_stride': [5, 2, 2, 2, 2, 2, 2], 'conv_kernel': [10, 3, 3, 3, 3, 2, 2], 'conv_bias': False, 'num_conv_pos_embeddings': 128, 'num_conv_pos_embedding_groups': 16, 'num_hidden_layers': 12, 'intermediate_size': 3072, 'hidden_act': 'gelu', 'num_attention_heads': 12, 'hidden_dropout': 0.1, 'attention_dropout': 0.1, 'activation_dropout': 0.0, 'feat_proj_dropout': 0.1, 'final_dropout': 0.0, 'layerdrop': 0.0, 'layer_norm_eps': 1e-05, 'initializer_range': 0.02, 'vocab_size': 32, 'do_stable_layer_norm': False, 'use_weighted_layer_sum': False, 'apply_spec_augment': True, 'mask_time_prob': 0.05, 'mask_time_length': 10, 'mask_time_min_masks': 2, 'mask_feature_prob': 0.0, 'mask_feature_length': 10, 'mask_feature_min_masks': 0, 'num_codevectors_per_group': 320, 'num_codevector_groups': 2, 'contrastive_logits_temperature': 0.1, 'feat_quantizer_dropout': 0.0, 'num_negatives': 100, 'codevector_dim': 256, 'proj_codevector_dim': 256, 'diversity_loss_weight': 0.1, 'ctc_loss_reduction': 'sum', 'ctc_zero_infinity': False, 'add_adapter': False, 'adapter_kernel_size': 3, 'adapter_stride': 2, 'num_adapter_layers': 3, 'output_hidden_size': 768, 'adapter_attn_dim': None, 'classifier_proj_size': 256, 'tdnn_dim': [512, 512, 512, 512, 1500], 'tdnn_kernel': [5, 3, 3, 1, 1], 'tdnn_dilation': [1, 2, 3, 1, 1], 'xvector_output_dim': 512, 'output_dir': '/kaggle/working/', 'overwrite_output_dir': False, 'do_train': False, 'do_eval': True, 'do_predict': False, 'evaluation_strategy': 'steps', 'prediction_loss_only': False, 'per_device_train_batch_size': 16, 'per_device_eval_batch_size': 8, 'per_gpu_train_batch_size': None, 'per_gpu_eval_batch_size': None, 'gradient_accumulation_steps': 1, 'eval_accumulation_steps': None, 'eval_delay': 0, 'learning_rate': 0.0001, 'weight_decay': 0.005, 'adam_beta1': 0.9, 'adam_beta2': 0.999, 'adam_epsilon': 1e-08, 'max_grad_norm': 1.0, 'num_train_epochs': 30, 'max_steps': -1, 'lr_scheduler_type': 'linear', 'lr_scheduler_kwargs': {}, 'warmup_ratio': 0.0, 'warmup_steps': 1000, 'log_level': 'passive', 'log_level_replica': 'warning', 'log_on_each_node': True, 'logging_dir': '/kaggle/working/runs/May24_20-48-13_890e8b3ca76b', 'logging_strategy': 'steps', 'logging_first_step': False, 'logging_steps': 500, 'logging_nan_inf_filter': True, 'save_strategy': 'steps', 'save_steps': 500, 'save_total_limit': 2, 'save_safetensors': True, 'save_on_each_node': False, 'save_only_model': False, 'no_cuda': False, 'use_cpu': False, 'use_mps_device': False, 'seed': 42, 'data_seed': None, 'jit_mode_eval': False, 'use_ipex': False, 'bf16': False, 'fp16': True, 'fp16_opt_level': 'O1', 'half_precision_backend': 'auto', 'bf16_full_eval': False, 'fp16_full_eval': False, 'tf32': None, 'local_rank': 0, 'ddp_backend': None, 'tpu_num_cores': None, 'tpu_metrics_debug': False, 'debug': [], 'dataloader_drop_last': False, 'eval_steps': 500, 'dataloader_num_workers': 0, 'dataloader_prefetch_factor': None, 'past_index': -1, 'run_name': '/kaggle/working/', 'disable_tqdm': False, 'remove_unused_columns': True, 'label_names': None, 'load_best_model_at_end': False, 'metric_for_best_model': None, 'greater_is_better': None, 'ignore_data_skip': False, 'fsdp': [], 'fsdp_min_num_params': 0, 'fsdp_config': {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, 'fsdp_transformer_layer_cls_to_wrap': None, 'accelerator_config': {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, 'deepspeed': None, 'label_smoothing_factor': 0.0, 'optim': 'adamw_torch', 'optim_args': None, 'adafactor': False, 'group_by_length': True, 'length_column_name': 'length', 'report_to': ['tensorboard', 'wandb'], 'ddp_find_unused_parameters': None, 'ddp_bucket_cap_mb': None, 'ddp_broadcast_buffers': None, 'dataloader_pin_memory': True, 'dataloader_persistent_workers': False, 'skip_memory_metrics': True, 'use_legacy_prediction_loop': False, 'push_to_hub': True, 'resume_from_checkpoint': None, 'hub_model_id': None, 'hub_strategy': 'every_save', 'hub_token': '<HUB_TOKEN>', 'hub_private_repo': False, 'hub_always_push': False, 'gradient_checkpointing': True, 'gradient_checkpointing_kwargs': None, 'include_inputs_for_metrics': False, 'fp16_backend': 'auto', 'push_to_hub_model_id': None, 'push_to_hub_organization': None, 'push_to_hub_token': '<PUSH_TO_HUB_TOKEN>', 'mp_parameters': '', 'auto_find_batch_size': False, 'full_determinism': False, 'torchdynamo': None, 'ray_scope': 'last', 'ddp_timeout': 1800, 'torch_compile': False, 'torch_compile_backend': None, 'torch_compile_mode': None, 'dispatch_batches': None, 'split_batches': None, 'include_tokens_per_second': False, 'include_num_input_tokens_seen': False, 'neftune_noise_alpha': None, 'optim_target_modules': None}
|
| 197 |
+
2024-05-24 20:48:42,631 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 198 |
+
2024-05-24 20:48:42,632 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 199 |
+
2024-05-24 20:49:00,381 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 200 |
+
2024-05-24 20:49:00,425 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 201 |
+
2024-05-24 20:49:00,425 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 202 |
+
2024-05-24 20:49:03,721 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 203 |
+
2024-05-24 20:49:03,725 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 204 |
+
2024-05-24 20:49:03,726 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 205 |
+
2024-05-24 20:49:08,089 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 206 |
+
2024-05-24 20:49:08,455 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 207 |
+
2024-05-24 20:49:08,456 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 208 |
+
2024-05-24 20:51:52,779 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 209 |
+
2024-05-24 20:51:52,781 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 210 |
+
2024-05-24 20:51:52,781 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 211 |
+
2024-05-24 20:51:56,593 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 212 |
+
2024-05-24 20:52:22,759 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 213 |
+
2024-05-24 20:52:22,759 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 214 |
+
2024-05-24 20:52:49,570 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
| 215 |
+
2024-05-24 20:52:49,572 INFO MainThread:34 [jupyter.py:save_ipynb():373] not saving jupyter notebook
|
| 216 |
+
2024-05-24 20:52:49,572 INFO MainThread:34 [wandb_init.py:_pause_backend():432] pausing backend
|
| 217 |
+
2024-05-24 20:52:52,107 INFO MainThread:34 [wandb_init.py:_resume_backend():437] resuming backend
|
wandb/run-20240524_202737-n1w0kmmv/run-n1w0kmmv.wandb
ADDED
|
Binary file (114 kB). View file
|
|
|