adapters run
Browse filesThis view is limited to 50 files because it contains too many changes. See raw diff
- outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/args.json +3 -0
- outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/cola_bert-base-uncased_train_loss.png +0 -0
- outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/cola_bert-base-uncased_validation_loss.png +0 -0
- outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/logfile.log +265 -0
- outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/adapter_config.json +3 -0
- outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/all_results.json +3 -0
- outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/all_results_val.json +3 -0
- outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/eval_res.json +3 -0
- outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/gpu_stats.json +3 -0
- outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/head_config.json +3 -0
- outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/pytorch_adapter.bin +3 -0
- outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/pytorch_model_head.bin +3 -0
- outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/special_tokens_map.json +3 -0
- outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/tokenizer.json +3 -0
- outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/tokenizer_config.json +3 -0
- outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/val_res.json +3 -0
- outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/vocab.txt +0 -0
- outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/adapter_config.json +3 -0
- outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/all_results.json +3 -0
- outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/all_results_val.json +3 -0
- outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/eval_res.json +3 -0
- outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/gpu_stats.json +3 -0
- outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/head_config.json +3 -0
- outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/pytorch_adapter.bin +3 -0
- outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/pytorch_model_head.bin +3 -0
- outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/special_tokens_map.json +3 -0
- outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/tokenizer.json +3 -0
- outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/tokenizer_config.json +3 -0
- outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/val_res.json +3 -0
- outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/vocab.txt +0 -0
- outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/adapter_config.json +3 -0
- outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/all_results.json +3 -0
- outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/all_results_val.json +3 -0
- outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/eval_res.json +3 -0
- outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/gpu_stats.json +3 -0
- outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/head_config.json +3 -0
- outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/pytorch_adapter.bin +3 -0
- outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/pytorch_model_head.bin +3 -0
- outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/special_tokens_map.json +3 -0
- outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/tokenizer.json +3 -0
- outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/tokenizer_config.json +3 -0
- outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/val_res.json +3 -0
- outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/vocab.txt +0 -0
- outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_3999/adapter_config.json +3 -0
- outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_3999/all_results.json +3 -0
- outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_3999/all_results_val.json +3 -0
- outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_3999/eval_res.json +3 -0
- outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_3999/gpu_stats.json +3 -0
- outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_3999/head_config.json +3 -0
- outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_3999/pytorch_adapter.bin +3 -0
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/args.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1415b67ee0305ee6cd9c8765f1d76df5f634333b0b510dda4678c2ac3a922b3b
|
| 3 |
+
size 977
|
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/cola_bert-base-uncased_train_loss.png
ADDED
|
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/cola_bert-base-uncased_validation_loss.png
ADDED
|
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/logfile.log
ADDED
|
@@ -0,0 +1,265 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
05/30/2024 15:13:21 - INFO - __main__ - Number of labels detected = 2
|
| 2 |
+
05/30/2024 15:13:22 - INFO - adapters.heads.model_mixin - Adding head 'default' with config {'head_type': 'masked_lm', 'vocab_size': 30522, 'embedding_size': 768, 'layers': 2, 'activation_function': 'gelu', 'layer_norm': True, 'bias': True, 'shift_labels': False, 'label2id': None}.
|
| 3 |
+
05/30/2024 15:13:23 - INFO - __main__ - Number of labels detected = 2
|
| 4 |
+
05/30/2024 15:13:23 - INFO - adapters.heads.model_mixin - Adding head 'cola' with config {'head_type': 'classification', 'num_labels': 2, 'layers': 2, 'activation_function': 'tanh', 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'use_pooler': False, 'bias': True, 'dropout_prob': None}.
|
| 5 |
+
05/30/2024 15:13:23 - INFO - adapters.configuration.model_adapters_config - Adding adapter 'cola'.
|
| 6 |
+
05/30/2024 15:13:23 - INFO - __main__ - ================================================================================
|
| 7 |
+
Name Architecture #Param %Param Active Train
|
| 8 |
+
--------------------------------------------------------------------------------
|
| 9 |
+
cola bottleneck 1,789,056 1.634 1 1
|
| 10 |
+
--------------------------------------------------------------------------------
|
| 11 |
+
Full model 109,482,240 100.000 0
|
| 12 |
+
================================================================================
|
| 13 |
+
05/30/2024 15:13:23 - INFO - __main__ - printing model
|
| 14 |
+
05/30/2024 15:13:23 - INFO - __main__ - BertAdapterModel(
|
| 15 |
+
(bert): BertModel(
|
| 16 |
+
(embeddings): BertEmbeddings(
|
| 17 |
+
(word_embeddings): Embedding(30522, 768, padding_idx=0)
|
| 18 |
+
(position_embeddings): Embedding(512, 768)
|
| 19 |
+
(token_type_embeddings): Embedding(2, 768)
|
| 20 |
+
(LayerNorm): LayerNorm((768,), eps=1e-12, elementwise_affine=True)
|
| 21 |
+
(dropout): Dropout(p=0.1, inplace=False)
|
| 22 |
+
)
|
| 23 |
+
(encoder): BertEncoder(
|
| 24 |
+
(layer): ModuleList(
|
| 25 |
+
(0-11): 12 x BertLayer(
|
| 26 |
+
(attention): BertAttention(
|
| 27 |
+
(self): BertSelfAttentionWithAdapters(
|
| 28 |
+
(query): LoRALinearTorch(
|
| 29 |
+
in_features=768, out_features=768, bias=True
|
| 30 |
+
(loras): ModuleDict()
|
| 31 |
+
)
|
| 32 |
+
(key): LoRALinearTorch(
|
| 33 |
+
in_features=768, out_features=768, bias=True
|
| 34 |
+
(loras): ModuleDict()
|
| 35 |
+
)
|
| 36 |
+
(value): LoRALinearTorch(
|
| 37 |
+
in_features=768, out_features=768, bias=True
|
| 38 |
+
(loras): ModuleDict()
|
| 39 |
+
)
|
| 40 |
+
(dropout): Dropout(p=0.1, inplace=False)
|
| 41 |
+
(prefix_tuning): PrefixTuningLayer(
|
| 42 |
+
(prefix_gates): ModuleDict()
|
| 43 |
+
(pool): PrefixTuningPool(
|
| 44 |
+
(prefix_tunings): ModuleDict()
|
| 45 |
+
)
|
| 46 |
+
)
|
| 47 |
+
)
|
| 48 |
+
(output): BertSelfOutputWithAdapters(
|
| 49 |
+
(dense): Linear(in_features=768, out_features=768, bias=True)
|
| 50 |
+
(LayerNorm): LayerNorm((768,), eps=1e-12, elementwise_affine=True)
|
| 51 |
+
(dropout): Dropout(p=0.1, inplace=False)
|
| 52 |
+
(adapters): ModuleDict(
|
| 53 |
+
(cola): Adapter(
|
| 54 |
+
(non_linearity): Activation_Function_Class(
|
| 55 |
+
(f): SiLU()
|
| 56 |
+
)
|
| 57 |
+
(adapter_down): Sequential(
|
| 58 |
+
(0): Linear(in_features=768, out_features=48, bias=True)
|
| 59 |
+
(1): Activation_Function_Class(
|
| 60 |
+
(f): SiLU()
|
| 61 |
+
)
|
| 62 |
+
)
|
| 63 |
+
(adapter_up): Linear(in_features=48, out_features=768, bias=True)
|
| 64 |
+
(dropout): Dropout(p=0.0, inplace=False)
|
| 65 |
+
)
|
| 66 |
+
)
|
| 67 |
+
(adapter_fusion_layer): ModuleDict()
|
| 68 |
+
)
|
| 69 |
+
)
|
| 70 |
+
(intermediate): BertIntermediate(
|
| 71 |
+
(dense): LoRALinearTorch(
|
| 72 |
+
in_features=768, out_features=3072, bias=True
|
| 73 |
+
(loras): ModuleDict()
|
| 74 |
+
)
|
| 75 |
+
(intermediate_act_fn): GELUActivation()
|
| 76 |
+
)
|
| 77 |
+
(output): BertOutputWithAdapters(
|
| 78 |
+
(dense): LoRALinearTorch(
|
| 79 |
+
in_features=3072, out_features=768, bias=True
|
| 80 |
+
(loras): ModuleDict()
|
| 81 |
+
)
|
| 82 |
+
(LayerNorm): LayerNorm((768,), eps=1e-12, elementwise_affine=True)
|
| 83 |
+
(dropout): Dropout(p=0.1, inplace=False)
|
| 84 |
+
(adapters): ModuleDict(
|
| 85 |
+
(cola): Adapter(
|
| 86 |
+
(non_linearity): Activation_Function_Class(
|
| 87 |
+
(f): SiLU()
|
| 88 |
+
)
|
| 89 |
+
(adapter_down): Sequential(
|
| 90 |
+
(0): Linear(in_features=768, out_features=48, bias=True)
|
| 91 |
+
(1): Activation_Function_Class(
|
| 92 |
+
(f): SiLU()
|
| 93 |
+
)
|
| 94 |
+
)
|
| 95 |
+
(adapter_up): Linear(in_features=48, out_features=768, bias=True)
|
| 96 |
+
(dropout): Dropout(p=0.0, inplace=False)
|
| 97 |
+
)
|
| 98 |
+
)
|
| 99 |
+
(adapter_fusion_layer): ModuleDict()
|
| 100 |
+
)
|
| 101 |
+
)
|
| 102 |
+
)
|
| 103 |
+
)
|
| 104 |
+
(pooler): BertPooler(
|
| 105 |
+
(dense): Linear(in_features=768, out_features=768, bias=True)
|
| 106 |
+
(activation): Tanh()
|
| 107 |
+
)
|
| 108 |
+
(invertible_adapters): ModuleDict()
|
| 109 |
+
(shared_parameters): ModuleDict()
|
| 110 |
+
(prefix_tuning): PrefixTuningPool(
|
| 111 |
+
(prefix_tunings): ModuleDict()
|
| 112 |
+
)
|
| 113 |
+
(prompt_tuning): PromptTuningLayer(
|
| 114 |
+
(base_model_embeddings): Embedding(30522, 768, padding_idx=0)
|
| 115 |
+
(prompt_tunings): ModuleDict()
|
| 116 |
+
)
|
| 117 |
+
)
|
| 118 |
+
(heads): ModuleDict(
|
| 119 |
+
(default): BertStyleMaskedLMHead(
|
| 120 |
+
(0): Linear(in_features=768, out_features=768, bias=True)
|
| 121 |
+
(1): Activation_Function_Class(
|
| 122 |
+
(f): GELUActivation()
|
| 123 |
+
)
|
| 124 |
+
(2): LayerNorm((768,), eps=1e-12, elementwise_affine=True)
|
| 125 |
+
(3): Linear(in_features=768, out_features=30522, bias=True)
|
| 126 |
+
)
|
| 127 |
+
(cola): ClassificationHead(
|
| 128 |
+
(0): Dropout(p=0.1, inplace=False)
|
| 129 |
+
(1): Linear(in_features=768, out_features=768, bias=True)
|
| 130 |
+
(2): Activation_Function_Class(
|
| 131 |
+
(f): Tanh()
|
| 132 |
+
)
|
| 133 |
+
(3): Dropout(p=0.1, inplace=False)
|
| 134 |
+
(4): Linear(in_features=768, out_features=2, bias=True)
|
| 135 |
+
)
|
| 136 |
+
)
|
| 137 |
+
)
|
| 138 |
+
05/30/2024 15:13:23 - INFO - __main__ - Sample 3397 of the training set: {'input_ids': [101, 9901, 13303, 2044, 1996, 12383, 1012, 102], 'token_type_ids': [0, 0, 0, 0, 0, 0, 0, 0], 'attention_mask': [1, 1, 1, 1, 1, 1, 1, 1], 'labels': 1}.
|
| 139 |
+
05/30/2024 15:13:23 - INFO - __main__ - Sample 2366 of the training set: {'input_ids': [101, 1996, 3586, 3631, 10647, 1005, 1055, 12277, 1012, 102], 'token_type_ids': [0, 0, 0, 0, 0, 0, 0, 0, 0, 0], 'attention_mask': [1, 1, 1, 1, 1, 1, 1, 1, 1, 1], 'labels': 1}.
|
| 140 |
+
05/30/2024 15:13:23 - INFO - __main__ - Sample 2356 of the training set: {'input_ids': [101, 2057, 19379, 26860, 2094, 4981, 2083, 1996, 4624, 1012, 102], 'token_type_ids': [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], 'attention_mask': [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], 'labels': 0}.
|
| 141 |
+
05/30/2024 15:13:23 - INFO - __main__ - Max training steps before recalculation = 10000
|
| 142 |
+
05/30/2024 15:13:23 - INFO - __main__ - num_update_steps_per_epoch initial = 855
|
| 143 |
+
05/30/2024 15:13:23 - INFO - __main__ - num training epochs initial = 3
|
| 144 |
+
05/30/2024 15:13:23 - INFO - __main__ - Adjusted num_train_epochs based on max_train_steps: 3
|
| 145 |
+
05/30/2024 15:13:24 - INFO - __main__ - num_update_steps_per_epoch before recalculation = 855
|
| 146 |
+
05/30/2024 15:13:24 - INFO - __main__ - num_update_steps_per_epoch after recalculation = 855
|
| 147 |
+
05/30/2024 15:13:24 - INFO - __main__ - num training epochs before recalculation = 12
|
| 148 |
+
05/30/2024 15:13:24 - INFO - __main__ - ***** Running training *****
|
| 149 |
+
05/30/2024 15:13:24 - INFO - __main__ - Num examples = 6840
|
| 150 |
+
05/30/2024 15:13:24 - INFO - __main__ - Num Epochs = 12
|
| 151 |
+
05/30/2024 15:13:24 - INFO - __main__ - Instantaneous batch size per device = 8
|
| 152 |
+
05/30/2024 15:13:24 - INFO - __main__ - Total train batch size (w. parallel, distributed & accumulation) = 8
|
| 153 |
+
05/30/2024 15:13:24 - INFO - __main__ - Gradient Accumulation steps = 1
|
| 154 |
+
05/30/2024 15:13:24 - INFO - __main__ - Total optimization steps = 10000
|
| 155 |
+
05/30/2024 15:13:29 - INFO - __main__ - epoch 0: {'matthews_correlation': 0.010974323267649383}
|
| 156 |
+
05/30/2024 15:13:29 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/adapter_config.json
|
| 157 |
+
05/30/2024 15:13:29 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/pytorch_adapter.bin
|
| 158 |
+
05/30/2024 15:13:29 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/head_config.json
|
| 159 |
+
05/30/2024 15:13:29 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/pytorch_model_head.bin
|
| 160 |
+
05/30/2024 15:13:34 - INFO - __main__ - epoch 0: {'matthews_correlation': -0.01399398063991135}
|
| 161 |
+
05/30/2024 15:13:34 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/adapter_config.json
|
| 162 |
+
05/30/2024 15:13:34 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/pytorch_adapter.bin
|
| 163 |
+
05/30/2024 15:13:34 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/head_config.json
|
| 164 |
+
05/30/2024 15:13:34 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/pytorch_model_head.bin
|
| 165 |
+
05/30/2024 15:14:28 - INFO - __main__ - epoch 1: {'matthews_correlation': 0.4559658705975695}
|
| 166 |
+
05/30/2024 15:14:28 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_999/adapter_config.json
|
| 167 |
+
05/30/2024 15:14:28 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_999/pytorch_adapter.bin
|
| 168 |
+
05/30/2024 15:14:28 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_999/head_config.json
|
| 169 |
+
05/30/2024 15:14:28 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_999/pytorch_model_head.bin
|
| 170 |
+
05/30/2024 15:14:34 - INFO - __main__ - epoch 1: {'matthews_correlation': 0.4744167854195685}
|
| 171 |
+
05/30/2024 15:14:34 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_999/adapter_config.json
|
| 172 |
+
05/30/2024 15:14:34 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_999/pytorch_adapter.bin
|
| 173 |
+
05/30/2024 15:14:34 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_999/head_config.json
|
| 174 |
+
05/30/2024 15:14:34 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_999/pytorch_model_head.bin
|
| 175 |
+
05/30/2024 15:15:27 - INFO - __main__ - epoch 2: {'matthews_correlation': 0.5100838064555473}
|
| 176 |
+
05/30/2024 15:15:27 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/adapter_config.json
|
| 177 |
+
05/30/2024 15:15:27 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/pytorch_adapter.bin
|
| 178 |
+
05/30/2024 15:15:27 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/head_config.json
|
| 179 |
+
05/30/2024 15:15:27 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/pytorch_model_head.bin
|
| 180 |
+
05/30/2024 15:15:33 - INFO - __main__ - epoch 2: {'matthews_correlation': 0.5024499972855541}
|
| 181 |
+
05/30/2024 15:15:33 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/adapter_config.json
|
| 182 |
+
05/30/2024 15:15:33 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/pytorch_adapter.bin
|
| 183 |
+
05/30/2024 15:15:33 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/head_config.json
|
| 184 |
+
05/30/2024 15:15:33 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/pytorch_model_head.bin
|
| 185 |
+
05/30/2024 15:16:27 - INFO - __main__ - epoch 3: {'matthews_correlation': 0.5128707947122957}
|
| 186 |
+
05/30/2024 15:16:27 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/adapter_config.json
|
| 187 |
+
05/30/2024 15:16:27 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/pytorch_adapter.bin
|
| 188 |
+
05/30/2024 15:16:27 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/head_config.json
|
| 189 |
+
05/30/2024 15:16:27 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/pytorch_model_head.bin
|
| 190 |
+
05/30/2024 15:16:33 - INFO - __main__ - epoch 3: {'matthews_correlation': 0.5177241811791387}
|
| 191 |
+
05/30/2024 15:16:33 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/adapter_config.json
|
| 192 |
+
05/30/2024 15:16:33 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/pytorch_adapter.bin
|
| 193 |
+
05/30/2024 15:16:33 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/head_config.json
|
| 194 |
+
05/30/2024 15:16:33 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/pytorch_model_head.bin
|
| 195 |
+
05/30/2024 15:17:26 - INFO - __main__ - epoch 4: {'matthews_correlation': 0.5238347808517775}
|
| 196 |
+
05/30/2024 15:17:26 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_3999/adapter_config.json
|
| 197 |
+
05/30/2024 15:17:26 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_3999/pytorch_adapter.bin
|
| 198 |
+
05/30/2024 15:17:26 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_3999/head_config.json
|
| 199 |
+
05/30/2024 15:17:26 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_3999/pytorch_model_head.bin
|
| 200 |
+
05/30/2024 15:17:31 - INFO - __main__ - epoch 4: {'matthews_correlation': 0.5447619216339334}
|
| 201 |
+
05/30/2024 15:17:31 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_3999/adapter_config.json
|
| 202 |
+
05/30/2024 15:17:31 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_3999/pytorch_adapter.bin
|
| 203 |
+
05/30/2024 15:17:31 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_3999/head_config.json
|
| 204 |
+
05/30/2024 15:17:31 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_3999/pytorch_model_head.bin
|
| 205 |
+
05/30/2024 15:18:25 - INFO - __main__ - epoch 5: {'matthews_correlation': 0.5416905121171213}
|
| 206 |
+
05/30/2024 15:18:25 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_4999/adapter_config.json
|
| 207 |
+
05/30/2024 15:18:25 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_4999/pytorch_adapter.bin
|
| 208 |
+
05/30/2024 15:18:25 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_4999/head_config.json
|
| 209 |
+
05/30/2024 15:18:25 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_4999/pytorch_model_head.bin
|
| 210 |
+
05/30/2024 15:18:31 - INFO - __main__ - epoch 5: {'matthews_correlation': 0.5440440628449357}
|
| 211 |
+
05/30/2024 15:18:31 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_4999/adapter_config.json
|
| 212 |
+
05/30/2024 15:18:31 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_4999/pytorch_adapter.bin
|
| 213 |
+
05/30/2024 15:18:31 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_4999/head_config.json
|
| 214 |
+
05/30/2024 15:18:31 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_4999/pytorch_model_head.bin
|
| 215 |
+
05/30/2024 15:19:25 - INFO - __main__ - epoch 7: {'matthews_correlation': 0.534406644277161}
|
| 216 |
+
05/30/2024 15:19:25 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_5999/adapter_config.json
|
| 217 |
+
05/30/2024 15:19:25 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_5999/pytorch_adapter.bin
|
| 218 |
+
05/30/2024 15:19:25 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_5999/head_config.json
|
| 219 |
+
05/30/2024 15:19:25 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_5999/pytorch_model_head.bin
|
| 220 |
+
05/30/2024 15:19:30 - INFO - __main__ - epoch 7: {'matthews_correlation': 0.5297630395991324}
|
| 221 |
+
05/30/2024 15:19:30 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_5999/adapter_config.json
|
| 222 |
+
05/30/2024 15:19:30 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_5999/pytorch_adapter.bin
|
| 223 |
+
05/30/2024 15:19:30 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_5999/head_config.json
|
| 224 |
+
05/30/2024 15:19:30 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_5999/pytorch_model_head.bin
|
| 225 |
+
05/30/2024 15:20:26 - INFO - __main__ - epoch 8: {'matthews_correlation': 0.5347381322825221}
|
| 226 |
+
05/30/2024 15:20:26 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_6999/adapter_config.json
|
| 227 |
+
05/30/2024 15:20:26 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_6999/pytorch_adapter.bin
|
| 228 |
+
05/30/2024 15:20:26 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_6999/head_config.json
|
| 229 |
+
05/30/2024 15:20:26 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_6999/pytorch_model_head.bin
|
| 230 |
+
05/30/2024 15:20:31 - INFO - __main__ - epoch 8: {'matthews_correlation': 0.5428023274928656}
|
| 231 |
+
05/30/2024 15:20:31 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_6999/adapter_config.json
|
| 232 |
+
05/30/2024 15:20:31 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_6999/pytorch_adapter.bin
|
| 233 |
+
05/30/2024 15:20:31 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_6999/head_config.json
|
| 234 |
+
05/30/2024 15:20:31 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_6999/pytorch_model_head.bin
|
| 235 |
+
05/30/2024 15:21:25 - INFO - __main__ - epoch 9: {'matthews_correlation': 0.5504217860036938}
|
| 236 |
+
05/30/2024 15:21:25 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_7999/adapter_config.json
|
| 237 |
+
05/30/2024 15:21:25 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_7999/pytorch_adapter.bin
|
| 238 |
+
05/30/2024 15:21:25 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_7999/head_config.json
|
| 239 |
+
05/30/2024 15:21:25 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_7999/pytorch_model_head.bin
|
| 240 |
+
05/30/2024 15:21:31 - INFO - __main__ - epoch 9: {'matthews_correlation': 0.5519661848311651}
|
| 241 |
+
05/30/2024 15:21:31 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_7999/adapter_config.json
|
| 242 |
+
05/30/2024 15:21:31 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_7999/pytorch_adapter.bin
|
| 243 |
+
05/30/2024 15:21:31 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_7999/head_config.json
|
| 244 |
+
05/30/2024 15:21:31 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_7999/pytorch_model_head.bin
|
| 245 |
+
05/30/2024 15:22:25 - INFO - __main__ - epoch 10: {'matthews_correlation': 0.5425982560876161}
|
| 246 |
+
05/30/2024 15:22:25 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_8999/adapter_config.json
|
| 247 |
+
05/30/2024 15:22:25 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_8999/pytorch_adapter.bin
|
| 248 |
+
05/30/2024 15:22:25 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_8999/head_config.json
|
| 249 |
+
05/30/2024 15:22:25 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_8999/pytorch_model_head.bin
|
| 250 |
+
05/30/2024 15:22:31 - INFO - __main__ - epoch 10: {'matthews_correlation': 0.5547785513559972}
|
| 251 |
+
05/30/2024 15:22:31 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_8999/adapter_config.json
|
| 252 |
+
05/30/2024 15:22:31 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_8999/pytorch_adapter.bin
|
| 253 |
+
05/30/2024 15:22:31 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_8999/head_config.json
|
| 254 |
+
05/30/2024 15:22:31 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_8999/pytorch_model_head.bin
|
| 255 |
+
05/30/2024 15:23:24 - INFO - __main__ - epoch 11: {'matthews_correlation': 0.5353569722427551}
|
| 256 |
+
05/30/2024 15:23:24 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_9999/adapter_config.json
|
| 257 |
+
05/30/2024 15:23:24 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_9999/pytorch_adapter.bin
|
| 258 |
+
05/30/2024 15:23:24 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_9999/head_config.json
|
| 259 |
+
05/30/2024 15:23:24 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_9999/pytorch_model_head.bin
|
| 260 |
+
05/30/2024 15:23:30 - INFO - __main__ - epoch 11: {'matthews_correlation': 0.5551794331953966}
|
| 261 |
+
05/30/2024 15:23:30 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_9999/adapter_config.json
|
| 262 |
+
05/30/2024 15:23:30 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_9999/pytorch_adapter.bin
|
| 263 |
+
05/30/2024 15:23:30 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_9999/head_config.json
|
| 264 |
+
05/30/2024 15:23:30 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_9999/pytorch_model_head.bin
|
| 265 |
+
05/30/2024 15:23:30 - INFO - __main__ - ***** Completed training *****
|
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/adapter_config.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5b2b147db5f916c6102e23f45ccc56602c8b2fec86d67932f37e22fe0b53180a
|
| 3 |
+
size 1044
|
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/all_results.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5310690ff28a2d6faafb09de8013388409720c1cef1cf768928ecac1172dae97
|
| 3 |
+
size 51
|
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/all_results_val.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0d74e0365b79dfa4aae0cff25add05de8bfa765e6c356e7ae68f99d5afefdf26
|
| 3 |
+
size 51
|
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/eval_res.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:bdd1b9e850a862dec87ebd3b858693a28bc01cdb6e5273aa433b90c068367022
|
| 3 |
+
size 185687
|
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/gpu_stats.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:71e3cc001200f49353669c3d09ae4e8f1fe5c6041f3dbc31456008005a743795
|
| 3 |
+
size 5979
|
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/head_config.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:acf384050d7546f70d6b5043d8cfcb2e13077f0438c8576ad3f2e95e87653dd1
|
| 3 |
+
size 421
|
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/pytorch_adapter.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0f2f32f2d4c165ce1039aebf59a44c138ecef5770af5c74e8e16fe08ffa837ed
|
| 3 |
+
size 7191062
|
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/pytorch_model_head.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a12292f4f4f0b4d65ddcc111ea18c3fa0ffaec2d22b344aa7de95427d0071d2f
|
| 3 |
+
size 2370664
|
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/special_tokens_map.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b6d346be366a7d1d48332dbc9fdf3bf8960b5d879522b7799ddba59e76237ee3
|
| 3 |
+
size 125
|
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/tokenizer.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:fa06adf10b0288c8416f63a3dc60480b8f3dd8820212397db6713cdc97f6ae43
|
| 3 |
+
size 711494
|
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/tokenizer_config.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3df22571b6fb8bcd6c00bf341bc0a9ad1a462b4eb1725bfefb72718999cff4a7
|
| 3 |
+
size 1216
|
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/val_res.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2cb915d497b4cd63cd625c1955b702cc429d88c51da269473d9c98e772858d6b
|
| 3 |
+
size 304769
|
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/vocab.txt
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/adapter_config.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5b2b147db5f916c6102e23f45ccc56602c8b2fec86d67932f37e22fe0b53180a
|
| 3 |
+
size 1044
|
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/all_results.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e7e4aa22e380a680fd4a19f99c0eb816c1aae5d95049578a40b736bc9e748b8a
|
| 3 |
+
size 49
|
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/all_results_val.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:50e80fa5806cfd3a5c2440ca8558750a34e661597623b719be1b741136167beb
|
| 3 |
+
size 49
|
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/eval_res.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e90d086e60c29edc32c774e1b06ecaeac4910894fff7d6bd85e1ffa76d33f27d
|
| 3 |
+
size 177106
|
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/gpu_stats.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ad359771e24846dbe1f43693962b772dac302f8f67f83bd0a9fd441339183f63
|
| 3 |
+
size 6055
|
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/head_config.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:acf384050d7546f70d6b5043d8cfcb2e13077f0438c8576ad3f2e95e87653dd1
|
| 3 |
+
size 421
|
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/pytorch_adapter.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a6a6a2da75d2d52918eed99023ade648224b53f1b052a7fefd287c1db703ef5e
|
| 3 |
+
size 7191062
|
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/pytorch_model_head.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9e8dbe58754bab6f464f6cabd87914cd61041d269abc3591ec70581d5bf02772
|
| 3 |
+
size 2370664
|
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/special_tokens_map.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b6d346be366a7d1d48332dbc9fdf3bf8960b5d879522b7799ddba59e76237ee3
|
| 3 |
+
size 125
|
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/tokenizer.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:fa06adf10b0288c8416f63a3dc60480b8f3dd8820212397db6713cdc97f6ae43
|
| 3 |
+
size 711494
|
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/tokenizer_config.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3df22571b6fb8bcd6c00bf341bc0a9ad1a462b4eb1725bfefb72718999cff4a7
|
| 3 |
+
size 1216
|
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/val_res.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3cac5ffd32f24311bfef2f760ad46b17cd7f4842f31285f70bfc8080e6efc285
|
| 3 |
+
size 291394
|
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/vocab.txt
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/adapter_config.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5b2b147db5f916c6102e23f45ccc56602c8b2fec86d67932f37e22fe0b53180a
|
| 3 |
+
size 1044
|
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/all_results.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8e97c5454580e0a577391ff62b4887884b38c2e4f0be2f10ffbb10f47128597c
|
| 3 |
+
size 49
|
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/all_results_val.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:127f036bddad652ea12e1419cb60723823195c930de8ebd75b54e2e45c52fc25
|
| 3 |
+
size 49
|
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/eval_res.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1af60c9ac2339b2336ec7fd4d917ee90426a34d1374b723ca380a57b2383e916
|
| 3 |
+
size 176551
|
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/gpu_stats.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a1c69db5093cae57ad1c3a12ddf5bc2459ddf560e6e25ee7bc7f2cd13f2448a7
|
| 3 |
+
size 6055
|
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/head_config.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:acf384050d7546f70d6b5043d8cfcb2e13077f0438c8576ad3f2e95e87653dd1
|
| 3 |
+
size 421
|
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/pytorch_adapter.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d7b7e22b16b8366ec49cd8340f63646d1e93afb58e8960e5c2dc3be6f02750f9
|
| 3 |
+
size 7191062
|
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/pytorch_model_head.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:92591c99260985732342ae6067dde82d0ffdebd8292a30519c9287041a119f6a
|
| 3 |
+
size 2370664
|
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/special_tokens_map.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b6d346be366a7d1d48332dbc9fdf3bf8960b5d879522b7799ddba59e76237ee3
|
| 3 |
+
size 125
|
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/tokenizer.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:fa06adf10b0288c8416f63a3dc60480b8f3dd8820212397db6713cdc97f6ae43
|
| 3 |
+
size 711494
|
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/tokenizer_config.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3df22571b6fb8bcd6c00bf341bc0a9ad1a462b4eb1725bfefb72718999cff4a7
|
| 3 |
+
size 1216
|
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/val_res.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:fa61010286d51bbc58c1daaa66385814510185276f3ea0eb6a17722faa002ab1
|
| 3 |
+
size 290478
|
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/vocab.txt
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_3999/adapter_config.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5b2b147db5f916c6102e23f45ccc56602c8b2fec86d67932f37e22fe0b53180a
|
| 3 |
+
size 1044
|
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_3999/all_results.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:12595e64dabbe54a88115349f4e3a6060d2960b905e840e2f4f7adc226c3dacf
|
| 3 |
+
size 49
|
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_3999/all_results_val.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:23a2f4d13089808af2fcce749642a6f8d4197d326479ffcd21832ed6b601858e
|
| 3 |
+
size 49
|
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_3999/eval_res.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a089981167598afb50094161b4fd2c9a87b26c758a9956bfe9d7ce5132996684
|
| 3 |
+
size 176154
|
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_3999/gpu_stats.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3227c81d715011800f02db7872d5df521454a5532c7371374497eca8e980033b
|
| 3 |
+
size 6055
|
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_3999/head_config.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:acf384050d7546f70d6b5043d8cfcb2e13077f0438c8576ad3f2e95e87653dd1
|
| 3 |
+
size 421
|
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_3999/pytorch_adapter.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f1882982079bf45a509155a1e297549f958523f630066bc22d94e6517ccc0642
|
| 3 |
+
size 7191062
|