jdorairaj commited on
Commit
7e61843
·
1 Parent(s): 2cb9512

adapters run

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/args.json +3 -0
  2. outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/cola_bert-base-uncased_train_loss.png +0 -0
  3. outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/cola_bert-base-uncased_validation_loss.png +0 -0
  4. outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/logfile.log +265 -0
  5. outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/adapter_config.json +3 -0
  6. outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/all_results.json +3 -0
  7. outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/all_results_val.json +3 -0
  8. outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/eval_res.json +3 -0
  9. outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/gpu_stats.json +3 -0
  10. outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/head_config.json +3 -0
  11. outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/pytorch_adapter.bin +3 -0
  12. outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/pytorch_model_head.bin +3 -0
  13. outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/special_tokens_map.json +3 -0
  14. outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/tokenizer.json +3 -0
  15. outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/tokenizer_config.json +3 -0
  16. outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/val_res.json +3 -0
  17. outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/vocab.txt +0 -0
  18. outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/adapter_config.json +3 -0
  19. outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/all_results.json +3 -0
  20. outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/all_results_val.json +3 -0
  21. outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/eval_res.json +3 -0
  22. outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/gpu_stats.json +3 -0
  23. outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/head_config.json +3 -0
  24. outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/pytorch_adapter.bin +3 -0
  25. outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/pytorch_model_head.bin +3 -0
  26. outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/special_tokens_map.json +3 -0
  27. outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/tokenizer.json +3 -0
  28. outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/tokenizer_config.json +3 -0
  29. outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/val_res.json +3 -0
  30. outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/vocab.txt +0 -0
  31. outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/adapter_config.json +3 -0
  32. outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/all_results.json +3 -0
  33. outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/all_results_val.json +3 -0
  34. outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/eval_res.json +3 -0
  35. outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/gpu_stats.json +3 -0
  36. outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/head_config.json +3 -0
  37. outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/pytorch_adapter.bin +3 -0
  38. outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/pytorch_model_head.bin +3 -0
  39. outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/special_tokens_map.json +3 -0
  40. outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/tokenizer.json +3 -0
  41. outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/tokenizer_config.json +3 -0
  42. outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/val_res.json +3 -0
  43. outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/vocab.txt +0 -0
  44. outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_3999/adapter_config.json +3 -0
  45. outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_3999/all_results.json +3 -0
  46. outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_3999/all_results_val.json +3 -0
  47. outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_3999/eval_res.json +3 -0
  48. outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_3999/gpu_stats.json +3 -0
  49. outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_3999/head_config.json +3 -0
  50. outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_3999/pytorch_adapter.bin +3 -0
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/args.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1415b67ee0305ee6cd9c8765f1d76df5f634333b0b510dda4678c2ac3a922b3b
3
+ size 977
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/cola_bert-base-uncased_train_loss.png ADDED
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/cola_bert-base-uncased_validation_loss.png ADDED
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/logfile.log ADDED
@@ -0,0 +1,265 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 05/30/2024 15:13:21 - INFO - __main__ - Number of labels detected = 2
2
+ 05/30/2024 15:13:22 - INFO - adapters.heads.model_mixin - Adding head 'default' with config {'head_type': 'masked_lm', 'vocab_size': 30522, 'embedding_size': 768, 'layers': 2, 'activation_function': 'gelu', 'layer_norm': True, 'bias': True, 'shift_labels': False, 'label2id': None}.
3
+ 05/30/2024 15:13:23 - INFO - __main__ - Number of labels detected = 2
4
+ 05/30/2024 15:13:23 - INFO - adapters.heads.model_mixin - Adding head 'cola' with config {'head_type': 'classification', 'num_labels': 2, 'layers': 2, 'activation_function': 'tanh', 'label2id': {'LABEL_0': 0, 'LABEL_1': 1}, 'use_pooler': False, 'bias': True, 'dropout_prob': None}.
5
+ 05/30/2024 15:13:23 - INFO - adapters.configuration.model_adapters_config - Adding adapter 'cola'.
6
+ 05/30/2024 15:13:23 - INFO - __main__ - ================================================================================
7
+ Name Architecture #Param %Param Active Train
8
+ --------------------------------------------------------------------------------
9
+ cola bottleneck 1,789,056 1.634 1 1
10
+ --------------------------------------------------------------------------------
11
+ Full model 109,482,240 100.000 0
12
+ ================================================================================
13
+ 05/30/2024 15:13:23 - INFO - __main__ - printing model
14
+ 05/30/2024 15:13:23 - INFO - __main__ - BertAdapterModel(
15
+ (bert): BertModel(
16
+ (embeddings): BertEmbeddings(
17
+ (word_embeddings): Embedding(30522, 768, padding_idx=0)
18
+ (position_embeddings): Embedding(512, 768)
19
+ (token_type_embeddings): Embedding(2, 768)
20
+ (LayerNorm): LayerNorm((768,), eps=1e-12, elementwise_affine=True)
21
+ (dropout): Dropout(p=0.1, inplace=False)
22
+ )
23
+ (encoder): BertEncoder(
24
+ (layer): ModuleList(
25
+ (0-11): 12 x BertLayer(
26
+ (attention): BertAttention(
27
+ (self): BertSelfAttentionWithAdapters(
28
+ (query): LoRALinearTorch(
29
+ in_features=768, out_features=768, bias=True
30
+ (loras): ModuleDict()
31
+ )
32
+ (key): LoRALinearTorch(
33
+ in_features=768, out_features=768, bias=True
34
+ (loras): ModuleDict()
35
+ )
36
+ (value): LoRALinearTorch(
37
+ in_features=768, out_features=768, bias=True
38
+ (loras): ModuleDict()
39
+ )
40
+ (dropout): Dropout(p=0.1, inplace=False)
41
+ (prefix_tuning): PrefixTuningLayer(
42
+ (prefix_gates): ModuleDict()
43
+ (pool): PrefixTuningPool(
44
+ (prefix_tunings): ModuleDict()
45
+ )
46
+ )
47
+ )
48
+ (output): BertSelfOutputWithAdapters(
49
+ (dense): Linear(in_features=768, out_features=768, bias=True)
50
+ (LayerNorm): LayerNorm((768,), eps=1e-12, elementwise_affine=True)
51
+ (dropout): Dropout(p=0.1, inplace=False)
52
+ (adapters): ModuleDict(
53
+ (cola): Adapter(
54
+ (non_linearity): Activation_Function_Class(
55
+ (f): SiLU()
56
+ )
57
+ (adapter_down): Sequential(
58
+ (0): Linear(in_features=768, out_features=48, bias=True)
59
+ (1): Activation_Function_Class(
60
+ (f): SiLU()
61
+ )
62
+ )
63
+ (adapter_up): Linear(in_features=48, out_features=768, bias=True)
64
+ (dropout): Dropout(p=0.0, inplace=False)
65
+ )
66
+ )
67
+ (adapter_fusion_layer): ModuleDict()
68
+ )
69
+ )
70
+ (intermediate): BertIntermediate(
71
+ (dense): LoRALinearTorch(
72
+ in_features=768, out_features=3072, bias=True
73
+ (loras): ModuleDict()
74
+ )
75
+ (intermediate_act_fn): GELUActivation()
76
+ )
77
+ (output): BertOutputWithAdapters(
78
+ (dense): LoRALinearTorch(
79
+ in_features=3072, out_features=768, bias=True
80
+ (loras): ModuleDict()
81
+ )
82
+ (LayerNorm): LayerNorm((768,), eps=1e-12, elementwise_affine=True)
83
+ (dropout): Dropout(p=0.1, inplace=False)
84
+ (adapters): ModuleDict(
85
+ (cola): Adapter(
86
+ (non_linearity): Activation_Function_Class(
87
+ (f): SiLU()
88
+ )
89
+ (adapter_down): Sequential(
90
+ (0): Linear(in_features=768, out_features=48, bias=True)
91
+ (1): Activation_Function_Class(
92
+ (f): SiLU()
93
+ )
94
+ )
95
+ (adapter_up): Linear(in_features=48, out_features=768, bias=True)
96
+ (dropout): Dropout(p=0.0, inplace=False)
97
+ )
98
+ )
99
+ (adapter_fusion_layer): ModuleDict()
100
+ )
101
+ )
102
+ )
103
+ )
104
+ (pooler): BertPooler(
105
+ (dense): Linear(in_features=768, out_features=768, bias=True)
106
+ (activation): Tanh()
107
+ )
108
+ (invertible_adapters): ModuleDict()
109
+ (shared_parameters): ModuleDict()
110
+ (prefix_tuning): PrefixTuningPool(
111
+ (prefix_tunings): ModuleDict()
112
+ )
113
+ (prompt_tuning): PromptTuningLayer(
114
+ (base_model_embeddings): Embedding(30522, 768, padding_idx=0)
115
+ (prompt_tunings): ModuleDict()
116
+ )
117
+ )
118
+ (heads): ModuleDict(
119
+ (default): BertStyleMaskedLMHead(
120
+ (0): Linear(in_features=768, out_features=768, bias=True)
121
+ (1): Activation_Function_Class(
122
+ (f): GELUActivation()
123
+ )
124
+ (2): LayerNorm((768,), eps=1e-12, elementwise_affine=True)
125
+ (3): Linear(in_features=768, out_features=30522, bias=True)
126
+ )
127
+ (cola): ClassificationHead(
128
+ (0): Dropout(p=0.1, inplace=False)
129
+ (1): Linear(in_features=768, out_features=768, bias=True)
130
+ (2): Activation_Function_Class(
131
+ (f): Tanh()
132
+ )
133
+ (3): Dropout(p=0.1, inplace=False)
134
+ (4): Linear(in_features=768, out_features=2, bias=True)
135
+ )
136
+ )
137
+ )
138
+ 05/30/2024 15:13:23 - INFO - __main__ - Sample 3397 of the training set: {'input_ids': [101, 9901, 13303, 2044, 1996, 12383, 1012, 102], 'token_type_ids': [0, 0, 0, 0, 0, 0, 0, 0], 'attention_mask': [1, 1, 1, 1, 1, 1, 1, 1], 'labels': 1}.
139
+ 05/30/2024 15:13:23 - INFO - __main__ - Sample 2366 of the training set: {'input_ids': [101, 1996, 3586, 3631, 10647, 1005, 1055, 12277, 1012, 102], 'token_type_ids': [0, 0, 0, 0, 0, 0, 0, 0, 0, 0], 'attention_mask': [1, 1, 1, 1, 1, 1, 1, 1, 1, 1], 'labels': 1}.
140
+ 05/30/2024 15:13:23 - INFO - __main__ - Sample 2356 of the training set: {'input_ids': [101, 2057, 19379, 26860, 2094, 4981, 2083, 1996, 4624, 1012, 102], 'token_type_ids': [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], 'attention_mask': [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], 'labels': 0}.
141
+ 05/30/2024 15:13:23 - INFO - __main__ - Max training steps before recalculation = 10000
142
+ 05/30/2024 15:13:23 - INFO - __main__ - num_update_steps_per_epoch initial = 855
143
+ 05/30/2024 15:13:23 - INFO - __main__ - num training epochs initial = 3
144
+ 05/30/2024 15:13:23 - INFO - __main__ - Adjusted num_train_epochs based on max_train_steps: 3
145
+ 05/30/2024 15:13:24 - INFO - __main__ - num_update_steps_per_epoch before recalculation = 855
146
+ 05/30/2024 15:13:24 - INFO - __main__ - num_update_steps_per_epoch after recalculation = 855
147
+ 05/30/2024 15:13:24 - INFO - __main__ - num training epochs before recalculation = 12
148
+ 05/30/2024 15:13:24 - INFO - __main__ - ***** Running training *****
149
+ 05/30/2024 15:13:24 - INFO - __main__ - Num examples = 6840
150
+ 05/30/2024 15:13:24 - INFO - __main__ - Num Epochs = 12
151
+ 05/30/2024 15:13:24 - INFO - __main__ - Instantaneous batch size per device = 8
152
+ 05/30/2024 15:13:24 - INFO - __main__ - Total train batch size (w. parallel, distributed & accumulation) = 8
153
+ 05/30/2024 15:13:24 - INFO - __main__ - Gradient Accumulation steps = 1
154
+ 05/30/2024 15:13:24 - INFO - __main__ - Total optimization steps = 10000
155
+ 05/30/2024 15:13:29 - INFO - __main__ - epoch 0: {'matthews_correlation': 0.010974323267649383}
156
+ 05/30/2024 15:13:29 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/adapter_config.json
157
+ 05/30/2024 15:13:29 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/pytorch_adapter.bin
158
+ 05/30/2024 15:13:29 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/head_config.json
159
+ 05/30/2024 15:13:29 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/pytorch_model_head.bin
160
+ 05/30/2024 15:13:34 - INFO - __main__ - epoch 0: {'matthews_correlation': -0.01399398063991135}
161
+ 05/30/2024 15:13:34 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/adapter_config.json
162
+ 05/30/2024 15:13:34 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/pytorch_adapter.bin
163
+ 05/30/2024 15:13:34 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/head_config.json
164
+ 05/30/2024 15:13:34 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/pytorch_model_head.bin
165
+ 05/30/2024 15:14:28 - INFO - __main__ - epoch 1: {'matthews_correlation': 0.4559658705975695}
166
+ 05/30/2024 15:14:28 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_999/adapter_config.json
167
+ 05/30/2024 15:14:28 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_999/pytorch_adapter.bin
168
+ 05/30/2024 15:14:28 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_999/head_config.json
169
+ 05/30/2024 15:14:28 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_999/pytorch_model_head.bin
170
+ 05/30/2024 15:14:34 - INFO - __main__ - epoch 1: {'matthews_correlation': 0.4744167854195685}
171
+ 05/30/2024 15:14:34 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_999/adapter_config.json
172
+ 05/30/2024 15:14:34 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_999/pytorch_adapter.bin
173
+ 05/30/2024 15:14:34 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_999/head_config.json
174
+ 05/30/2024 15:14:34 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_999/pytorch_model_head.bin
175
+ 05/30/2024 15:15:27 - INFO - __main__ - epoch 2: {'matthews_correlation': 0.5100838064555473}
176
+ 05/30/2024 15:15:27 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/adapter_config.json
177
+ 05/30/2024 15:15:27 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/pytorch_adapter.bin
178
+ 05/30/2024 15:15:27 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/head_config.json
179
+ 05/30/2024 15:15:27 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/pytorch_model_head.bin
180
+ 05/30/2024 15:15:33 - INFO - __main__ - epoch 2: {'matthews_correlation': 0.5024499972855541}
181
+ 05/30/2024 15:15:33 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/adapter_config.json
182
+ 05/30/2024 15:15:33 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/pytorch_adapter.bin
183
+ 05/30/2024 15:15:33 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/head_config.json
184
+ 05/30/2024 15:15:33 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/pytorch_model_head.bin
185
+ 05/30/2024 15:16:27 - INFO - __main__ - epoch 3: {'matthews_correlation': 0.5128707947122957}
186
+ 05/30/2024 15:16:27 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/adapter_config.json
187
+ 05/30/2024 15:16:27 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/pytorch_adapter.bin
188
+ 05/30/2024 15:16:27 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/head_config.json
189
+ 05/30/2024 15:16:27 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/pytorch_model_head.bin
190
+ 05/30/2024 15:16:33 - INFO - __main__ - epoch 3: {'matthews_correlation': 0.5177241811791387}
191
+ 05/30/2024 15:16:33 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/adapter_config.json
192
+ 05/30/2024 15:16:33 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/pytorch_adapter.bin
193
+ 05/30/2024 15:16:33 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/head_config.json
194
+ 05/30/2024 15:16:33 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/pytorch_model_head.bin
195
+ 05/30/2024 15:17:26 - INFO - __main__ - epoch 4: {'matthews_correlation': 0.5238347808517775}
196
+ 05/30/2024 15:17:26 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_3999/adapter_config.json
197
+ 05/30/2024 15:17:26 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_3999/pytorch_adapter.bin
198
+ 05/30/2024 15:17:26 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_3999/head_config.json
199
+ 05/30/2024 15:17:26 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_3999/pytorch_model_head.bin
200
+ 05/30/2024 15:17:31 - INFO - __main__ - epoch 4: {'matthews_correlation': 0.5447619216339334}
201
+ 05/30/2024 15:17:31 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_3999/adapter_config.json
202
+ 05/30/2024 15:17:31 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_3999/pytorch_adapter.bin
203
+ 05/30/2024 15:17:31 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_3999/head_config.json
204
+ 05/30/2024 15:17:31 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_3999/pytorch_model_head.bin
205
+ 05/30/2024 15:18:25 - INFO - __main__ - epoch 5: {'matthews_correlation': 0.5416905121171213}
206
+ 05/30/2024 15:18:25 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_4999/adapter_config.json
207
+ 05/30/2024 15:18:25 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_4999/pytorch_adapter.bin
208
+ 05/30/2024 15:18:25 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_4999/head_config.json
209
+ 05/30/2024 15:18:25 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_4999/pytorch_model_head.bin
210
+ 05/30/2024 15:18:31 - INFO - __main__ - epoch 5: {'matthews_correlation': 0.5440440628449357}
211
+ 05/30/2024 15:18:31 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_4999/adapter_config.json
212
+ 05/30/2024 15:18:31 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_4999/pytorch_adapter.bin
213
+ 05/30/2024 15:18:31 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_4999/head_config.json
214
+ 05/30/2024 15:18:31 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_4999/pytorch_model_head.bin
215
+ 05/30/2024 15:19:25 - INFO - __main__ - epoch 7: {'matthews_correlation': 0.534406644277161}
216
+ 05/30/2024 15:19:25 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_5999/adapter_config.json
217
+ 05/30/2024 15:19:25 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_5999/pytorch_adapter.bin
218
+ 05/30/2024 15:19:25 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_5999/head_config.json
219
+ 05/30/2024 15:19:25 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_5999/pytorch_model_head.bin
220
+ 05/30/2024 15:19:30 - INFO - __main__ - epoch 7: {'matthews_correlation': 0.5297630395991324}
221
+ 05/30/2024 15:19:30 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_5999/adapter_config.json
222
+ 05/30/2024 15:19:30 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_5999/pytorch_adapter.bin
223
+ 05/30/2024 15:19:30 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_5999/head_config.json
224
+ 05/30/2024 15:19:30 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_5999/pytorch_model_head.bin
225
+ 05/30/2024 15:20:26 - INFO - __main__ - epoch 8: {'matthews_correlation': 0.5347381322825221}
226
+ 05/30/2024 15:20:26 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_6999/adapter_config.json
227
+ 05/30/2024 15:20:26 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_6999/pytorch_adapter.bin
228
+ 05/30/2024 15:20:26 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_6999/head_config.json
229
+ 05/30/2024 15:20:26 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_6999/pytorch_model_head.bin
230
+ 05/30/2024 15:20:31 - INFO - __main__ - epoch 8: {'matthews_correlation': 0.5428023274928656}
231
+ 05/30/2024 15:20:31 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_6999/adapter_config.json
232
+ 05/30/2024 15:20:31 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_6999/pytorch_adapter.bin
233
+ 05/30/2024 15:20:31 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_6999/head_config.json
234
+ 05/30/2024 15:20:31 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_6999/pytorch_model_head.bin
235
+ 05/30/2024 15:21:25 - INFO - __main__ - epoch 9: {'matthews_correlation': 0.5504217860036938}
236
+ 05/30/2024 15:21:25 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_7999/adapter_config.json
237
+ 05/30/2024 15:21:25 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_7999/pytorch_adapter.bin
238
+ 05/30/2024 15:21:25 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_7999/head_config.json
239
+ 05/30/2024 15:21:25 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_7999/pytorch_model_head.bin
240
+ 05/30/2024 15:21:31 - INFO - __main__ - epoch 9: {'matthews_correlation': 0.5519661848311651}
241
+ 05/30/2024 15:21:31 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_7999/adapter_config.json
242
+ 05/30/2024 15:21:31 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_7999/pytorch_adapter.bin
243
+ 05/30/2024 15:21:31 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_7999/head_config.json
244
+ 05/30/2024 15:21:31 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_7999/pytorch_model_head.bin
245
+ 05/30/2024 15:22:25 - INFO - __main__ - epoch 10: {'matthews_correlation': 0.5425982560876161}
246
+ 05/30/2024 15:22:25 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_8999/adapter_config.json
247
+ 05/30/2024 15:22:25 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_8999/pytorch_adapter.bin
248
+ 05/30/2024 15:22:25 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_8999/head_config.json
249
+ 05/30/2024 15:22:25 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_8999/pytorch_model_head.bin
250
+ 05/30/2024 15:22:31 - INFO - __main__ - epoch 10: {'matthews_correlation': 0.5547785513559972}
251
+ 05/30/2024 15:22:31 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_8999/adapter_config.json
252
+ 05/30/2024 15:22:31 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_8999/pytorch_adapter.bin
253
+ 05/30/2024 15:22:31 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_8999/head_config.json
254
+ 05/30/2024 15:22:31 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_8999/pytorch_model_head.bin
255
+ 05/30/2024 15:23:24 - INFO - __main__ - epoch 11: {'matthews_correlation': 0.5353569722427551}
256
+ 05/30/2024 15:23:24 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_9999/adapter_config.json
257
+ 05/30/2024 15:23:24 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_9999/pytorch_adapter.bin
258
+ 05/30/2024 15:23:24 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_9999/head_config.json
259
+ 05/30/2024 15:23:24 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_9999/pytorch_model_head.bin
260
+ 05/30/2024 15:23:30 - INFO - __main__ - epoch 11: {'matthews_correlation': 0.5551794331953966}
261
+ 05/30/2024 15:23:30 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_9999/adapter_config.json
262
+ 05/30/2024 15:23:30 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_9999/pytorch_adapter.bin
263
+ 05/30/2024 15:23:30 - INFO - adapters.loading - Configuration saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_9999/head_config.json
264
+ 05/30/2024 15:23:30 - INFO - adapters.loading - Module weights saved in ./outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_9999/pytorch_model_head.bin
265
+ 05/30/2024 15:23:30 - INFO - __main__ - ***** Completed training *****
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/adapter_config.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5b2b147db5f916c6102e23f45ccc56602c8b2fec86d67932f37e22fe0b53180a
3
+ size 1044
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/all_results.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5310690ff28a2d6faafb09de8013388409720c1cef1cf768928ecac1172dae97
3
+ size 51
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/all_results_val.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0d74e0365b79dfa4aae0cff25add05de8bfa765e6c356e7ae68f99d5afefdf26
3
+ size 51
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/eval_res.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bdd1b9e850a862dec87ebd3b858693a28bc01cdb6e5273aa433b90c068367022
3
+ size 185687
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/gpu_stats.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:71e3cc001200f49353669c3d09ae4e8f1fe5c6041f3dbc31456008005a743795
3
+ size 5979
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/head_config.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:acf384050d7546f70d6b5043d8cfcb2e13077f0438c8576ad3f2e95e87653dd1
3
+ size 421
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/pytorch_adapter.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0f2f32f2d4c165ce1039aebf59a44c138ecef5770af5c74e8e16fe08ffa837ed
3
+ size 7191062
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/pytorch_model_head.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a12292f4f4f0b4d65ddcc111ea18c3fa0ffaec2d22b344aa7de95427d0071d2f
3
+ size 2370664
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/special_tokens_map.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b6d346be366a7d1d48332dbc9fdf3bf8960b5d879522b7799ddba59e76237ee3
3
+ size 125
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fa06adf10b0288c8416f63a3dc60480b8f3dd8820212397db6713cdc97f6ae43
3
+ size 711494
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/tokenizer_config.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3df22571b6fb8bcd6c00bf341bc0a9ad1a462b4eb1725bfefb72718999cff4a7
3
+ size 1216
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/val_res.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2cb915d497b4cd63cd625c1955b702cc429d88c51da269473d9c98e772858d6b
3
+ size 304769
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_0/vocab.txt ADDED
The diff for this file is too large to render. See raw diff
 
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/adapter_config.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5b2b147db5f916c6102e23f45ccc56602c8b2fec86d67932f37e22fe0b53180a
3
+ size 1044
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/all_results.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e7e4aa22e380a680fd4a19f99c0eb816c1aae5d95049578a40b736bc9e748b8a
3
+ size 49
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/all_results_val.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:50e80fa5806cfd3a5c2440ca8558750a34e661597623b719be1b741136167beb
3
+ size 49
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/eval_res.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e90d086e60c29edc32c774e1b06ecaeac4910894fff7d6bd85e1ffa76d33f27d
3
+ size 177106
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/gpu_stats.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ad359771e24846dbe1f43693962b772dac302f8f67f83bd0a9fd441339183f63
3
+ size 6055
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/head_config.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:acf384050d7546f70d6b5043d8cfcb2e13077f0438c8576ad3f2e95e87653dd1
3
+ size 421
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/pytorch_adapter.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a6a6a2da75d2d52918eed99023ade648224b53f1b052a7fefd287c1db703ef5e
3
+ size 7191062
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/pytorch_model_head.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e8dbe58754bab6f464f6cabd87914cd61041d269abc3591ec70581d5bf02772
3
+ size 2370664
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/special_tokens_map.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b6d346be366a7d1d48332dbc9fdf3bf8960b5d879522b7799ddba59e76237ee3
3
+ size 125
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fa06adf10b0288c8416f63a3dc60480b8f3dd8820212397db6713cdc97f6ae43
3
+ size 711494
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/tokenizer_config.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3df22571b6fb8bcd6c00bf341bc0a9ad1a462b4eb1725bfefb72718999cff4a7
3
+ size 1216
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/val_res.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3cac5ffd32f24311bfef2f760ad46b17cd7f4842f31285f70bfc8080e6efc285
3
+ size 291394
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_1999/vocab.txt ADDED
The diff for this file is too large to render. See raw diff
 
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/adapter_config.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5b2b147db5f916c6102e23f45ccc56602c8b2fec86d67932f37e22fe0b53180a
3
+ size 1044
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/all_results.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8e97c5454580e0a577391ff62b4887884b38c2e4f0be2f10ffbb10f47128597c
3
+ size 49
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/all_results_val.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:127f036bddad652ea12e1419cb60723823195c930de8ebd75b54e2e45c52fc25
3
+ size 49
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/eval_res.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1af60c9ac2339b2336ec7fd4d917ee90426a34d1374b723ca380a57b2383e916
3
+ size 176551
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/gpu_stats.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a1c69db5093cae57ad1c3a12ddf5bc2459ddf560e6e25ee7bc7f2cd13f2448a7
3
+ size 6055
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/head_config.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:acf384050d7546f70d6b5043d8cfcb2e13077f0438c8576ad3f2e95e87653dd1
3
+ size 421
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/pytorch_adapter.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d7b7e22b16b8366ec49cd8340f63646d1e93afb58e8960e5c2dc3be6f02750f9
3
+ size 7191062
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/pytorch_model_head.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:92591c99260985732342ae6067dde82d0ffdebd8292a30519c9287041a119f6a
3
+ size 2370664
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/special_tokens_map.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b6d346be366a7d1d48332dbc9fdf3bf8960b5d879522b7799ddba59e76237ee3
3
+ size 125
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fa06adf10b0288c8416f63a3dc60480b8f3dd8820212397db6713cdc97f6ae43
3
+ size 711494
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/tokenizer_config.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3df22571b6fb8bcd6c00bf341bc0a9ad1a462b4eb1725bfefb72718999cff4a7
3
+ size 1216
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/val_res.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fa61010286d51bbc58c1daaa66385814510185276f3ea0eb6a17722faa002ab1
3
+ size 290478
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_2999/vocab.txt ADDED
The diff for this file is too large to render. See raw diff
 
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_3999/adapter_config.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5b2b147db5f916c6102e23f45ccc56602c8b2fec86d67932f37e22fe0b53180a
3
+ size 1044
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_3999/all_results.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:12595e64dabbe54a88115349f4e3a6060d2960b905e840e2f4f7adc226c3dacf
3
+ size 49
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_3999/all_results_val.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:23a2f4d13089808af2fcce749642a6f8d4197d326479ffcd21832ed6b601858e
3
+ size 49
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_3999/eval_res.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a089981167598afb50094161b4fd2c9a87b26c758a9956bfe9d7ce5132996684
3
+ size 176154
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_3999/gpu_stats.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3227c81d715011800f02db7872d5df521454a5532c7371374497eca8e980033b
3
+ size 6055
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_3999/head_config.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:acf384050d7546f70d6b5043d8cfcb2e13077f0438c8576ad3f2e95e87653dd1
3
+ size 421
outputs/cola/bert-base-uncased_adapterstrain_val_0.0001_65_8_10000/step_3999/pytorch_adapter.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f1882982079bf45a509155a1e297549f958523f630066bc22d94e6517ccc0642
3
+ size 7191062