Rishu115 commited on
Commit
6c16e36
·
1 Parent(s): 446703f

Training in progress epoch 0

Browse files
.gitattributes CHANGED
@@ -32,3 +32,5 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
32
  *.zip filter=lfs diff=lfs merge=lfs -text
33
  *.zst filter=lfs diff=lfs merge=lfs -text
34
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
32
  *.zip filter=lfs diff=lfs merge=lfs -text
33
  *.zst filter=lfs diff=lfs merge=lfs -text
34
  *tfevents* filter=lfs diff=lfs merge=lfs -text
35
+ checkpoint/variables/variables.data-00000-of-00001 filter=lfs diff=lfs merge=lfs -text
36
+ checkpoint/variables/variables.index filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ tags:
3
+ - generated_from_keras_callback
4
+ model-index:
5
+ - name: Rishu115/mlm-bert-train_finalTraining_changedLR2
6
+ results: []
7
+ ---
8
+
9
+ <!-- This model card has been generated automatically according to the information Keras had access to. You should
10
+ probably proofread and complete it, then remove this comment. -->
11
+
12
+ # Rishu115/mlm-bert-train_finalTraining_changedLR2
13
+
14
+ This model was trained from scratch on an unknown dataset.
15
+ It achieves the following results on the evaluation set:
16
+ - Train Loss: 0.9465
17
+ - Validation Loss: 0.8696
18
+ - Epoch: 0
19
+
20
+ ## Model description
21
+
22
+ More information needed
23
+
24
+ ## Intended uses & limitations
25
+
26
+ More information needed
27
+
28
+ ## Training and evaluation data
29
+
30
+ More information needed
31
+
32
+ ## Training procedure
33
+
34
+ ### Training hyperparameters
35
+
36
+ The following hyperparameters were used during training:
37
+ - optimizer: {'name': 'AdamWeightDecay', 'learning_rate': {'class_name': 'WarmUp', 'config': {'initial_learning_rate': 5e-05, 'decay_schedule_fn': {'class_name': 'PolynomialDecay', 'config': {'initial_learning_rate': 5e-05, 'decay_steps': 47396, 'end_learning_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}, '__passive_serialization__': True}, 'warmup_steps': 1000, 'power': 1.0, 'name': None}}, 'decay': 0.0, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-08, 'amsgrad': False, 'weight_decay_rate': 0.01}
38
+ - training_precision: float32
39
+
40
+ ### Training results
41
+
42
+ | Train Loss | Validation Loss | Epoch |
43
+ |:----------:|:---------------:|:-----:|
44
+ | 0.9465 | 0.8696 | 0 |
45
+
46
+
47
+ ### Framework versions
48
+
49
+ - Transformers 4.23.1
50
+ - TensorFlow 2.10.0
51
+ - Datasets 2.10.1
52
+ - Tokenizers 0.13.2
added_tokens.json ADDED
@@ -0,0 +1,216 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "a.": 30534,
3
+ "acknowledgement": 30656,
4
+ "addressee": 30617,
5
+ "adj": 30551,
6
+ "andor": 30555,
7
+ "asap": 30559,
8
+ "aspxxmllocationteamsmlicodcod_v": 30683,
9
+ "atd": 30645,
10
+ "attachments": 30556,
11
+ "average fuel": 30694,
12
+ "b.": 30545,
13
+ "baf": 30531,
14
+ "baseline rotterdam": 30707,
15
+ "benchmark ports": 30693,
16
+ "bhd": 30660,
17
+ "bij": 30672,
18
+ "bkg": 30569,
19
+ "bl": 30522,
20
+ "blading": 30547,
21
+ "bls": 30626,
22
+ "booked_by": 30602,
23
+ "bookings": 30542,
24
+ "bremerhaven": 30641,
25
+ "brgds": 30670,
26
+ "bunker fuel": 30712,
27
+ "bunker prices": 30691,
28
+ "bunker qualities": 30705,
29
+ "bunker reimbursement": 30702,
30
+ "bunkerworld": 30706,
31
+ "c.": 30590,
32
+ "calculated average": 30710,
33
+ "cnee": 30648,
34
+ "cntr": 30604,
35
+ "co.": 30557,
36
+ "cof": 30625,
37
+ "confidentiality": 30624,
38
+ "consignee": 30529,
39
+ "cont": 30669,
40
+ "contract outset": 30704,
41
+ "covid": 30564,
42
+ "credit_party": 30609,
43
+ "d.": 30550,
44
+ "ddf": 30565,
45
+ "dears": 30595,
46
+ "delete": 30552,
47
+ "demurrage": 30543,
48
+ "demurragedetention": 30570,
49
+ "departutearrival": 30666,
50
+ "dest": 30616,
51
+ "destination lsmgo": 30695,
52
+ "dg": 30638,
53
+ "discharging": 30661,
54
+ "disclaimer": 30563,
55
+ "dit": 30680,
56
+ "docs": 30598,
57
+ "e.": 30560,
58
+ "edi": 30629,
59
+ "een": 30677,
60
+ "etd": 30528,
61
+ "ets": 30541,
62
+ "eur": 30553,
63
+ "expressly": 30620,
64
+ "ext": 30568,
65
+ "f.": 30588,
66
+ "fak": 30571,
67
+ "fax": 30532,
68
+ "fcl": 30659,
69
+ "fmc": 30628,
70
+ "fnd": 30548,
71
+ "formserver": 30681,
72
+ "forwarded": 30678,
73
+ "forwarder": 30596,
74
+ "forwarders": 30647,
75
+ "forwarding": 30575,
76
+ "freundlichen": 30655,
77
+ "fuel price": 30696,
78
+ "fuel type": 30713,
79
+ "fw": 30540,
80
+ "gated": 30631,
81
+ "gatein": 30685,
82
+ "gcss": 30606,
83
+ "geen": 30664,
84
+ "groeten": 30635,
85
+ "h.": 30593,
86
+ "hasu": 30574,
87
+ "haulage": 30587,
88
+ "hbl": 30674,
89
+ "i.": 30614,
90
+ "ifo": 30720,
91
+ "ifo 180": 30729,
92
+ "ifo 380": 30728,
93
+ "imo": 30654,
94
+ "imo compliant": 30699,
95
+ "imo2020": 30730,
96
+ "indien": 30679,
97
+ "intl": 30619,
98
+ "inv": 30640,
99
+ "invoice": 30523,
100
+ "invoice_party": 30607,
101
+ "invoices": 30583,
102
+ "kgs": 30636,
103
+ "l.": 30572,
104
+ "lading": 30537,
105
+ "lng 380e": 30734,
106
+ "lng mgoe": 30735,
107
+ "longstanding": 30610,
108
+ "low sulfur": 30692,
109
+ "ls 180": 30732,
110
+ "ls 380": 30731,
111
+ "lsmgo": 30717,
112
+ "m.": 30544,
113
+ "maeu": 30524,
114
+ "marine gas": 30689,
115
+ "mbl": 30597,
116
+ "mdo": 30733,
117
+ "metric tonne": 30718,
118
+ "mgo": 30721,
119
+ "mnbu": 30589,
120
+ "moerdijk": 30687,
121
+ "mrku": 30525,
122
+ "mrsu": 30567,
123
+ "msds": 30632,
124
+ "msk": 30603,
125
+ "msku": 30527,
126
+ "msl": 30675,
127
+ "multimodal": 30682,
128
+ "n.": 30577,
129
+ "niet": 30651,
130
+ "ningbo": 30622,
131
+ "notify": 30533,
132
+ "o.": 30576,
133
+ "obl": 30579,
134
+ "oceanfreight": 30546,
135
+ "ocf": 30549,
136
+ "oil lsmgo": 30690,
137
+ "onc": 30653,
138
+ "oncarriage": 30643,
139
+ "ons": 30662,
140
+ "origin lsmgo": 30698,
141
+ "outside eca": 30688,
142
+ "p.": 30566,
143
+ "payer": 30530,
144
+ "png": 30673,
145
+ "ponu": 30580,
146
+ "prepaid": 30561,
147
+ "price_owner": 30627,
148
+ "prices used": 30703,
149
+ "pte": 30639,
150
+ "pty": 30634,
151
+ "pvt": 30591,
152
+ "queries": 30615,
153
+ "quotation": 30642,
154
+ "r.": 30605,
155
+ "reefer": 30582,
156
+ "reexport": 30667,
157
+ "revert": 30594,
158
+ "revise": 30599,
159
+ "rme": 30723,
160
+ "rmf": 30724,
161
+ "rmg": 30725,
162
+ "rmh": 30726,
163
+ "rmk": 30727,
164
+ "s.": 30535,
165
+ "safmarine": 30671,
166
+ "sdn": 30665,
167
+ "sealand": 30539,
168
+ "seca component": 30714,
169
+ "sender": 30536,
170
+ "ship bunker": 30711,
171
+ "shipandbunker": 30708,
172
+ "shipper": 30526,
173
+ "shippers": 30608,
174
+ "soonest": 30637,
175
+ "sudu": 30554,
176
+ "sulfur bunker": 30719,
177
+ "sulphur": 30658,
178
+ "surcharge": 30592,
179
+ "surcharges": 30618,
180
+ "swb": 30611,
181
+ "t.": 30573,
182
+ "tcku": 30584,
183
+ "tclu": 30613,
184
+ "tcnu": 30612,
185
+ "teamsmlicod_layouts": 30684,
186
+ "telex": 30601,
187
+ "tetra": 30649,
188
+ "tghu": 30676,
189
+ "thc": 30558,
190
+ "thcd": 30562,
191
+ "tllu": 30630,
192
+ "tp": 30686,
193
+ "tpdoc": 30586,
194
+ "transport_document_receiver": 30623,
195
+ "transporter": 30657,
196
+ "trhu": 30621,
197
+ "twill": 30650,
198
+ "u.": 30652,
199
+ "uw": 30663,
200
+ "v.": 30538,
201
+ "vgm": 30581,
202
+ "vlsfo": 30716,
203
+ "vlsfo low": 30715,
204
+ "vlsfo price": 30701,
205
+ "vlsfo prices": 30697,
206
+ "vlsfo pricing": 30700,
207
+ "vlsfo st": 30709,
208
+ "voor": 30646,
209
+ "vriendelijke": 30600,
210
+ "vsl": 30644,
211
+ "w.": 30633,
212
+ "warmest": 30578,
213
+ "waybill": 30585,
214
+ "wti": 30722,
215
+ "x.": 30668
216
+ }
checkpoint/keras_metadata.pb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a5a5569e874bfb45260a5b5c736f15a1240bc9cba88bfce101b2dac130892bd7
3
+ size 168139
checkpoint/saved_model.pb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:656525d057726c4bd92461d90185021a81a5c026bb03638a3a903825005a3e4d
3
+ size 9157727
checkpoint/variables/variables.data-00000-of-00001 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5db4e5a1b9cd03955319c6059066b2f1ae6d6dad251bb3f816b1d960dc42ab0a
3
+ size 1316390277
checkpoint/variables/variables.index ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2ee25386f12ce681b03b7abcac6178d40f0cbf386dfb9e31a1da3d522bcce462
3
+ size 40749
config.json ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "/dbfs/mnt/crm_email_classification/manikant/bert_pretrain_V2/bert_model_Updated_V2",
3
+ "architectures": [
4
+ "BertForMaskedLM"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "gradient_checkpointing": false,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 3072,
14
+ "layer_norm_eps": 1e-12,
15
+ "max_position_embeddings": 512,
16
+ "model_type": "bert",
17
+ "num_attention_heads": 12,
18
+ "num_hidden_layers": 12,
19
+ "pad_token_id": 0,
20
+ "position_embedding_type": "absolute",
21
+ "torch_dtype": "float32",
22
+ "transformers_version": "4.23.1",
23
+ "type_vocab_size": 2,
24
+ "use_cache": true,
25
+ "vocab_size": 30736
26
+ }
special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
tf_model.h5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d12b3b8cb527965a86e3263c8e0406d3c8ea2158a5921ef856a9e70cd3ec9ff6
3
+ size 535003352
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "do_lower_case": true,
4
+ "mask_token": "[MASK]",
5
+ "model_max_length": 512,
6
+ "name_or_path": "/dbfs/mnt/crm_email_classification/manikant/bert_pretrain_V2/4L_allIndex_tokenizer",
7
+ "pad_token": "[PAD]",
8
+ "sep_token": "[SEP]",
9
+ "special_tokens_map_file": null,
10
+ "strip_accents": null,
11
+ "tokenize_chinese_chars": true,
12
+ "tokenizer_class": "BertTokenizer",
13
+ "unk_token": "[UNK]"
14
+ }
vocab.txt ADDED
The diff for this file is too large to render. See raw diff