Sayan01 commited on
Commit
f82139b
·
1 Parent(s): cdd3bb9

Training in progress, epoch 1

Browse files
config.json CHANGED
@@ -1,19 +1,27 @@
1
  {
2
- "_name_or_path": "google/bert_uncased_L-2_H-128_A-2",
 
 
 
 
 
 
3
  "architectures": [
4
  "BertForSequenceClassification"
5
  ],
6
  "attention_probs_dropout_prob": 0.1,
 
7
  "classifier_dropout": null,
 
8
  "hidden_act": "gelu",
9
  "hidden_dropout_prob": 0.1,
10
- "hidden_size": 128,
11
  "id2label": {
12
  "0": "entailment",
13
  "1": "not_entailment"
14
  },
15
  "initializer_range": 0.02,
16
- "intermediate_size": 512,
17
  "label2id": {
18
  "entailment": "0",
19
  "not_entailment": "1"
@@ -21,13 +29,15 @@
21
  "layer_norm_eps": 1e-12,
22
  "max_position_embeddings": 512,
23
  "model_type": "bert",
24
- "num_attention_heads": 2,
25
- "num_hidden_layers": 2,
26
  "pad_token_id": 0,
27
  "position_embedding_type": "absolute",
 
28
  "problem_type": "single_label_classification",
 
29
  "torch_dtype": "float32",
30
- "transformers_version": "4.20.1",
31
  "type_vocab_size": 2,
32
  "use_cache": true,
33
  "vocab_size": 30522
 
1
  {
2
+ "_name_or_path": "huawei-noah/TinyBERT_General_4L_312D",
3
+ "adapters": {
4
+ "adapters": {},
5
+ "config_map": {},
6
+ "fusion_config_map": {},
7
+ "fusions": {}
8
+ },
9
  "architectures": [
10
  "BertForSequenceClassification"
11
  ],
12
  "attention_probs_dropout_prob": 0.1,
13
+ "cell": {},
14
  "classifier_dropout": null,
15
+ "emb_size": 312,
16
  "hidden_act": "gelu",
17
  "hidden_dropout_prob": 0.1,
18
+ "hidden_size": 312,
19
  "id2label": {
20
  "0": "entailment",
21
  "1": "not_entailment"
22
  },
23
  "initializer_range": 0.02,
24
+ "intermediate_size": 1200,
25
  "label2id": {
26
  "entailment": "0",
27
  "not_entailment": "1"
 
29
  "layer_norm_eps": 1e-12,
30
  "max_position_embeddings": 512,
31
  "model_type": "bert",
32
+ "num_attention_heads": 12,
33
+ "num_hidden_layers": 4,
34
  "pad_token_id": 0,
35
  "position_embedding_type": "absolute",
36
+ "pre_trained": "",
37
  "problem_type": "single_label_classification",
38
+ "structure": [],
39
  "torch_dtype": "float32",
40
+ "transformers_version": "4.17.0",
41
  "type_vocab_size": 2,
42
  "use_cache": true,
43
  "vocab_size": 30522
logs/1657875633.2928188/events.out.tfevents.1657875633.3130d29d5be1.142.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e91b5e896e35b2bfbd6c57260302613406e8de6f332db6a2cf2ec82bd23d1bd1
3
+ size 4783
logs/1657877289.747412/events.out.tfevents.1657877289.3130d29d5be1.142.3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7940361a4e8b35cd8c734783fe1ecf5e8c640ee5521e0d0b778af44473365568
3
+ size 4783
logs/events.out.tfevents.1657875633.3130d29d5be1.142.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1144fd121976ed025eb3d497d0db9f11eb292a3dd9ffc762fdbfd67ce9dd00af
3
+ size 3571
logs/events.out.tfevents.1657877289.3130d29d5be1.142.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c0f5d478d06cb5e0492993a37fe7f9c3c379c3c77801f56f0087042db977c91a
3
+ size 4051
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9ea04600c88de119aa14eebdb5b004ad4c07b728d41a04381cbd55414c1f8cea
3
- size 17561831
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e963ed457af2b5f745a551a432e25a5b6a10293c5fccb3af447d495219922d33
3
+ size 57436935
special_tokens_map.json CHANGED
@@ -1,7 +1 @@
1
- {
2
- "cls_token": "[CLS]",
3
- "mask_token": "[MASK]",
4
- "pad_token": "[PAD]",
5
- "sep_token": "[SEP]",
6
- "unk_token": "[UNK]"
7
- }
 
1
+ {"unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]"}
 
 
 
 
 
 
tokenizer_config.json CHANGED
@@ -1,14 +1 @@
1
- {
2
- "cls_token": "[CLS]",
3
- "do_lower_case": true,
4
- "mask_token": "[MASK]",
5
- "model_max_length": 512,
6
- "name_or_path": "bert-base-uncased",
7
- "pad_token": "[PAD]",
8
- "sep_token": "[SEP]",
9
- "special_tokens_map_file": null,
10
- "strip_accents": null,
11
- "tokenize_chinese_chars": true,
12
- "tokenizer_class": "BertTokenizer",
13
- "unk_token": "[UNK]"
14
- }
 
1
+ {"do_lower_case": true, "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "tokenize_chinese_chars": true, "strip_accents": null, "model_max_length": 512, "special_tokens_map_file": null, "name_or_path": "bert-base-uncased", "tokenizer_class": "BertTokenizer"}
 
 
 
 
 
 
 
 
 
 
 
 
 
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:992d7eebc18ef81555911bbef662e0dde5c4306806ecd28fe8f037a66bc5eacd
3
- size 3375
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4d0e38336921e437638b72b7a7ad7c7c75f823a442cc96d52d62ebf541738640
3
+ size 3055