Sayan01 commited on
Commit
76e7e27
·
1 Parent(s): 14e9e2d

Training in progress, epoch 1

Browse files
config.json CHANGED
@@ -1,5 +1,6 @@
1
  {
2
- "_name_or_path": "google/bert_uncased_L-6_H-768_A-12",
 
3
  "adapters": {
4
  "adapters": {},
5
  "config_map": {},
@@ -7,34 +8,32 @@
7
  "fusions": {}
8
  },
9
  "architectures": [
10
- "BertForSequenceClassification"
11
  ],
12
- "attention_probs_dropout_prob": 0.1,
13
- "classifier_dropout": null,
14
- "hidden_act": "gelu",
15
- "hidden_dropout_prob": 0.1,
16
- "hidden_size": 768,
17
  "id2label": {
18
  "0": "entailment",
19
  "1": "not_entailment"
20
  },
21
  "initializer_range": 0.02,
22
- "intermediate_size": 3072,
23
  "label2id": {
24
  "entailment": "0",
25
  "not_entailment": "1"
26
  },
27
- "layer_norm_eps": 1e-12,
28
  "max_position_embeddings": 512,
29
- "model_type": "bert",
30
- "num_attention_heads": 12,
31
- "num_hidden_layers": 6,
32
  "pad_token_id": 0,
33
- "position_embedding_type": "absolute",
34
  "problem_type": "single_label_classification",
 
 
 
 
35
  "torch_dtype": "float32",
36
  "transformers_version": "4.21.3",
37
- "type_vocab_size": 2,
38
- "use_cache": true,
39
  "vocab_size": 30522
40
  }
 
1
  {
2
+ "_name_or_path": "distilbert-base-uncased",
3
+ "activation": "gelu",
4
  "adapters": {
5
  "adapters": {},
6
  "config_map": {},
 
8
  "fusions": {}
9
  },
10
  "architectures": [
11
+ "DistilBertForSequenceClassification"
12
  ],
13
+ "attention_dropout": 0.1,
14
+ "dim": 768,
15
+ "dropout": 0.1,
16
+ "hidden_dim": 3072,
 
17
  "id2label": {
18
  "0": "entailment",
19
  "1": "not_entailment"
20
  },
21
  "initializer_range": 0.02,
 
22
  "label2id": {
23
  "entailment": "0",
24
  "not_entailment": "1"
25
  },
 
26
  "max_position_embeddings": 512,
27
+ "model_type": "distilbert",
28
+ "n_heads": 12,
29
+ "n_layers": 6,
30
  "pad_token_id": 0,
 
31
  "problem_type": "single_label_classification",
32
+ "qa_dropout": 0.1,
33
+ "seq_classif_dropout": 0.2,
34
+ "sinusoidal_pos_embds": false,
35
+ "tie_weights_": true,
36
  "torch_dtype": "float32",
37
  "transformers_version": "4.21.3",
 
 
38
  "vocab_size": 30522
39
  }
logs/1663304779.8945827/events.out.tfevents.1663304779.cd02b94e3af8.5814.4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2f2d269b5491445066f65ba4f705f7cd1d1b306577043233e33286a40b3fa5d3
3
+ size 5437
logs/1663305699.3569372/events.out.tfevents.1663305699.cd02b94e3af8.5814.6 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a47acabe3a336cbef74ba2d3f54a8cf9cfde64db5fe0ba1de3d398f773450813
3
+ size 5437
logs/events.out.tfevents.1663302864.cd02b94e3af8.5814.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e94be61ddd1d218cce0d09a76d1042bfe890598f246aca9e30466f056c2923c8
3
- size 4631
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:256212836eec81d4f0dcdac4a464f8298fc88327d9c401038d52bc9e1477590c
3
+ size 14087
logs/events.out.tfevents.1663303742.cd02b94e3af8.5814.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a8d92f1e085e6affb8f7bb8e04c87b77e4d6adba4a49c23345263f53a80eeae4
3
+ size 363
logs/events.out.tfevents.1663304773.cd02b94e3af8.5814.3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9cd8c1b61017506c2b3b465fb28524e459fec81bc4a4cef144b7ced0e5176a1a
3
+ size 4721
logs/events.out.tfevents.1663305696.cd02b94e3af8.5814.5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:17d7e47f034bb29b015a104efba4d4cdaf9ab6aa948a47f77a7433a95e3282dd
3
+ size 4586
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1aa1ae283b84f37efb07ebfc01393aa69d8e59ca68deb868b0103cd8c8856149
3
- size 267864813
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:de8f16a0e2beeb1c437a68937553020bedfe3b7889b4f9404a6546bb071dddae
3
+ size 267854321