Sayan01 commited on
Commit
417e2e4
·
1 Parent(s): e08e54f

Training in progress, epoch 1

Browse files
config.json CHANGED
@@ -1,6 +1,5 @@
1
  {
2
- "_name_or_path": "distilbert-base-uncased",
3
- "activation": "gelu",
4
  "adapters": {
5
  "adapters": {},
6
  "config_map": {},
@@ -8,32 +7,34 @@
8
  "fusions": {}
9
  },
10
  "architectures": [
11
- "DistilBertForSequenceClassification"
12
  ],
13
- "attention_dropout": 0.1,
14
- "dim": 768,
15
- "dropout": 0.1,
16
- "hidden_dim": 3072,
 
17
  "id2label": {
18
  "0": "entailment",
19
  "1": "not_entailment"
20
  },
21
  "initializer_range": 0.02,
 
22
  "label2id": {
23
  "entailment": "0",
24
  "not_entailment": "1"
25
  },
 
26
  "max_position_embeddings": 512,
27
- "model_type": "distilbert",
28
- "n_heads": 12,
29
- "n_layers": 6,
30
  "pad_token_id": 0,
 
31
  "problem_type": "single_label_classification",
32
- "qa_dropout": 0.1,
33
- "seq_classif_dropout": 0.2,
34
- "sinusoidal_pos_embds": false,
35
- "tie_weights_": true,
36
  "torch_dtype": "float32",
37
  "transformers_version": "4.21.3",
 
 
38
  "vocab_size": 30522
39
  }
 
1
  {
2
+ "_name_or_path": "google/bert_uncased_L-6_H-768_A-12",
 
3
  "adapters": {
4
  "adapters": {},
5
  "config_map": {},
 
7
  "fusions": {}
8
  },
9
  "architectures": [
10
+ "BertForSequenceClassification"
11
  ],
12
+ "attention_probs_dropout_prob": 0.1,
13
+ "classifier_dropout": null,
14
+ "hidden_act": "gelu",
15
+ "hidden_dropout_prob": 0.1,
16
+ "hidden_size": 768,
17
  "id2label": {
18
  "0": "entailment",
19
  "1": "not_entailment"
20
  },
21
  "initializer_range": 0.02,
22
+ "intermediate_size": 3072,
23
  "label2id": {
24
  "entailment": "0",
25
  "not_entailment": "1"
26
  },
27
+ "layer_norm_eps": 1e-12,
28
  "max_position_embeddings": 512,
29
+ "model_type": "bert",
30
+ "num_attention_heads": 12,
31
+ "num_hidden_layers": 6,
32
  "pad_token_id": 0,
33
+ "position_embedding_type": "absolute",
34
  "problem_type": "single_label_classification",
 
 
 
 
35
  "torch_dtype": "float32",
36
  "transformers_version": "4.21.3",
37
+ "type_vocab_size": 2,
38
+ "use_cache": true,
39
  "vocab_size": 30522
40
  }
logs/1663299284.0096412/events.out.tfevents.1663299284.cd02b94e3af8.3404.3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c4fdaa9b84b2116f2d4eedb5228b414c992c4d1fc2955a3bdcd8a860db3c4f7d
3
+ size 5437
logs/events.out.tfevents.1663298059.cd02b94e3af8.3404.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:eb2fdd11ec9796d0410902230e7e3e3a68610f29040f0247933bab5bc9862b73
3
- size 4586
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f6cbe7e6d6390e1f87e0e1496e84f51c5d03d668a149e949233c6760f292d399
3
+ size 10808
logs/events.out.tfevents.1663299279.cd02b94e3af8.3404.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:659377f41c236c5d51108524dd023c107d593bcbb4ab84a676a52d3bdab12ce6
3
+ size 4631
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a651a5dfe4a8f5c59fdb7b37447db58c95af90d3a05e2f1f5d124754514bc0a3
3
- size 267854321
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cd7f14ed8348d149a2a10dcc13ad1ae2342830ee8f94b2758599c3e8993bd06a
3
+ size 267864813
tokenizer.json CHANGED
@@ -1,11 +1,6 @@
1
  {
2
  "version": "1.0",
3
- "truncation": {
4
- "direction": "Right",
5
- "max_length": 512,
6
- "strategy": "LongestFirst",
7
- "stride": 0
8
- },
9
  "padding": null,
10
  "added_tokens": [
11
  {
 
1
  {
2
  "version": "1.0",
3
+ "truncation": null,
 
 
 
 
 
4
  "padding": null,
5
  "added_tokens": [
6
  {